Speech & Audio AI

Quantifying Speaker Embedding Phonological Rule Interactions in Accented Speech Synthesis

TThanathai LertpetchpunYYoonjeong LeeTThanapat TrachuJJihwan LeeTTiantian FengDDani ByrdSShrikanth Narayanan
arXiv ID
2601.14417
Published
January 20, 2026
Authors
7
Hugging Face Likes
5
Comments
2

Abstract

Many spoken languages, including English, exhibit wide variation in dialects and accents, making accent control an important capability for flexible text-to-speech (TTS) models. Current TTS systems typically generate accented speech by conditioning on speaker embeddings associated with specific accents. While effective, this approach offers limited interpretability and controllability, as embeddings also encode traits such as timbre and emotion. In this study, we analyze the interaction between speaker embeddings and linguistically motivated phonological rules in accented speech synthesis. Using American and British English as a case study, we implement rules for flapping, rhoticity, and vowel correspondences. We propose the phoneme shift rate (PSR), a novel metric quantifying how strongly embeddings preserve or override rule-based transformations. Experiments show that combining rules with embeddings yields more authentic accents, while embeddings can attenuate or overwrite rules, revealing entanglement between accent and speaker identity. Our findings highlight rules as a lever for accent control and a framework for evaluating disentanglement in speech generation.

Keywords

text-to-speechspeaker embeddingsphonological rulesaccent controlphoneme shift rateflappingrhoticityvowel correspondences

More in Speech & Audio AI

View all
Quantifying Speaker Embedding Phonological Rule Interactions in Accented Speech Synthesis | Paperchime