Editing
Advancements in Neural Network-Based TTS (2000s)
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==== <big>Articulatory Features-Based TTS</big> ==== Articulatory feature-based Text-to-Speech (TTS) is a concept that involves using articulatory features, which represent the movements and positions of the speech articulators (such as the tongue, lips, and jaw), as the basis for synthesizing speech. This approach aims to capture the detailed articulatory information present in the speech signal, allowing for more natural and expressive speech synthesis.<ref name=":3">Singampalli, V. D. (2010). ''Statistical identification of articulatory roles in speech production'' (Order No. 10131268). Available from ProQuest Dissertations & Theses A&I. (1810640121). Retrieved from <nowiki>http://server.proxy-ub.rug.nl/login?url=https://www.proquest.com/dissertations-theses/statistical-identification-articulatory-roles/docview/1810640121/se-2</nowiki></ref> '''Innovation:''' * Utilization of Biophysical Phonetics: By incorporating articulatory models and other biophysical phonetic information, this approach enhances the quality and naturalness of speech synthesis. * Enhanced Speech Quality: Articulatory Features-Based TTS improves the quality of synthesized speech by considering articulatory features, making it more closely resemble natural human speech. * Addressing Shortcomings of Traditional TTS: This method aims to compensate for the limitations of traditional TTS systems, particularly in terms of naturalness and quality of speech. * Improved Control Capabilities: Articulatory Features-Based TTS offers enhanced control over speech synthesis, enabling users to adjust parameters such as pitch, speed, and other characteristics. * Data-Driven Learning: This approach leans towards data-driven learning, reducing reliance on manual rules and models for speech synthesis. * These innovations have the potential to enhance the performance of speech synthesis systems, bringing them closer to natural human speech while providing greater control over the synthesized output.<ref name=":3" />
Summary:
Please note that all contributions to MSc Voice Technology are considered to be released under the Creative Commons Attribution (see
MSc Voice Technology:Copyrights
for details). If you do not want your writing to be edited mercilessly and redistributed at will, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource.
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
Edit source
View history
More
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information