Editing
Development of End-to-End Models
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
==== 1.The Impact on the Field of Speech Recognition ==== ===== Simplify Processes and Reduce Labor Costs: ===== The end-to-end model eliminates the need for multiple processing steps and manual feature engineering, reducing the time and effort to develop and maintain the model. In the aspect of speech recognition, the end-to-end model simplifies the architecture of the speech recognition system, eliminating the need for complex hand-designed feature extractors or acoustic models. ===== Better Performance: ===== End-to-end models can extract relevant features directly from the original data, sometimes better than traditional processes to capture the complex relationships of the data, so the performance of the model may be improved. The end-to-end model can better capture the complex relationship between audio and text, which usually enables speech recognition to achieve higher recognition accuracy than traditional methods. ===== More Extensive Applications: ===== The end-to-end model can be applied to a variety of tasks, such as speech recognition, machine translation, image processing and so on, making the solutions of these tasks easier to implement and popularize. For speech recognition, this simplified method can be easily applied to a variety of speech recognition tasks, including multiple languages, accents from all over the world, language habits of various cultural backgrounds and speech recognition in specific fields. ===== Faster Experimental Iteration: ===== The simplified modeling process enables researchers to try new ideas and algorithms faster and faster, accelerating research progress. The end-to-end model provides more flexibility and innovative possibilities for researchers in the field of speech recognition. ===== More Efficient Training: ===== End-to-end models may require more data to train, and more computer resources to process and train these large models. But at the same time, end-to-end training can provide more efficient training for speech recognition, and direct training from original audio to text can make more efficient use of a large number of tagged data and improve the performance of the model. ===== Listen, Attend and Spell (LAS) ===== Listen, Attend and Spell (LAS), is an attention-based neural network that can di-rectly transcribe acoustic signals to characters. LAS is based on the sequence to sequence framework with a pyramid structure in the encoder that reduces the number of timesteps that the decoder has to attend to. LAS is trained end-to-end and has two main components: a listener and a speller. The listener is a pyramidal recurrent net- work encoder that accepts filter bank spectra as inputs. The speller is an attention- based recurrent network decoder that emits characters as outputs. The network produces character sequences without making any independence assumptions be- tween the characters. This is the key improvement of LAS over previous end-to-end [[wikipedia:Connectionist_temporal_classification|CTC models]]. On a subset of the Google voice search task, LAS achieves a word error rate (WER) of 14.1% without a dictionary or a language model, and 10.3% with language model rescoring over the top 32 beams. By comparison, the state-of-the-art [https://static.googleusercontent.com/media/research.google.com/zh-CN//pubs/archive/43455.pdf CLDNN]-HMM model achieves a WER of 8.0%.<ref>Chan, William, Navdeep Jaitly, Quoc V. Le和Oriol Vinyals. "Listen, Attend and Spell". arXiv, 19.08.2015. http://arxiv.org/abs/1508.01211.</ref>
Summary:
Please note that all contributions to MSc Voice Technology are considered to be released under the Creative Commons Attribution (see
MSc Voice Technology:Copyrights
for details). If you do not want your writing to be edited mercilessly and redistributed at will, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource.
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
Edit source
View history
More
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information