Carnegie Mellon's Harpy System: Difference between revisions
No edit summary |
|||
Line 1: | Line 1: | ||
Sherry Yu-Ting | Sherry Yu-Ting | ||
Wansu | Wansu Zhu | ||
Jingsi Huang | |||
Wenjun Meng | Wenjun Meng |
Revision as of 21:05, 18 September 2023
Sherry Yu-Ting
Wansu Zhu
Jingsi Huang
Wenjun Meng
Carnegie Mellon's Harpy System
Jump to navigationJump to search
Introduction
Harpy System is a result of combing and improving the best features from two speech systems previously developed at Carnegie Mellon University: Hearsay-I ( Erman 1974 ) and Dragon ( Baker 1975 ). Two feature of the Harpy system that led to a successful demonstration are its representation of knowledge and the use of new search techniques. In developing speech recognition systems, it is necessary to devise a means of acquiring and representing the many diverse types of knowledge that characterize speech. We must also develop matching and searching techniques that convert this passive knowledge into an active process for understanding the utterance in the presence of error, noise, and uncertainty.
When an utterance is input to the computer. Several pre-processing steps are taken to prepare the data for recognition. The utterance is segmented into acoustic units and analyzed to determine the segmented features and parameters. The goal of recognition tasks is to find an optimal sequence of phones satisfying two criteria: the sequence must represent a legal path through the knowledge network and should consist of phones with high acoustic matches.
Harpy Systems uses a beam search to locate this optimal sequence of phones. Harpy begins the beam search by taking all legal phones from the start of the sentence from the knowledge network and entering them in a recognition tree. Next, a path probability is calculated for each candidate. The path with the best probability is determined and remaining candidates are compared with it. Those that fall below a threshold of acceptability are eliminated from further searching. The successors of the surviving candidates are expanded based on the information in the knowledge network.
Hearsay-I System | Dragon System | Harpy System | |
---|---|---|---|
Sources of Knowledge |
|
|
|
Knowledge Representation | Procedural embedding. | Markove networks. | Transition networks. |
Search Strategy | Best-First with backtracking. | All paths in parallel ( search all the possible syntactic
and acoustic paths through the network in parallel to determine optimal path ) with no backtracking. |
''best few'' in parallel with no backtracking.
Beam Search
|
Segmentation | Yes
( uses it to reduce effective utterance length ) |
No | Yes |
Strength |
|
|
N/A |
Weakness |
|
|
N/A |
- simple intro of the relationship between hearsay-I, Dragon system, and harpy system
- general info about how harpy system works (features, methods
- simply mention the advatages of harpy sytem
Historical Context
Before 1970s, the speech recognition systems available in the world could only understand 16 words spoken in English. With the goal of ending up with a machine capable of understanding a minimum of 1,000 words, speech recognition technology made major strides in the 1970s, thanks to interest and funding from the U.S. Department of Defense. The DoD’s DARPA Speech Understanding Research (SUR) program, from 1971 to 1976, funds five years of speech recognition research leading to the creation of the Harpy by Carnegie Mellon, a machine capable of understanding 1,011 words, approximately the vocabulary of an average three-year-old.
The Harpy connected speech recognition system is the result of an attempt to understand the relative importance of various design choices of two earlier speech recognition systems developed at Carnegie-Mellon University: the Hearsay-I system and the Dragon system. Systematic performance analysis of various design choices of these two systems resulted in the Harpy system.
Hearsay-I System
- The Hearsay-I System must use all these sources of knowledge effectively in the Hearsay System, this is achieved by representing knowledge as as a set of cooperating parallel process. The Hearsay System also use the source of knowledge (acoustics-phonetics, prosody, lexicon, syntax, semantics and pragmatics) to generate hypothesis about what words might appear. The Hearsay System is incrementally trying to match words to the un-analyzed portion of the utterance
- If the resulting word matches are not optimal, the system backs up and tries a different path. The technique was known as probabilistic tree. The Hearsay System cannot recognize sentences in unrestricted English and can only recognize simple phrase structure languages.
- To be acceptable as an input medium, the system must not only perform with high accuracy, but it must also minimize the combinatorial explosion so as to respond as fast as a human would in a similar situation.
Dragon System
- The Dragon system is interesting for its simplicity of design, mathematical tractability and high accuracy performance, which can achieve 100% accuracy on a set of speech data using only syntactic and acoustic-phonetic knowledge;
- It achieves recognition by updating state probabilities of a Markov network. All possible paths are searched in parallel by a dynamic programming scheme and lead through states representing all syntactic and phonetic pronunciations;
- The most significant feature of the system is that it contains no coded heuristics except for the acoustic probability matching routine. It shows that there is a search algorithm which guarantees a recognition and in a deterministic amount of time.
Hearsay-I System | Dragon System | |
---|---|---|
Model | cooperating parallel process | probabilistic function of Markov Process. |
knowledge representation | Procedures. | Markov Networks. |
Unit of Speech
Signal Representation |
segmentation | 10 milli-second sample |
search strategy | best-first | all paths in parallel |
search time | 8-50 times real-time | 45-200 times real-time |
What Have Been Tested | 144 connected utterances
676 words 5 speakers 4 tasks: Chess, News Retrieval, Medical Diagnosis, Desk Calculator 28-76 word vocabularies |
Janice, if you can't find relevant info you can put N/A |
Performance | % of words spotted and identified correctly
89% with all the sources of knowledge 67% without semantics 44% without syntax and semantics (as of October, 1973) |
Janice, if you can't find relevant info you can put N/A |
Approaches | acoustics-phonetics, prosody, lexicon, syntax, semantics and pragmatics | Treats speech recognition as a mathematical computation problem
rather than as an artificial intelligence problem. |
Harpy Sytem
- detailed info of the current system
- with a table comparing three systems
The Harpy system attempts to combine the best features of the Hearsay-I system and the Dragon system. For instance, it uses a mathematically tractable model, as in the Dragon system, and speech-dependent heuristics, as in the Hearsay-I system. Its optimizations on language representation and search strategy have been shown effective in improving the model speed and accuracy. Moreover, it incorporates new techniques for further improvements
![An old rock in the desert](/assets/images/shiprock.jpg "Shiprock, New Mexico by Beau Rogers")(
https://www.google.com/search?sca_esv=566211836&sxsrf=AM9HkKn0aLPx4kaMULpyMsPvlbOO9Ug7Fw:1695026670732&q=harpy+system+speech&tbm=isch&source=lnms&sa=X&ved=2ahUKEwjI19-04rOBAxWG2qQKHRWCCk0Q0pQJegQIDBAB&cshid=1695026736590469&biw=1280&bih=571&dpr=1.5#imgrc=fCDs8_p8t6rcBM)
1. Model
The model of the Harpy system is a dynamic programming system with heuristics to reduce the search space, and thus increasing the speed. The system uses a network contains all possible synthetic paths and all the pronunciations of the paths. The synthetic and pronunciation knowledges are combined together in the ‘synthetic-phonetic’ paths of the network, which contains inter-state connections.
2. Representation of knowledge
The majority of the knowledge represented in the Harpy system is represented within the network, which is a set of states with inter-state connections. The network is generated by a network compiler from the BNF grammar (citation specification, the phonetic dictionary of the lexical words, and the inter-word juncture rules. In the phonetic dictionary, the Harpy system uses a speaker-dependent acoustic-phonetic templates to represent the acoustic realization of the phones. Thus, each state in the network contains the following information: the word (the terminal symbol from the BNF grammar), a unique ID number (every BNF terminal symbol with a unique symbol), the phone (from either the phonetic dictionary or the word juncture rules), a list of prior and following states (all the states that may transition into the current state and all the states to which the current state may transition)
2.1 Generation
The generation of the network in the following steps: the BNF grammar is used to generate a grammar network; each state in the grammar network, which represents a terminal symbol in the BNF grammar, is expanded with a phonetic subnetwork that represents the dictionary spelling for the terminal symbol; word juncture rules are added, and finally, special reduction heuristics are applied.
3. Data dependent transition probabilities
The Harpy system uses data-dependent transition probabilities. This is realized by heuristically calculate the transition probabilities for each state during the recognition process from speech-dependent knowledge, and only intrinsic phonemic durations in the Harpy system.
The system uses minimum and maximum expected durations for each phone and the current duration of the states to calculate the transition probabilities. The current duration for a state is the number of time samples for which the state was its own best prior state.
- The inter-state transition probability from a state is 1 if the current duration for that state is not less than the minimum expected duration for the phone of that state. If the current duration is less than the minimum expected, then the inter-state transition probability from that state is reduced by a heuristic amount that is proportional to the difference between the minimum expected and the actual current duration. If I is the minimum expected duration, C is the current duration (with C<I) then the inter-state transition probability is HI-C where H is a heuristic value. Similarly, if the current duration is not greater than the maximum expected duration, then the intra-state transition probability is 1, otherwise it is where A is the maximum expected duration and C and H are as before. The current value used for H is.1. A value of 0 for H would be tantamount to a total rejection of a path should the state duration fall outside the interval between the minimum and maximum expected durations. A large value for H(close to 1) and a small interval between the minimum and maximum expected durations would produce a Gaussian like distribution for the transition probabilities.
Condition | Transition probability |
The current duration ≥ the minimum expected duration | 1 |
The current duration < the minimum expected duration | HI-C |
The current duration ≤ the maximum expected duration | 1 |
The current duration > the maximum expected duration | HC-A |
4. Segmentation
The Harpy system segments the input data into units larger than 10 mili-seconds time samples. The segmentation algorithm used in the Harpy system works as follows: The 10 milli-second time samples are processed one at a time to extend the "current" segment. As each 10 milli-second time sample is processed, it is compared against the first 10 milli-second sample of the current segment and against the middle sample. The current segment is considered complete when the distance between the current 10 milli-second sample and either the first sample or the middle sample exceeds a heuristic threshold; the current 10 milli-second sample then becomes the 1st sample of the next segment. The linear predictor coefficients that are to represent the now complete current segment are obtained from the sum of the autocorrelation coefficients of all the samples in the segment.
5. Increasing speed and accuracy
5.1 Prior state information saving
The Harpy system optimized the prior state information saving strategy in Dragon system to reach higher recognition speed. Harpy saves a small number of prior states (100 times samples worth) in memory at a time and when more room is needed, they are saved on external storage (a high speed drum). For the algorithms, Harpy calculates the probabilities by logs and thus do addition, which eliminates the need to scale.
5.2 Search space reduction
The Harpy system uses heuristics to reduce the calculations and search spaces, and therefore increases the speed of recognition. The formula for the reaction time shown as below:
(rec.time in seconds)= (#time samples)(.022+.0004(#templates)+.00027(#states)+.00005(# pointers)
The formula shows the 4 main areas where Harpy applies heuristics: the number of time samples used, the number of templates used for the acoustic match probabilities per time sample, the number of states checked per time sample, and the number of pointers checked per time sample.
The methods used by Harpy to reduce all these number checked per time are based on the notion that not all network state probabilities need be updated every time sample. At time sample 0, there is probability 1 of being in the initial state and probability of being in all other states. The only states that have non-zero probability at time sample 1 are those states that can be reached in one transition from the initial state. Similarly, the only states that have non-zero probability at time ample N are those states that can be reached in N transitions (both inter and intra-state) from the initial state. At M time samples from the end of the utterance, the only states that need be checked are those from which the final state can be reached in M transitions. Unfortunately, this significantly reduces the number of states that need to be checked only near the start or end of the utterance. Intuitively, the only paths that need to be checked at every time sample are those that are "obvious".
The number of pointers checked at every time sample is dependent on the number of states checked.
The Harpy system uses ‘best-several’ search strategy to reduce the number of paths checked and guarantee the performance in large search space conditions. That is, Harpy searches for a few ‘best’ paths in parallel. ‘best’ is calculated by heuristics.
5.3 State Size reduction
The recognition time is dependent upon the confusability among the paths in the Harpy system. The developers removed null states, redundant states, and subsumed common states to reduce the network size and complexity. The former two removals have no effect on speech recognition accuracy, while the last removal has the potential influencing the recognition process.
- A null state is a state contains no synthetic or phonetic information other than the connections to its prior and following states. Harpy removes null states by linking each of it prior states to each of its following states and deleting the null states. However, the removal of null states increases the number of pointers.
- Two redundant states are 1) same terminal lexical symbol and same prior states or 2) same terminal lexical symbol and same following states. One of the states is selected and removed by Harpy.
- Two states are common if they both have the same phone and 1) same prior states or 2) same following states. Similarly, one state is selected and removed. To avoid danger of completely losing grammatical information due to subsumption of homonyms, the special character ‘!’ may use in the acoustic dictionary to indicate the network compiler.
Impact
The Harpy system was a dark horse during speech recognition development. It showed a greater performance than previous systems.
- It was the first machine that was capable of understanding 1,011 words (wiki link, which was a substantial improvement as previous model can recognize only less than 20 words.
- Harpy was also capable of recognizing complete sentences.
- Moreover, the optimization of the previous systems and newly incorporated heuristics both benefited the speed and accuracy of the recognition.
- The system also performed quite well for data with background noise.
Key Innovations
- how the optimization make the improvements
- implications
Future research
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum. Cite error: The opening <ref>
tag is malformed or has a bad name
ChatGPT
As for the research and development of Harpy System, ChatGPT perceived it as a non-existent concept. Perhaps the training data of ChatGPT does not extensively cover the history of speech recognition.
References
To insert a reference, type <ref> and paste the source your exported from Zotero (or whatever reference manager you're using) in the pop up box which appears. Make sure links in citations are clickable using proper formatting. Once you do this, a footnote will appear.Cite error: The opening <ref>
tag is malformed or has a bad name and a reference comes at the end automatically. Please use this method to cite for Wiki articles only, not for your thesis.
<ref>Raj Reddy. (2013). CMU Hearsay 1973 (Short Version). Retrieved September 17, 2023, from https://www.youtube.com/watch?v=c70QlwttnVg.
- ↑ Raj Reddy. (2013). CMU Hearsay 1973 (Short Version). Retrieved September 17, 2023, from https://www.youtube.com/watch?v=c70QlwttnVg.
- ↑ Raj Reddy. (2013). CMU Harpy System 1976 - Beam Search. Retrieved September 17, 2023, from https://www.youtube.com/watch?v=32KKg3aP3Vw.
- ↑ Lowerre, B., & Reddy, R. (1976). The harpy speech recognition system: performance with large vocabularies. The Journal of the Acoustical Society of America, 60(S1), S10-S11.