Carnegie Mellon's Harpy System

From MSc Voice Technology
Revision as of 22:10, 18 September 2023 by S5721415 (talk | contribs) (→‎References)
Jump to navigation Jump to search

Sherry Yu-Ting

Wansu Zhu

Jingsi Huang

Wenjun Meng

Carnegie Mellon's Harpy System

Jump to navigationJump to search

Introduction

Harpy System is a result of combing and improving the best features from two speech systems previously developed at Carnegie Mellon University: Hearsay-I ( Erman 1974 ) and Dragon ( Baker 1975 ). Two feature of the Harpy system that led to a successful demonstration are its representation of knowledge and the use of new search techniques. In developing speech recognition systems, it is necessary to devise a means of acquiring and representing the many diverse types of knowledge that characterize speech. We must also develop matching and searching techniques that convert this passive knowledge into an active process for understanding the utterance in the presence of error, noise, and uncertainty.


When an utterance is input to the computer. Several pre-processing steps are taken to prepare the data for recognition. The utterance is segmented into acoustic units and analyzed to determine the segmented features and parameters. The goal of recognition tasks is to find an optimal sequence of phones satisfying two criteria: the sequence must represent a legal path through the knowledge network and should consist of phones with high acoustic matches.


Harpy Systems uses a beam search to locate this optimal sequence of phones. Harpy begins the beam search by taking all legal phones from the start of the sentence from the knowledge network and entering them in a recognition tree. Next, a path probability is calculated for each candidate. The path with the best probability is determined and remaining candidates are compared with it. Those that fall below a threshold of acceptability are eliminated from further searching. The successors of the surviving candidates are expanded based on the information in the knowledge network.

Hearsay-I System Dragon System Harpy System
Sources of Knowledge
  • Acoustic-Phonetics (the characteristics of the sounds)
  • Prosodics (the stress and intonation patterns of speech)
  • Lexicon (a dictionary of legal words)
  • Syntax (the grammatical structure of the language)
  • Semantics (the meaning of words and sentences)
  • Pragmatics (the context of the conversation)
  • Almost no speech dependent heuristics.
  • The integrated knowledge of both syntax and lexical spellings, a set of user dependent acoustic-phonetic templates, and an acoustic-phonetic probability matching routine.
  • Mathematical acceptability available ---one of probabilistic function of Markov Process.
  • The combination of speech dependent heuristics ( Hearsay-I System ) and mathematical tractable model ( Dragon System ). This leverages the use of heuristics to reduce the search space therefore speed increase.
  • The system extensively use a network that represents both all legal syntactic paths and pronunciations of these legal paths.
  • No a-priori transition probabilities.
Knowledge Representation Procedural embedding. Markove networks. Transition networks.
Search Strategy Best-First with backtracking. All paths in parallel ( search all the possible syntactic

and acoustic paths through the network in parallel to determine optimal path ) with no backtracking.

''best few'' in parallel with no backtracking.

Beam Search

  • Locates optimal solution
  • Examines near-miss alternatives
  • No backtracking
Segmentation Yes

( uses it to reduce effective utterance length )

No Yes
Strength
  • Segmentation of the acoustic signal can effectively reduce the amount of speech data searched.
  • Heuristic speech knowledge serves as a practical guide to mapping and rating words.
  • The network data representation of combined knowledge of syntax and the phonetic dictionary spellings is tractable.
  • The dynamic programming scheme for searching all the network
  • paths in parallel guarantees both a recognition will always be achieved and that the recognition will be the globally optimum one. ( with the given model )
N/A
Weakness
  • Best-first strategy involves backtracking, which can be costly in large search space.
  • Searching all paths ( even in parallel ) is time consuming.
N/A
  • simple intro of the relationship between hearsay-I, Dragon system, and harpy system
  • general info about how harpy system works (features, methods
  • simply mention the advatages of harpy sytem

Historical Context

Before 1970s, the speech recognition systems available in the world could only understand 16 words spoken in English. With the goal of ending up with a machine capable of understanding a minimum of 1,000 words, speech recognition technology made major strides in the 1970s, thanks to interest and funding from the U.S. Department of Defense. The DoD’s DARPA Speech Understanding Research (SUR) program, from 1971 to 1976, funds five years of speech recognition research leading to the creation of the Harpy by Carnegie Mellon, a machine capable of understanding 1,011 words, approximately the vocabulary of an average three-year-old.

The Harpy connected speech recognition system is the result of an attempt to understand the relative importance of various design choices of two earlier speech recognition systems developed at Carnegie-Mellon University: the Hearsay-I system and the Dragon system. Systematic performance analysis of various design choices of these two systems resulted in the Harpy system.

Hearsay-I System

  • The Hearsay-I System must use all these sources of knowledge effectively in the Hearsay System, this is achieved by representing knowledge as as a set of cooperating parallel process. The Hearsay System also use the source of knowledge (acoustics-phonetics, prosody, lexicon, syntax, semantics and pragmatics) to generate hypothesis about what words might appear. The Hearsay System is incrementally trying to match words to the un-analyzed portion of the utterance
  • If the resulting word matches are not optimal, the system backs up and tries a different path. The technique was known as probabilistic tree. The Hearsay System cannot recognize sentences in unrestricted English and can only recognize simple phrase structure languages.
  • To be acceptable as an input medium, the system must not only perform with high accuracy, but it must also minimize the combinatorial explosion so as to respond as fast as a human would in a similar situation.

Dragon System

  • The Dragon system is interesting for its simplicity of design, mathematical tractability and high accuracy performance, which can achieve 100% accuracy on a set of speech data using only syntactic and acoustic-phonetic knowledge;
  • It achieves recognition by updating state probabilities of a Markov network. All possible paths are searched in parallel by a dynamic programming scheme and lead through states representing all syntactic and phonetic pronunciations;
  • The most significant feature of the system is that it contains no coded heuristics except for the acoustic probability matching routine. It shows that there is a search algorithm which guarantees a recognition and in a deterministic amount of time.
Hearsay-I System Dragon System
Model cooperating parallel process probabilistic function of Markov Process
knowledge representation Procedures Markov Networks
Unit of Speech

Signal Representation

segmentation 10 milli-second sample
search strategy best-first all paths in parallel
search time 8-50 times real-time 45-200 times real-time
What Have Been Tested 144 connected utterances

676 words

5 speakers

4 tasks: Chess, News Retrieval, Medical Diagnosis, Desk Calculator

28-76 word vocabularies

102 utterances

564 words

Performance % of words spotted and identified correctly

89% with all the sources of knowledge

67% without semantics

44% without syntax and semantics

(as of October, 1973)

85% of the words identified correctly
Approaches acoustics-phonetics, prosody, lexicon, syntax, semantics and pragmatics Treats speech recognition as a mathematical computation problem

rather than as an artificial intelligence problem.

Harpy Sytem

  • detailed info of the current system
  • with a table comparing three systems

The Harpy system attempts to combine the best features of the Hearsay-I system and the Dragon system. For instance, it uses a mathematically tractable model, as in the Dragon system, and speech-dependent heuristics, as in the Hearsay-I system. Its optimizations on language representation and search strategy have been shown effective in improving the model speed and accuracy. Moreover, it incorporates new techniques for further improvements

![An old rock in the desert](/assets/images/shiprock.jpg "Shiprock, New Mexico by Beau Rogers")(https://www.google.com/search?sca_esv=566211836&sxsrf=AM9HkKn0aLPx4kaMULpyMsPvlbOO9Ug7Fw:1695026670732&q=harpy+system+speech&tbm=isch&source=lnms&sa=X&ved=2ahUKEwjI19-04rOBAxWG2qQKHRWCCk0Q0pQJegQIDBAB&cshid=1695026736590469&biw=1280&bih=571&dpr=1.5#imgrc=fCDs8_p8t6rcBM)



1.    Model

The model of the Harpy system is a dynamic programming system with heuristics to reduce the search space, and thus increasing the speed. The system uses a network contains all possible synthetic paths and all the pronunciations of the paths. The synthetic and pronunciation knowledges are combined together in the ‘synthetic-phonetic’ paths of the network, which contains inter-state connections.


2.    Representation of knowledge

The majority of the knowledge represented in the Harpy system is represented within the network, which is a set of states with inter-state connections. The network is generated by a network compiler from the BNF grammar (citation specification, the phonetic dictionary of the lexical words, and the inter-word juncture rules. In the phonetic dictionary, the Harpy system uses a speaker-dependent acoustic-phonetic templates to represent the acoustic realization of the phones. Thus, each state in the network contains the following information: the word (the terminal symbol from the BNF grammar), a unique ID number (every BNF terminal symbol with a unique symbol), the phone (from either the phonetic dictionary or the word juncture rules), a list of prior and following states (all the states that may transition into the current state and all the states to which the current state may transition)


2.1 Generation

The generation of the network in the following steps: the BNF grammar is used to generate a grammar network; each state in the grammar network, which represents a terminal symbol in the BNF grammar, is expanded with a phonetic subnetwork that represents the dictionary spelling for the terminal symbol; word juncture rules are added, and finally, special reduction heuristics are applied.


3.    Data dependent transition probabilities

The Harpy system uses data-dependent transition probabilities. This is realized by heuristically calculate the transition probabilities for each state during the recognition process from speech-dependent knowledge, and only intrinsic phonemic durations in the Harpy system.

The system uses minimum and maximum expected durations for each phone and the current duration of the states to calculate the transition probabilities. The current duration for a state is the number of time samples for which the state was its own best prior state.

  • The inter-state transition probability from a state is 1 if the current duration for that state is not less than the minimum expected duration for the phone of that state. If the current duration is less than the minimum expected, then the inter-state transition probability from that state is reduced by a heuristic amount that is proportional to the difference between the minimum expected and the actual current duration. If I is the minimum expected duration, C is the current duration (with C<I) then the inter-state transition probability is HI-C where H is a heuristic value. Similarly, if the current duration is not greater than the maximum expected duration, then the intra-state transition probability is 1, otherwise it is   where A is the maximum expected duration and C and H are as before. The current value used for H is.1. A value of 0 for H would be tantamount to a total rejection of a path should the state duration fall outside the interval between the minimum and maximum expected durations. A large value for H(close to 1) and a small interval between the minimum and maximum expected durations would produce a Gaussian like distribution for the transition probabilities.
Condition Transition probability
The current duration ≥ the minimum expected duration 1
The current duration < the minimum expected duration HI-C
The current duration ≤ the maximum expected duration 1
The current duration > the maximum expected duration HC-A


4.    Segmentation

The Harpy system segments the input data into units larger than 10 mili-seconds time samples. The segmentation algorithm used in the Harpy system works as follows: The 10 milli-second time samples are processed one at a time to extend the "current" segment. As each 10 milli-second time sample is processed, it is compared against the first 10 milli-second sample of the current segment and against the middle sample. The current segment is considered complete when the distance between the current 10 milli-second sample and either the first sample or the middle sample exceeds a heuristic threshold; the current 10 milli-second sample then becomes the 1st sample of the next segment. The linear predictor coefficients that are to represent the now complete current segment are obtained from the sum of the autocorrelation coefficients of all the samples in the segment.


5.    Increasing speed and accuracy

5.1 Prior state information saving

The Harpy system optimized the prior state information saving strategy in Dragon system to reach higher recognition speed. Harpy saves a small number of prior states (100 times samples worth) in memory at a time and when more room is needed, they are saved on external storage (a high speed drum). For the algorithms, Harpy calculates the probabilities by logs and thus do addition, which eliminates the need to scale.

5.2 Search space reduction

The Harpy system uses heuristics to reduce the calculations and search spaces, and therefore increases the speed of recognition. The formula for the reaction time shown as below:

(rec.time in seconds)= (#time samples)(.022+.0004(#templates)+.00027(#states)+.00005(# pointers)

The formula shows the 4 main areas where Harpy applies heuristics: the number of time samples used, the number of templates used for the acoustic match probabilities per time sample, the number of states checked per time sample, and the number of pointers checked per time sample.


The methods used by Harpy to reduce all these number checked per time are based on the notion that not all network state probabilities need be updated every time sample. At time sample 0, there is probability 1 of being in the initial state and probability of being in all other states. The only states that have non-zero probability at time sample 1 are those states that can be reached in one transition from the initial state. Similarly, the only states that have non-zero probability at time ample N are those states that can be reached in N transitions (both inter and intra-state) from the initial state. At M time samples from the end of the utterance, the only states that need be checked are those from which the final state can be reached in M transitions. Unfortunately, this significantly reduces the number of states that need to be checked only near the start or end of the utterance. Intuitively, the only paths that need to be checked at every time sample are those that are "obvious".


The number of pointers checked at every time sample is dependent on the number of states checked.


The Harpy system uses ‘best-several’ search strategy to reduce the number of paths checked and guarantee the performance in large search space conditions. That is, Harpy searches for a few ‘best’ paths in parallel. ‘best’ is calculated by heuristics.


5.3 State Size reduction

The recognition time is dependent upon the confusability among the paths in the Harpy system. The developers removed null states, redundant states, and subsumed common states to reduce the network size and complexity. The former two removals have no effect on speech recognition accuracy, while the last removal has the potential influencing the recognition process.

  • A null state is a state contains no synthetic or phonetic information other than the connections to its prior and following states. Harpy removes null states by linking each of it prior states to each of its following states and deleting the null states. However, the removal of null states increases the number of pointers.
  • Two redundant states are 1) same terminal lexical symbol and same prior states or 2) same terminal lexical symbol and same following states. One of the states is selected and removed by Harpy.
  • Two states are common if they both have the same phone and 1) same prior states or 2) same following states. Similarly, one state is selected and removed. To avoid danger of completely losing grammatical information due to subsumption of homonyms, the special character ‘!’ may use in the acoustic dictionary to indicate the network compiler.

Impact

The Harpy system was a dark horse during speech recognition development. It showed a greater performance than previous systems.

  • It was the first machine that was capable of understanding 1,011 words (wiki link, which was a substantial improvement as previous model can recognize only less than 20 words.
  • Harpy was also capable of recognizing complete sentences.
  • Moreover, the optimization of the previous systems and newly incorporated heuristics both benefited the speed and accuracy of the recognition.
  • The system also performed quite well for data with background noise.

Key Innovations

  • how the optimization make the improvements
  • implications

Future research

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum. Cite error: The opening <ref> tag is malformed or has a bad name

ChatGPT

As for the research and development of Harpy System, ChatGPT perceived it as a non-existent concept. Perhaps the training data of ChatGPT does not extensively cover the history of speech recognition.

References

To insert a reference, type <ref> and paste the source your exported from Zotero (or whatever reference manager you're using) in the pop up box which appears. Make sure links in citations are clickable using proper formatting. Once you do this, a footnote will appear.Cite error: The opening <ref> tag is malformed or has a bad name and a reference comes at the end automatically. Please use this method to cite for Wiki articles only, not for your thesis.

[1][2][3]>

<ref>Raj Reddy. (2013). CMU Hearsay 1973 (Short Version). Retrieved September 17, 2023, from https://www.youtube.com/watch?v=c70QlwttnVg.

[4] [5]

  1. Pinola, M. (2011). Speech Recognition Through the Decades: How We Ended Up With Siri. https://www.pcworld.com/article/477914/speech_recognition_through_the_decades_how_we_ended_up_with_siri.html
  2. Lowerre, B. T. (1974). Comparison of Two Speech Understanding Systems. The Journal of the Acoustical Society of America, 56, S27. https://pubs.aip.org/asa/jasa/article/56/S1/S27/740227/Comparison-of-two-speech-understanding-systems
  3. Lowerre, B. T. (1976). The HARPY Speech Recognition System. Department of Computer Science, Carnegie-Mellon University.
  4. Raj Reddy. (2013). CMU Harpy System 1976 - Beam Search. Retrieved September 17, 2023, from https://www.youtube.com/watch?v=32KKg3aP3Vw.
  5. Lowerre, B., & Reddy, R. (1976). The harpy speech recognition system: performance with large vocabularies. The Journal of the Acoustical Society of America, 60(S1), S10-S11.