Speech Recognition: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Samuel C. Smith
imported>Samuel C. Smith
Line 9: Line 9:
[[Image:Waveform_I_went_to_the_store_yesterday.jpg|thumb|Waveform]] [[Image:Spectrogram_I_went_to_the_store_yesterday.jpg|thumb|Spectrogram of "I went to the store yesterday."]] Writing systems are ancient, going back as far as the Sumerians of 6,000 years ago.  The phonograph, which allowed the analog recording and playback of speech, dates to 1877.  Speech recognition had to await the development of computer, however, due to multifarious problems with the recognition of speech.
[[Image:Waveform_I_went_to_the_store_yesterday.jpg|thumb|Waveform]] [[Image:Spectrogram_I_went_to_the_store_yesterday.jpg|thumb|Spectrogram of "I went to the store yesterday."]] Writing systems are ancient, going back as far as the Sumerians of 6,000 years ago.  The phonograph, which allowed the analog recording and playback of speech, dates to 1877.  Speech recognition had to await the development of computer, however, due to multifarious problems with the recognition of speech.


First, speech is not simply spoken text--in the same way that [[Miles Davis]] playing ''So What'' can hardly be captured by a note-for-note rendition as sheet music.  What humans understand as discrete words with clear boundaries are actually delivered as a continuous stream of sounds.  ''Iwenttothestoreyesterday'', rather than ''I went to the store yesterday''.  Words can also blend, with ''Whaddayawa?'' representing ''What do you want?''
First, speech is not simply spoken text--in the same way that Miles Davis playing ''So What'' can hardly be captured by a note-for-note rendition as sheet music.  What humans understand as discrete words with clear boundaries are actually delivered as a continuous stream of sounds.  ''Iwenttothestoreyesterday'', rather than ''I went to the store yesterday''.  Words can also blend, with ''Whaddayawa?'' representing ''What do you want?''


Second, there is no one-to-one correlation between the sounds and letters.  In English, there are slightly more than five vowels--''a'', ''e'', ''i'', ''o'', ''u'', and sometimes ''y''.  There are more than twenty different vowel sounds, though, and the exact count can vary.  The reverse problem also occurs, where more than one letter can represent a given sound.  The letter ''c'' can have the same sound as the letter ''k'' or as the letter ''s''.
Second, there is no one-to-one correlation between the sounds and letters.  In English, there are slightly more than five vowels--''a'', ''e'', ''i'', ''o'', ''u'', and sometimes ''y''.  There are more than twenty different vowel sounds, though, and the exact count can vary.  The reverse problem also occurs, where more than one letter can represent a given sound.  The letter ''c'' can have the same sound as the letter ''k'' or as the letter ''s''.
Line 21: Line 21:
Audrey worked by recognizing [[phoneme]]s, or individual sounds.  The phonemes were correlated to reference models of phonemes that were generated by training the recognizer.  Over the next two decades, researchers spent large amounts of time and money trying to improve upon this concept, with little success.  Computer hardware improved by leaps and bounds, speech synthesis improved steadily, and [[Noam Chomsky]]'s idea of generative grammar suggested that language could be analyzed programmatically.  None of this, however, seemed to improve the state of the art in speech recognition.
Audrey worked by recognizing [[phoneme]]s, or individual sounds.  The phonemes were correlated to reference models of phonemes that were generated by training the recognizer.  Over the next two decades, researchers spent large amounts of time and money trying to improve upon this concept, with little success.  Computer hardware improved by leaps and bounds, speech synthesis improved steadily, and [[Noam Chomsky]]'s idea of generative grammar suggested that language could be analyzed programmatically.  None of this, however, seemed to improve the state of the art in speech recognition.


In 1969, John R. Pierce wrote a forthright letter to the Journal of the Acoustical Society of America, where much of the research on speech recognition was published.  Pierce was one of the pioneers in satellite communications, and an executive vice president at Bell Labs, which was a leader in speech recognition research.  Pierce said everyone involved was wasting time and money.
In 1969, [[John R. Pierce]] wrote a forthright letter to the Journal of the Acoustical Society of America, where much of the research on speech recognition was published.  Pierce was one of the pioneers in satellite communications, and an executive vice president at Bell Labs, which was a leader in speech recognition research.  Pierce said everyone involved was wasting time and money.


<blockquote>It would be too simple to say that work in speech recognition is carried out simply because one can get money for it. . . .The attraction is perhaps similar to the attraction of schemes for turning water into gasoline, extracting gold from the sea, curing cancer, or going to the moon.  One doesn't attract thoughtlessly given dollars by means of schemes for cutting the cost of soap by 10%.  To sell suckers, one uses deceit and offers glamor.<ref> J.R. Pierce: Whither Speech Recognition.  Journal of the Acoustical Society of America.  '''46''', 1049-1051 (1969)</ref></blockquote>
<blockquote>It would be too simple to say that work in speech recognition is carried out simply because one can get money for it. . . .The attraction is perhaps similar to the attraction of schemes for turning water into gasoline, extracting gold from the sea, curing cancer, or going to the moon.  One doesn't attract thoughtlessly given dollars by means of schemes for cutting the cost of soap by 10%.  To sell suckers, one uses deceit and offers glamor.<ref> J.R. Pierce: Whither Speech Recognition.  Journal of the Acoustical Society of America.  '''46''', 1049-1051 (1969)</ref></blockquote>
Line 33: Line 33:
Throughout the 1970s and 1980s research continued.  By the 1980s, most researchers were using hidden Markov models, which are behind all contemporary speech recognizers.  In the latter part of the 1980s and in the 1990s, DARPA (the renamed ARPA) funded several initiatives.  The first initiative was similar to the previous challenge: the requirement was still a one-thousand word vocabulary, but this time a rigorous performance standard was devised.  This initiative produced systems that lowered the word error rate from ten percent to a few percent.  Additional initiatives have focused on improving algorithms and improving computational efficiency.<ref> Pieraccini, R. and Lubensky, D.: Spoken Language Communication with Machines: The Long and Winding Road from Research to Business.  In Innovations in Applied Artificial Intelligence/Lecture Notes in Artificial Intelligence 3533, pp. 6-15, M. Ali and F. Esposito (Eds.) 2005</ref>
Throughout the 1970s and 1980s research continued.  By the 1980s, most researchers were using hidden Markov models, which are behind all contemporary speech recognizers.  In the latter part of the 1980s and in the 1990s, DARPA (the renamed ARPA) funded several initiatives.  The first initiative was similar to the previous challenge: the requirement was still a one-thousand word vocabulary, but this time a rigorous performance standard was devised.  This initiative produced systems that lowered the word error rate from ten percent to a few percent.  Additional initiatives have focused on improving algorithms and improving computational efficiency.<ref> Pieraccini, R. and Lubensky, D.: Spoken Language Communication with Machines: The Long and Winding Road from Research to Business.  In Innovations in Applied Artificial Intelligence/Lecture Notes in Artificial Intelligence 3533, pp. 6-15, M. Ali and F. Esposito (Eds.) 2005</ref>


In 2001, Microsoft released a speech recognition system that worked with Office XP.  It neatly encapsulates how far the technology had come in fifty years, and what the limitations still were.  The system had to be trained to a specific user's voice, using the works of great authors that were provided, such as Edgar Allen Poe's ''Fall of the House of Usher'', and Bill Gates' ''The Way Forward''.  Even after training, the system was fragile enough that a warning was provided, "If you change the room in which you use Microsoft Speech Recognition and your accuracy drops, run the Microphone Wizard again."  On the plus side, the system did work in real time, and it did recognize connected speech.
In 2001, [[Microsoft]] released a speech recognition system that worked with Office XP.  It neatly encapsulates how far the technology had come in fifty years, and what the limitations still were.  The system had to be trained to a specific user's voice, using the works of great authors that were provided, such as Edgar Allen Poe's ''Fall of the House of Usher'', and Bill Gates' ''The Way Forward''.  Even after training, the system was fragile enough that a warning was provided, "If you change the room in which you use Microsoft Speech Recognition and your accuracy drops, run the Microphone Wizard again."  On the plus side, the system did work in real time, and it did recognize connected speech.


== Speech Recognition Today ==
== Speech Recognition Today ==

Revision as of 13:01, 17 August 2008

This article is developed but not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
Gallery [?]
 
This editable, developed Main Article is subject to a disclaimer.
Nuvola apps kbounce green.png
Nuvola apps kbounce green.png
This article is currently being developed as part of an Eduzendium student project. The project's homepage is at CZ:CIS 700 Special Topics 2008. One of the goals of the course is to provide students with insider experience in collaborative educational projects, and so you are warmly invited to join in here, or to leave comments on the discussion page. However, please refrain from removing this notice.
Besides, many other Eduzendium articles welcome your collaboration!



Speech Recognition is one of the main elements of natural language processing, or computer speech technology. Speech recognition is equivalent to taking dictation: converting speech into comprehensible data. This is a skill that is done seemingly without effort by humans, but requires formidable processing and algorithmic resources from computers.


History of Speech Recognition

Waveform
Spectrogram of "I went to the store yesterday."

Writing systems are ancient, going back as far as the Sumerians of 6,000 years ago. The phonograph, which allowed the analog recording and playback of speech, dates to 1877. Speech recognition had to await the development of computer, however, due to multifarious problems with the recognition of speech.

First, speech is not simply spoken text--in the same way that Miles Davis playing So What can hardly be captured by a note-for-note rendition as sheet music. What humans understand as discrete words with clear boundaries are actually delivered as a continuous stream of sounds. Iwenttothestoreyesterday, rather than I went to the store yesterday. Words can also blend, with Whaddayawa? representing What do you want?

Second, there is no one-to-one correlation between the sounds and letters. In English, there are slightly more than five vowels--a, e, i, o, u, and sometimes y. There are more than twenty different vowel sounds, though, and the exact count can vary. The reverse problem also occurs, where more than one letter can represent a given sound. The letter c can have the same sound as the letter k or as the letter s.

In addition, people who speak the same language do not make the same sounds. There are different dialects--the word 'water' could be pronounced watter, wadder, woader, wattah, and so on. Each person has a distinctive pitch when they speak--men typically having the lowest pitch, women and children have a higher pitch (though there is wide variation and overlap within each group.) Pronunciation is also colored by adjacent sounds, the speed at which the user is talking, and even by the user's health. Consider how pronunciation changes when a person has a cold.

Lastly, consider that not all sounds are meaningful speech. Regular speech is filled with interjections that do not have meaning: Oh, like, you know, well. There are also sounds that are a part of speech that are not considered words: er, um, uh. Coughing, sneezing, laughing, sobbing, even hiccupping can be a part of what is spoken. And the environment adds its own noises; speech recognition is difficult even for humans in noisy places.

Despite the manifold difficulties, speech recognition has been attempted for almost as long as there have been digital computers. As early as 1952, researchers at Bell Labs had developed an Automatic Digit Recognizer, or "Audrey". Audrey attained an accuracy of 97 to 99 percent if the speaker was male, and if the speaker paused 350 milliseconds between words, and if the speaker limited his vocabulary to the digits from one to nine, plus "oh", and if the machine could be adjusted to the speaker's speech profile. Results dipped as low as 60 percent if the recognizer was not adjusted.[1]

Audrey worked by recognizing phonemes, or individual sounds. The phonemes were correlated to reference models of phonemes that were generated by training the recognizer. Over the next two decades, researchers spent large amounts of time and money trying to improve upon this concept, with little success. Computer hardware improved by leaps and bounds, speech synthesis improved steadily, and Noam Chomsky's idea of generative grammar suggested that language could be analyzed programmatically. None of this, however, seemed to improve the state of the art in speech recognition.

In 1969, John R. Pierce wrote a forthright letter to the Journal of the Acoustical Society of America, where much of the research on speech recognition was published. Pierce was one of the pioneers in satellite communications, and an executive vice president at Bell Labs, which was a leader in speech recognition research. Pierce said everyone involved was wasting time and money.

It would be too simple to say that work in speech recognition is carried out simply because one can get money for it. . . .The attraction is perhaps similar to the attraction of schemes for turning water into gasoline, extracting gold from the sea, curing cancer, or going to the moon. One doesn't attract thoughtlessly given dollars by means of schemes for cutting the cost of soap by 10%. To sell suckers, one uses deceit and offers glamor.[2]

Pierce's 1969 letter marked the end of official research at Bell Labs for nearly a decade. The defense research agency ARPA, however, chose to persevere. In 1971 they sponsored a research initiative to develop a speech recognizer that could handle at least 1,000 words and understand connected speech, i.e., speech without clear pauses between each word. The recognizer could assume a low-background-noise environment, and it did not need to work in real time.

By 1976, three contractors had developed six systems. The most successful system, developed by Carnegie Mellon University, was called Harpy. Harpy was slow—a four-second sentence would have taken more than five minutes to process. It also still required speakers to 'train' it by speaking sentences to build up a reference model. Nonetheless, it did recognize a thousand-word vocabulary, and it did support connected speech.[3]

Research continued on several paths, but Harpy was the model for future success. It used hidden Markov models and statistical modeling to extract meaning from speech. In essence, speech was broken up into overlapping small chunks of sound, and probabilistic models inferred the most likely words or parts of words in each chunk, and then the same model was applied again to the aggregate of the overlapping chunks. The procedure is computationally intensive, but it has proven to be the most successful.

Throughout the 1970s and 1980s research continued. By the 1980s, most researchers were using hidden Markov models, which are behind all contemporary speech recognizers. In the latter part of the 1980s and in the 1990s, DARPA (the renamed ARPA) funded several initiatives. The first initiative was similar to the previous challenge: the requirement was still a one-thousand word vocabulary, but this time a rigorous performance standard was devised. This initiative produced systems that lowered the word error rate from ten percent to a few percent. Additional initiatives have focused on improving algorithms and improving computational efficiency.[4]

In 2001, Microsoft released a speech recognition system that worked with Office XP. It neatly encapsulates how far the technology had come in fifty years, and what the limitations still were. The system had to be trained to a specific user's voice, using the works of great authors that were provided, such as Edgar Allen Poe's Fall of the House of Usher, and Bill Gates' The Way Forward. Even after training, the system was fragile enough that a warning was provided, "If you change the room in which you use Microsoft Speech Recognition and your accuracy drops, run the Microphone Wizard again." On the plus side, the system did work in real time, and it did recognize connected speech.

Speech Recognition Today

Technology

Speech is derived from unique sounds created from the vocal chords of the human species. Through the constant exposure to speech during a child’s development a child is able to “learn” to understand similar sounding words from different people due to the phonetic similarities in the words. The mental capabilities of the human brain helps humans achieve this remarkable capability. So far we have only been able to reproduce this in computers on a limited basis. Current voice recognition technologies work on the ability to mathematically analyze the sound waves formed by our voices through resonance and spectrum analysis. Computer systems first record the sound waves spoken into a microphone through a digital to analog converter. The analog or continuous sound wave that we produce when we say a word is sliced up into small time fragments. From there these fragments are measured based on their amplitude levels, where the amplitude is described as the level of compression of air released from a person’s mouth. To measure the amplitudes and convert a sound wave to digital format the industry has commonly used the Nyquist-Shannon Theorem.[5]

Nyquist-Shannon Theorem
The Nyquist –Shannon theorem was developed in 1928 to show that a given analog frequency could be most accurately recreated by a digital frequency that is twice the original analog frequency. This is because, as Nyquist proved, an audible frequency must be sampled once for compression and once for rarefaction. For example, a 20 kHz audio signal can be accurately represented as a digital sample at 44.1 kHz. Interpreting Samples for Voice Recognition In speech recognition programs software will convert spoken instructions into digital samplings. These samplings will be measured against a stored database of recognized instructions. If the sample matches a stored instruction then the software executes a command. While this concept sounds simple enough, matching the sample with a stored instruction can be very difficult.

Recognizing Commands
The most important goal of current speech recognition software is to recognize commands. This increases the functionality of speech software. Software such as Sync is built into many new vehicles, supposedly allowing users to access all of the car’s electronic accessories, hands-free. This software has a small learning curve where it asks you a series of questions and based on the way you say some common words it is able to derive some constants to factor in its speech recognition algorithms and provide better recognition in the future. Current tech reviewers have said the technology is much improved from the early 1990’s but will not be replacing hand controls any time soon. [6]

Business

Major Speech Technology Companies

NICE Systems (NASDAQ: NICE and Tel Aviv: Nice), headquartered in Israel and founded in 1986, specialize in digital recording and archiving technologies. In 2007 they made $523 million in revenue in 2007. For more information visit http://www.nice.com.

Verint Systems Inc.(OTC:VRNT), headquartered in Melville, New York and founded in 1994 self-define themselves as “A leading provider of actionable intelligence solutions for workforce optimization, IP video, communications interception, and public safety.”[7] For more information visit http://verint.com.

Nuance (NASDAQ: NUAN) headquartered in Burlington, develops speech and image technologies for business and customer service uses. For more information visit http://www.nuance.com/.

Vlingo, headquartered in Cambridge, MA, develops speech recognition technology that interfaces with wireless/mobile technologies. Vlingo has recently teamed up with Yahoo! providing the speech recognition technology for Yahoo!’s mobile search service, oneSearch. For more information visit http://vlingo.com

Patent Infringement Lawsuits

NICE v. Verint

The technologies produced by both NICE and Verint are highly dependent on speech recognition. Both companies are highly competitive in the field of call/contact centers and security intelligence. This commonality is also found in their technologies resulting multiple patent infringement lawsuits. Since 2004 NICE and Verint have both filed multiple patent infringement lawsuits upon each other. They are discussed below.

  • January 2008: Patent infringement case brought by NICE against Verint had a deadlocked jury and is currently awaiting new trial.
  • May 19, 2008: It was announced that the US District Court for the Northern District of Georgia in favor of Verint in their infringement suit against Nice Systems, and further awarded Verint $3.3 million in damages.[8] The suit was over Verint’s patent No. 6,404,857 for a device which can monitor several telephone communications and “relates to emotion detection, word spotting, and talk-over detection in speech analytics solutions.”[9]
  • May 27, 2008: US District Court for the Northern District of Georgia found that the technology of Verint Systems Inc. did not infringe on the NICE patent No. 6,871,299. The technology in this patent deals with IP recording.
  • August 2008: Verint is currently awaiting a trial date for the patent infringement case they brought against NICE that involves their patented screen-capture technology. Verint stands to receive over $30 million in damages and an injunction prohibiting NICE from developing or selling any products with infringed technology.

And these lawsuits between NICE and Verint are not likely to end anytime soon, evident by Shlomo Shamir, NICE’s presidents, comment that “We believe that Verint’s patents asserted against NICE are invalid, and we intend to continue seeking invalidation of these patents.”[10]

Nuance v. vlingo

June 16, 2008: Nuance filed a claim in the United States District Court for the Eastern District of Texas against vlingo for patent infringement of Patent No. 6,766,295, “Adaptation of a Speech Recognition System across Multiple Remote Sessions with a Speaker”. The technology described in this patent is an “adaption of speech recognition system across multiple remote sessions with speaker (patent foot note),” which is essentially the result of storing multiple acoustical models for one speaker. If the suit is found in favor of Nuance, they stand to receive damages and an injunction against the use of their technologies. [11]

Klausner Technologies v. AT&T and Apple

June 16 2008: Klausner Technologies settled the infringement lawsuit they brought against AT&T and Apple over patents for Klausner’s visual voicemail. AT&T and Apple paid Klausner an undisclosed sum and agreed to license the technology from Klausner.

Voice User Interface

The design cycle of a voice user interface (VUI) parallels that of the graphical user interface(GUI), both following the standard software development life cycle (SDLC). Although SDLC varies slightly based on the developer and the company, it is generally agreed upon as (1) identification of needs (2) gathering of requirements (3) prototyping (4) detailed design (2) programming (3) product usability testing (4) deployment (5) maintenance.[12][13] The VUI cycle must however incorporate input from a much more diverse group of specialist than the UI cycle, including such people as speech scientists, linguists, human factors experts and business experts. Once a business has adequately articulated their needs to software developers (step 1 of SDLC), developers must learn about their user (step 2 of SDLC). To do this, designers identify personas, model users who encompass the characteristics, personalities and skill levels of potential users. Designing VUIs based on personas helps to ensure that speech prompts are appropriate and understandable to users in addition to helping determine proper dialog style. Dialog style defines the way a user can respond to a speech application. The four main types of dialog styles include (1) directed dialog where the user is given a number of choices and must respond with one of those options (2) natural language dialog where the user is given free range to say what they would like (3) mixed initiative applications which use a more confined type of natural language and defer to directed dialog if the user is not understood and (4) Form Filling in which the user answers a series of straight forward questions as they would if they were filling out a form.[14]


Designers also must carefully decide how to create prompts within the VUI. The two main mechanisms are (1) text-to-speech (TTS) and (2) audio recordings. Typical tools break prompts down into words and phrases which can then be combined during application runtime. The algorithms used within the speech application determine the word and phrase combination.


Two main forms of speech recognition that have been used in the past are (1) constrained grammars which contain a large bank of word that the application can identify and (2) scripted interactions which take you through a step by step process, and which each step moves to word bank more directed at the user needs (ex: 411 call moving from “city and state” bank to “Philadelphia, Pennsylvania bank).[15] To reduce the constriction placed on users, many applications have implemented Statistical Language Models (SLMs) in which the large grammars used in the previous techniques are replaced by the statistical “averaging of a large corpus of sample responses recorded live from users responding to a prompt.”[16]


The building of these VUI involves extensive time and money. As a result a result the speech industry has answered back with a number of application templates. Companies such as ScanSoft, Vialto, Nuance and Unisys provide templates for such things as 411 directory assistance, voice-activated dialing, corporate greeter and a number of other automated call center applications.


It is crucial for developers to test their product throughout the design process including after prototyping, initial development and product completion, making usability testing an iterative process in which designers test, revise and test again. Usability testing is crucial for developers to learn from their users, to discover short-comings within products and to learn how to make their products better. Two common VUI testing techniques are (1) Wizard of Oz (WOZ) testing and (2) prototype testing. WOZ is described as “a human tester plays the recorded prompts to a panel of around 100 people who are asked to navigate the prompts and complete a particular task.” This technique focuses on the understandability and effectiveness of the VUI prompts. Other VUI usability testing techniques include statistical benchmarks, Holistic Usability Measure developed by Juan Gilbert where clients distribute 100 points across characteristics the developer deems important (i.e. accuracy, understandability and efficiency), and Melanie Polkosky’s statistical measure known as factor analysis that incorporates user- goal orientation, speech characteristics, verbosity and customer service behavior. [17]


It is important to note that there are many ways to develop UIs and VUIs other than those described here, however the techniques above describe some of the most “common” methodologies involved in VUI design.

The Future of Speech Recognition

Emerging Technologies

Mobile Search Applications

In recent years there has been a steady movement towards the development of speech technologies to replace or enhance text input applications. We see this trend in products such as audio search engines, voicemail to text programs, dictation programs and desktop “say what you see” commands..[18] Recently both Yahoo! and Microsoft have launched voice-based mobile search applications. The concept behind Yahoo! oneSearch and Microsoft Tellme are very similar, it is their implementation and the speech technology used in their applications which differ. With both products users speak a “search word” into their mobile phone or PDA while holding down the green talk button, the request is sent to a server that analyzes the sound bit and then the results of search appear on the mobile device.[19] OneSearch is currently available for select Blackberry users in the US and can be downloaded from Yahoo!. Tellme is available for Blackberry users in the US by download from Microsoft and pre-installed on all Helio’s Mystos.

Yahoo! has partnered with vlingo for the speech recognition feature of oneSearch. This voice technology allows users to say exactly what they want, they do not need to know or use special commands, speak slowly or overly articulately. Vlingo’s technology implements Adaptive Hierarchical Language Models (A-HLMs) which allows oneSearch to use regional details and user patters to adapt to the characteristics of its surroundings including word pronunciation, accents and acoustics.[20]

Microsoft’s subsidiary Tellme took a different approach to the speech recognition element of their mobile search application. Users are instructed to say specific phrases such as “traffic,” “map” or the name of business. Tellme’s senior product manager David Mitby, explains why they chose to limit speech parameters: "[because] very solid smartphone users pick this up for the first time, it’s not clear what to do. It’s not intuitive media yet."[21]

Future Trends & Applications

The Medical Industry
For years the medical industry has been touting electronic medical records (EMR). Unfortunately the industry has been slow to adopt EMRs and some companies are betting that the reason is because of data entry. There isn’t enough people to enter the multitude of current patient’s data into electronic format and because of that the paper record prevails. A company called Nuance (also featured in other areas here, and developer of the software called Dragon Dictate) is betting that they can find a market selling their voice recognition software to physicians who would rather speak patients data than handwrite all medical information into a person’s file. [22]

References

  1. K.H. Davis, R. Biddulph, S. Balashek: Automatic recognition of spoken digits. Journal of the Acoustical Society of America. 24, 637-642 (1952)
  2. J.R. Pierce: Whither Speech Recognition. Journal of the Acoustical Society of America. 46, 1049-1051 (1969)
  3. Robert D. Rodman. Computer Speech Technology. Massachusetts: Artech House 1999
  4. Pieraccini, R. and Lubensky, D.: Spoken Language Communication with Machines: The Long and Winding Road from Research to Business. In Innovations in Applied Artificial Intelligence/Lecture Notes in Artificial Intelligence 3533, pp. 6-15, M. Ali and F. Esposito (Eds.) 2005
  5. Jurafsy, M. and Martin, J. An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition. New Jersey: Prentice Hall 2006
  6. http://etech.eweek.com/content/enterprise_applications/recognizing_speech_recognition.html
  7. see "About Verint"
  8. “Verint Wins Patent Lawsuit Against Nice Systems on Speech Analytics Technology,” Verint May 19, 2008 < http://www.swpp.org/pressrelease/2008/0519VerintSpeechAnalyticsPatentRelease.pdf>
  9. L. Klie, “Speech Fills the Docket,” Speech Technology, vol. 13, no. 6, July/August 2008
  10. L. Klie, “Speech Fills the Docket,” Speech Technology, vol. 13, no. 6, July/August 2008
  11. “Nuance Asserts Intellectual Property in Speech Adaptation Domain, Files Patent Infringement,” Reuters, June 16, 2008 <http://www.reuters.com/article/pressRelease/idUS194736+16-Jun-2008+BW20080616>
  12. B. Scholz, “Building Speech Applications: Part One – The Application and its Development Infrastructure,” Speech Technology, March/April 2004, accessed 8/9/2008, < http://www.speechtechmag.com/Articles/Editorial/Feature/Building-Speech-Applications-Part-One---The-Application-and-its-Development-Infrastructure-29437.aspx>
  13. “SDLC Resources,” MKS, accessed 8/17/2008, <http://www.mks.com/sdlc>
  14. >B. Scholz, “Building Speech Applications: Part One – The Application and its Development Infrastructure,”
  15. “Revolutionizing Voice UI for Mobil” vlingo. May 2008, accessed 8/9/2008, <http://vlingo.com/pdf/Vlingo%20Voice%20Recognition%20White%20Paper%20-%20May%202008v2.pdf>
  16. Ibid.
  17. M. Polkosky, “What is Speech Usability Anyway?,” Speech Technology, Nov. 7, 2005, accessed August 17, 2008, < http://www.speechtechmag.com/Articles/Editorial/Feature/What-Is-Speech-Usability-Anyway-29601.aspx>
  18. Microsoft provides this feature on Vista operating systems. To learn more go to http://www.microsoft.com/enable/products/windowsvista/speech.aspx
  19. S.H. Wildstrom, “LOOK MA, NO-HANDS SEARCH; Voice-based, mobile search from Microsoft and Yahoo! are imperfect but promising,” BuisnessWeek, McGraw-Hill, Inc. vol. 4088, June 16, 2008.
  20. E. Keleher and B. Monaghan, “Vlingo Introduces Localized Voice-Recognition Support for Yahoo! oneSearch,” vlingo, June 17, 2008, http://vlingo.com/pdf/vlingo%20Introduces%20Localized%20Voice-rec%20for%20Yahoo%20oneSearch%20FINAL.pdf, accessed: August 8, 2008.
  21. R. Joe, “Multiple-Modality Disorder,” Speech Techology, vol 13 no. 5, June 2008.
  22. http://www.1450.com/speech_enable_emr.pdf