The role of automated speech recognition in endoscopic data collection. Academic Article uri icon

Overview

abstract

  • Speech recognition technology has developed substantially in the past half decade. Currently, large vocabulary, speaker independent, discrete recognizers are the state-of-the-art. This will change. Moderate sized, continuous recognition systems now exist in research settings. However, it is unlikely that such systems will be widely available until the mid to late 1990's. The accuracy rates of current speech recognition systems are high. Consequently, speech accuracy is not the current limiting aspect of using ASR. The limiting aspect of using ASR technology is the approach to integrating speech functionality into applications. One approach is to use ATNs as models of natural language to support both an input strategy and a text generation system. ATNs provide approaches to both syntactical correctness and semantic richness. This is an approach which plays to the strengths of the discrete nature of current speech technology and also provides a methodology for the capture and archiving of highly detailed information. The ATN approach avoids the natural language parsing problem created by a fully free form dictation interface. Evolving along with the underlying speech technology are standards in the definitions and criteria used in endoscopic practice. There are clear benefits from standards in this area. However, it is likely that this will also take several years and may never yield a universally accepted lexicon. Furthermore, there will be user interface barriers to surmount in any system attempting to use speech as an input modality. Because of the relatively large vocabularies used in medical discourse, the user interface will need to be carefully crafted.(ABSTRACT TRUNCATED AT 250 WORDS)

publication date

  • July 1, 1992

Research

keywords

  • Data Collection
  • Endoscopy
  • Pattern Recognition, Automated
  • Speech

Identity

Scopus Document Identifier

  • 0026731866

Digital Object Identifier (DOI)

  • 10.1055/s-2007-1010528

PubMed ID

  • 1396387

Additional Document Info

volume

  • 24 Suppl 2