Speech recognition for command entry in multimodal interaction
โ Scribed by DAVID A TYFA; MARK HOWES
- Publisher
- Elsevier Science
- Year
- 2000
- Tongue
- English
- Weight
- 279 KB
- Volume
- 52
- Category
- Article
- ISSN
- 1071-5819
No coin nor oath required. For personal study only.
โฆ Synopsis
Two experiments investigated the cognitive e$ciency of using speech recognition in combination with the mouse and keyboard for a range of word processing tasks. The "rst experiment examined the potential of this multimodal combination to increase performance by engaging concurrent multiple resources. Speech and mouse responses were compared when using menu and direct (toolbar icon) commands, making for a fairer comparison than in previous research which has been biased against the mouse. Only a limited basis for concurrent resource use was found, with speech leading to poorer task performance with both command types. Task completion times were faster with direct commands for both speech and mouse responses, and direct commands were preferred. In the second experiment, participants were free to choose command type, and nearly always chose to use direct commands with both response modes. Speech performance was again worse than mouse, except for tasks which involved a large amount of hand and eye movement, or where direct speech was used but mouse commands were made via menus. In both experiments recognition errors were low, and although they had some detrimental e!ect on speech use, problems in combining speech and manual modes were highlighted. Potential verbal interference e!ects when using speech are discussed.
2000 Academic Press
๐ SIMILAR VOLUMES
We propose a general-purpose color-based object model called the Multimodal Neighborhood Signature (MNS) with applications in object recognition and image retrieval. Object modeling is example-based and can cope with many appearance variations due to the image formation/rendering process. The local
In this paper, by using the formulation of the missing-data problem, a general framework for statistical acoustic modelling of speech is presented. With the motivation of utilizing bi-directional contextual dependence in acoustic modelling, a bi-directional hidden Markov modelling approach for speec