Gibbon et al., 2000 - Google Patents
Audio-visual and multimodal speech-based systemsGibbon et al., 2000
- Document ID
- 9477144308234280980
- Author
- Gibbon D
- Mertins I
- Moore R
- Publication year
- Publication venue
- Handbook of multimodal and spoken dialogue systems: Resources, terminology and product evaluation
External Links
Snippet
Communication between humans uses many modalities. We communicate not only via verbal language, but also through our use of intonation, gaze, hand gestures, body gestures, and facial expressions. Using these modalities, we can add, modify, and substitute …
- 230000001815 facial 0 abstract description 119
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/06—Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
- G10L21/10—Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids transforming into visible information
- G10L2021/105—Synthesis of the lips movements from speech, e.g. for talking heads
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/065—Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/001—Teaching or communicating with blind persons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/20—Handling natural language data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/66—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Benoit et al. | Audio-visual and multimodal speech systems | |
Brand | Voice puppetry | |
Gibbon et al. | Audio-visual and multimodal speech-based systems | |
Chen | Joint processing of audio-visual information for the recognition of emotional expressions in human-computer interaction | |
KR101604593B1 (en) | Method for modifying a representation based upon a user instruction | |
Hong et al. | Real-time speech-driven face animation with expressions using neural networks | |
Kipp | Gesture generation by imitation: From human behavior to computer character animation | |
Mattheyses et al. | Audiovisual speech synthesis: An overview of the state-of-the-art | |
Jaimes et al. | Multimodal human–computer interaction: A survey | |
Fanelli et al. | A 3-d audio-visual corpus of affective communication | |
Naert et al. | A survey on the animation of signing avatars: From sign representation to utterance synthesis | |
Delgado et al. | Spoken, multilingual and multimodal dialogue systems: development and assessment | |
Pelachaud et al. | Final report to NSF of the standards for facial animation workshop | |
DeCarlo et al. | Specifying and animating facial signals for discourse in embodied conversational agents | |
Thangthai et al. | Synthesising visual speech using dynamic visemes and deep learning architectures | |
Kacorri | TR-2015001: A survey and critique of facial expression synthesis in sign language animation | |
Madeo et al. | Studies in automated hand gesture analysis: an overview of functional types and gesture phases | |
Liu et al. | Real-time speech-driven animation of expressive talking faces | |
Khan | An Approach of Lip Synchronization With Facial Expression Rendering for an ECA | |
Cafaro et al. | Nonverbal behavior in multimodal performances | |
Chollet et al. | Multimodal human machine interactions in virtual and augmented reality | |
Fares | Multimodal Expressive Gesturing With Style | |
Huang et al. | Sensitive talking heads [Applications corner] | |
Cafaro et al. | Nonverbal Behavior in | |
Fanelli et al. | Acquisition of a 3d audio-visual corpus of affective speech |