dblp: ICMI 2002 (original) (raw)



default search action
- combined dblp search
- author search
- venue search
- publication search
Authors:
- no matches

Venues:
- no matches

Publications:
- no matches


4th ICMI 2002: Pittsburgh, Pennsylvania, USA

jump to- Perceptual Computing Systems and Methods
- Speech Generation and Recognition
- Dialogue Processing and Language Understanding
- Signing, Gesturing, and Writing
- Translation and Multilingual Interfaces
- Gaze Tracking and Lipreading
- Applications of Multimodal Interfaces
- Sensors, Tools, and Platforms for Multimodal Interfaces
- User Studies and System Evaluation
- Face Detection, Recognition, and Synthesis

SPARQL queries 
Refine list

refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as

4th IEEE International Conference on Multimodal Interfaces (ICMI 2002), 14-16 October 2002, Pittsburgh, PA, USA. IEEE Computer Society 2002, ISBN 0-7695-1834-6
Perceptual Computing Systems and Methods

Nuria Oliver
, Eric Horvitz, Ashutosh Garg:
Layered Representations for Human Activity Recognition. 3-8
Christian Bauckhage
, Jannik Fritsch, Katharina J. Rohlfing, Sven Wachsmuth, Gerhard Sagerer:
Evaluating Integrated Speech- and Image Understanding. 9-14
Dan Maynes-Aminzade, Randy Pausch, Steven M. Seitz:
Techniques for Interactive Audience Participation. 15-20
Paulo Barthelmess, Clarence A. Ellis:
Perceptual Collaboration in Neem. 21-26
Ediz Polat, Mohammed Yeasin, Rajeev Sharma:
A Tracking Framework for Collaborative Human Computer Interaction. 27-32
Federico Fontana
, Davide Rocchesso
, Laura Ottaviani:
A Structural Approach to Distance Rendering in Personal Auditory Displays. 33-38
Andrea Fusiello, Antonello Panuccio, Vittorio Murino, Federico Fontana
, Davide Rocchesso
:
A Multimodal Electronic Travel Aid Device. 39-46
Speech Generation and Recognition

Ivica Rogina, Thomas Schaaf
:
Lecture and Presentation Tracking in an Intelligent Meeting Room. 47-52
Ryuta Taguma, Tatsuhiro Moriyama, Koji Iwano, Sadaoki Furui:
Parallel Computing-Based Architecture for Mixed-Initiative Spoken Dialogue. 53-58
Satoshi Nakamura, Panikos Heracleous:
3-D N-Best Search for Simultaneous Recognition of Distant-Talking Speech of Multiple Talkers. 59-63
Pui-Fung Wong, Man-Hung Siu:
Integration of Tone Related Feature for Chinese Speech Recognition. 64-68
Marc Mersiol, Noël Chateau, Valérie Maffiolo:
Talking Heads: Which Matching between Faces and Synthetic Voices?. 69-74
Ye Tian, Ji Wu, Zuoying Wang, Dajin Lu:
Robust Noisy Speech Recognition with Adaptive Frequency Bank Selection. 75-80
Ziqiang Wang, Yang Liu, Peng Ding, Bo Xu:
Covariance-Tied Clustering Method In Speaker Identification. 81-86
Dialogue Processing and Language Understanding

Joyce Y. Chai, Shimei Pan, Michelle X. Zhou, Keith Houck:
Context-Based Multimodal Input Understanding in Conversational Systems. 87-92
Helen Wright Hastie, Michael Johnston, Patrick Ehlen:
Context-Sensitive Help for Multimodal Dialogue. 93-98
Frédéric Landragin, Nadia Bellalem, Laurent Romary:
Referring to Objects with Spoken and Haptic Modalities. 99-104
Deb Roy:
Towards Visually-Grounded Spoken Language Acquisition. 105-110
Christian Elting, Gregor Möhler:
Modeling Output in the EMBASSI Multimodal Dialog System. 111-116
Aseel Ibrahim, Pontus Johansson:
Multimodal Dialogue Systems for Interactive TVApplications. 117-122
Candace L. Sidner, Myroslava O. Dzikovska:
Human - Robot Interaction: Engagement between Humans and Robots for Hosting Activities. 123-128
Jack Mostow, Joseph E. Beck, Raghu Chalasani, Andrew Cuneo, Peng Jia:
Viewing and Analyzing Multimodal Human-computer Tutorial Dialogue: A Database Approach. 129-134
Sorin Dusan, James L. Flanagan:
Adaptive Dialog Based upon Multimodal Language Acquisition. 135-140
Hartwig Holzapfel, Christian Fügen, Matthias Denecke, Alex Waibel:
Integrating Emotional Cues into a Framework for Dialogue Management. 141-148
Signing, Gesturing, and Writing

Haifeng Li, Thierry Artières
, Patrick Gallinari:
Data Driven Design of an ANN/HMM System for On-line Unconstrained Handwritten Character Recognition. 149-154
Lei Chen, Mary P. Harper, Francis K. H. Quek:
Gesture Patterns during Speech Repairs. 155-160
Sanshzar Kettebekov, Mohammed Yeasin, Rajeev Sharma:
Prosody Based Co-analysis for Continuous Recognition of Coverbal Gestures. 161-166
Aleix M. Martínez, Ronnie B. Wilbur, Robin Shay, Avinash C. Kak:
Purdue RVL-SLLL ASL Database for Automatic Recognition of American Sign Language. 167-172
Frédéric Landragin:
The Role of Gesture in Multimodal Referring Actions. 173-178
Yingen Xiong, Francis K. H. Quek, David McNeill:
Hand Gesture Symmetric Behavior Detection and Analysis in Natural Conversation. 179-184
Jose L. Hernandez-Rebollar, Robert W. Lindeman
, Nicholas Kyriakopoulos:
A Multi-Class Pattern Recognition System for Practical Finger Spelling Translation. 185-190
Andrea Corradini, Richard M. Wesson, Philip R. Cohen
:
A Map-Based System Using Speech and 3D Gestures for Pervasive Computing. 191-196
Edward Lin, Andy Cassidy, Dan Hook, Avinash Baliga, Tsuhan Chen:
Hand Tracking Using Spatial Gesture Modeling and Visual Feedback for a Virtual DJ System. 197-202
Haifeng Li, Thierry Artières
, Patrick Gallinari:
State Sharing in a Hybrid Neuro-Markovian On-Line Handwriting Recognition System through a Simple Hierarchical Clustering Algorithm. 203-210
Translation and Multilingual Interfaces

Ryosuke Isotani, Kiyoshi Yamabana, Shinichi Ando, Ken Hanazawa, Shin-ya Ishikawa, Tadashi Emori, Ken-ichi Iso, Hiroaki Hattori, Akitoshi Okumura, Takao Watanabe:
An Automatic Speech Translation System on PDAs for Travel Conversation. 211-216
Jing Zhang, Xilin Chen, Jie Yang, Alex Waibel:
A PDA-Based Sign Translator. 217-222
Loredana Taddei, Erica Costantini, Alon Lavie:
The NESPOLE! Multimodal Interface for Cross-lingual Communication - Experience and Lessons Learned . 223-228
Dequan Zheng, Yi Hu, Tiejun Zhao, Hao Yu, Sheng Li:
Research of Machine Learning Method for Specific Information Recognition on the Internet. 229-234
Erica Costantini, Fabio Pianesi, Susanne Burger:
The Added Value of Multimodality in the NESPOLE! Speech-to-Speech Translation System: an Experimental Study. 235-240
Shigeo Morishima
, Satoshi Nakamura:
Multi-Modal Translation System and Its Evaluation. 241-246
Zhirong Wang, Umut Topkara, Tanja Schultz
, Alex Waibel:
Towards Universal Speech Recognition. 247-252
Fei Huang, Stephan Vogel:
Improved Named Entity Translation and Bilingual Named Entity Extraction. 253-260
Gaze Tracking and Lipreading

Rowel Atienza, Alexander Zelinsky:
Active Gaze Tracking for Human-Robot Interaction. 261-266
David Demirdjian, Trevor Darrell:
3-D Articulated Pose Tracking for Untethered Diectic Reference. 267-272
Rainer Stiefelhagen:
Tracking Focus of Attention in Meetings. 273-280
Qiang Wang
, Haizhou Ai, Guangyou Xu:
A Probabilistic Dynamic Contour Model for Accurate and Robust Lip Tracking. 281-286
Chen Yu, Dana H. Ballard, Shenghuo Zhu:
Attentional Object Spotting by Integrating Multimodal Input. 287-292
Zhilin Wu, Petar S. Aleksic, Aggelos K. Katsaggelos
:
Lip Tracking for MPEG-4 Facial Animation. 293-298
Taeyoon Kim, Yongsung Kang, Hanseok Ko
:
Achieving Real-Time Lip Synch via SVM-Based Phoneme Classification and Lip Shape Refinement. 299-304
Satoshi Nakamura, Ken'ichi Kumatani, Satoshi Tamura:
Multi-Modal Temporal Asynchronicity Modeling by Product HMMs for Robust. 305-312
Applications of Multimodal Interfaces

Elena V. Zudilova, Peter M. A. Sloot, Robert G. Belleman:
A Multi-Modal Interface for an Interactive Simulated Vascular Reconstruction System. 313-318
Helen Petrie, Wendy Fisher, Ine Langer, Gerhard Weber
, Keith Gladstone, Cathy Rundle, Liesbeth Pyfers:
Universal Interfaces to Multimedia Documents. 319-324
Ali Zandifar, Ramani Duraiswami
, Antoine Chahine, Larry S. Davis:
A Video Based Interface to Textual Information for the Visually Impaired. 325-330
Rajarathinam Arangarasan, George N. Phillips Jr.:
Modular Approach of Multimodal Integration in a Virtual Environment. 331-336
Georg Niklfeld, Michael Pucher, Robert Finan, Wolfgang Eckhart:
Mobile Multi-Modal Data Services for GPRS Phones and Beyond. 337-342
Brad A. Myers, Robert G. Malkin, Michael Bett, Alex Waibel, Ben Bostwick, Robert C. Miller, Jie Yang, Matthias Denecke, Edgar Seemann, Jie Zhu, Choon Hong Peck, Dave Kong, Jeffrey Nichols, William L. Scherlis:
Flexi-Modal and Multi-Machine User Interfaces. 343-348
Nils Krahnstoever, Sanshzar Kettebekov, Mohammed Yeasin, Rajeev Sharma:
A Real-Time Framework for Natural Multimodal Interaction with Large Screen Displays. 349-354
Anoop K. Sinha, James A. Landay:
Embarking on Multimodal Interface Design. 355-360
Lars Bo Larsen, Morten Damm Jensen, Wisdom Kobby Vodzi:
Multi Modal User Interaction in an Automatic Pool Trainer. 361-366
Daniel P. Siewiorek, Asim Smailagic, Matthew Hornyak:
Multimodal Contextual Car-Driver Interface. 367-376
Sensors, Tools, and Platforms for Multimodal Interfaces

Jeffrey Nichols, Brad A. Myers, Thomas K. Harris, Roni Rosenfeld
, Stefanie Shriver, Michael Higgins, Joseph Hughes:
Requirements for Automatically Generating Multi-Modal Interfaces for Complex Appliances. 377-382
Huazhong Ning, Liang Wang, Weiming Hu, Tieniu Tan:
Articulated Model Based People Tracking Using Motion Models. 383-388
Kevin W. Wilson, Vibhav Rangarajan, Neal Checka, Trevor Darrell:
Audiovisual Arrays for Untethered Spoken Interfaces. 389-394
Sen Wang, Wei Wei Zhang, Yangsheng Wang:
Fingerprint Classification by Directional Fields. 395-399
Dirk Focken, Rainer Stiefelhagen:
Towards Vision-Based 3-D People Tracking in a Smart Room. 400-405
Helena M. Mentis, Geri Gay:
Using TouchPad Pressure to Detect Negative Affect. 406-410
Marc Erich Latoschik:
Designing Transition Networks for Multimodal VR-Interactions Using a Markup Language. 411-416
Tomoko Yonezawa, Kenji Mase:
Musically Expressive Doll in Face-to-Face Communication. 417-422
Xilin Chen, Jie Yang:
Towards Monitoring Human Activities Using an Omnidirectional Camera. 423-428
Weikai Xie, Yuanchun Shi, Guangyou Xu, Yanhua Mao:
Smart Platform - A Software Infrastructure for Smart Space (SISS). 429-436
User Studies and System Evaluation

Rob Gray, Hong Z. Tan
, J. Jay Young:
Do Multimodal Signals Need to Come from the Same Place? Crossmodal Attentional Links Between Proximal and Distal Surfaces. 437-441
Jan Kleindienst, Ladislav Serédi, Pekka Kapanen, Janne Bergman:
CATCH-2004 Multi-Modal Browser: Overview Description with Usability Analysis. 442-447
Philip R. Cohen
, Rachel Coulston, Kelly Krout:
Multimodal Interaction During Multiparty Dialogues: Initial Results. 448-453
Yasmine Arafa, Abe Mamdani:
Multi-Modal Embodied Agents Scripting. 454-459
Jason Williams, Georg Michelitsch, Gregor Möhler, Stefan Rapp:
A Methodology for Evaluating Multimodality in a Home Entertainment System. 460-465
Changseok Cho, Huichul Yang, Gerard Jounghyun Kim, Sung Ho Han:
Body-Based Interfaces. 466-472
Jérôme Grosjean, Jean-Marie Burkhardt
, Sabine Coquillart, Paul Richard:
Evaluation of the Command and Control Cube. 473-478
Ernesto Arroyo, Ted Selker, Alexandre Stouffs
:
Interruptions as Multimodal Outputs: Which are the Less Disruptive?. 479-482
Gregory Aist, Barry Kort, Rob Reilly, Jack Mostow, Rosalind W. Picard:
Experimentally Augmenting an Intelligent Tutoring System with Human-Supplied Capabilities: Adding Human-Provided Emotional Scaffolding to an Automated Reading Tutor that Listens. 483-490
Jeffrey F. Cohn, Karen L. Schmidt, Ralph Gross, Paul Ekman:
Individual Differences in Facial Expression: Stability over Time, Relation to Self-Reported Emotion, and Ability to Inform Person Identification. 491-498
Face Detection, Recognition, and Synthesis

Michael M. Cohen, Dominic W. Massaro, Rashid Clark:
Training a Talking Head. 499-504
Piero Cosi, Emanuela Magno Caldognetto, Giulio Perin, Claudio Zmarich:
Labial Coarticulation Modeling for Realistic Facial Animation. 505-510
Ziyou Xiong, Yunqiang Chen, Roy Wang, Thomas S. Huang:
Improved Information Maximization based Face and Facial Feature Detection from Real-time Video and Application in a Multi-Modal Person Identification System. 511-516
Dalong Jiang, Wen Gao, Zhiguo Li, Zhaoqi Wang:
Animating Arbitrary Topology 3D Facial Model Using the MPEG-4 FaceDefTables. 517-522
Wei Wang, Shiguang Shan
, Wen Gao, Bo Cao, Baocai Yin:
An Improved Active Shape Model for Face Alignment. 523-528
Beat Fasel:
Head-Pose Invariant Facial Expression Recognition Using Convolutional Neural Networks. 529-534
Wenjun Lao, Dehui Kong, Baocai Yin:
An Improved Algorithm for Hairstyle Dynamics. 535-540

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from
to the list of external document links (if available).
load links from unpaywall.org
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the
of the Internet Archive (if available).
load content from archive.org
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from
,
, and
to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from
and
to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from
.
load data from openalex.org
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
dblp was originally created in 1993 at:
since 2018, dblp has been operated and maintained by:







