Eye movement of perceivers during audiovisualspeech perception (original) (raw)
References
Abry, C, Lallouache, M.-T., &Cathiard, M.-A. (1996). How can coarticulation models account for speech sensitivity to audio-visual desynchronization? In D. Stork & M. Hennecke (Eds.),Speechreading by humans and machines (NATO-ASI SeriesF: Computer and Systems Sciences, Vol. 150, pp. 247–256). Berlin: Springer-Verlag. Google Scholar
Benoît, C, Lallouache, [M.-] T., Mohamadi, T., &Abry, C. (1992). A set of French visemes for visual speech synthesis. In G. Bailly, C. Benoît, & T. R. Sawalis (Eds.),Talking machines: Theories,models, and designs (pp. 335–348). Amsterdam: Elsevier. Google Scholar
Bertelson, P., &Radeau, M. (1976). Ventriloquism, sensoryinteraction, and response bias: Remarks on the paper by Choe, Welch, Gilford, and Juola.Perception & Psychophysics,19, 531–535. Google Scholar
Brooke, N. M., &Summerfield, A. Q. (1983). Analysis, synthesis, and perception of visible articulatory movements.Journalof Phonetics,11, 63–76. Google Scholar
Carpenter, R. H. S. (1988).Movements of the eyes (2nd rev. ed.). London: Pion. Google Scholar
Carpenter, R. H. S. (Ed.) (1991).Eye movements. London: Macmillan. Google Scholar
Cathiard, M.A., Lallouache, M.-T., &Abry, C. (1996). Does movement on the lips mean movement in the mind? In D. Stork & M. Hennecke (Eds.),Speechreading by humans and machines (NATO-ASI Series F: Computer and Systems Sciences, Vol. 150, pp. 211–219). Berlin: Springer-Verlag. Google Scholar
Demorest, M., &Bernstein, L. (1992). Sources of variability in speechreadingsentences: A generalizability analysis.Journal of Speech & Hearing Research,35, 876–891. Google Scholar
Driver, J. (1996). Enhancement of selective listening by illusory mislocation of speech sounds due to lip-reading.Nature,381, 66–68. ArticlePubMed Google Scholar
Eigsti, I. M., Munhall, K. G., Yano, S., &Vatikiotis-Bateson, E. (1995). Effects of listener expectation on eye movement behavior during audiovisual perception.Journal of the Acoustical Society of America,97, 3286. Article Google Scholar
Gailey, L. (1987). Psychological parametersof lip-reading skill. In R. Dodd & B. Campbell (Eds.),Hearing by eye: The psychology of lip-reading (pp. 115–141). Hillsdale, NJ: Erlbaum. Google Scholar
Gray, H. (1977).Gray’s anatomy. New York: Crown. Google Scholar
Green, K. P., &Kuhl, P. K. (1989). The role of visual information in the processing of place and manner features in speech perception.Perception & Psychophysics,45, 34–42. Google Scholar
Green, K. P., &Kuhl, P. K. (1991). Integral processing of visual place and auditory voicing information during phonetic perception.Journal of Experimental Psychology: Human Perception & Performance,17, 278–288. Article Google Scholar
Grüsser, O.-J., &Landis, T. (Eds.) (1991).Visual agnosias and other disturbances of visual perception and cognition (Vision and Visual Dysfunction, Vol. 12). London: Macmillan. Google Scholar
Jacobs, A., &Lévy-Schoen, A. (1988). Breaking down saccade latency into central and peripheral processingtimes in a visual dual task. In G. Lüer, U. Lass, & J. Shallo-Hoffman (Eds.),Eye movement research: Physiological and psychological aspects (pp. 267–285). Lewiston, NY: Hogrefe. Google Scholar
Jeffers, J., &Barley, M. (1971).Speechreading (lipreading). Spring-field, IL: C. C. Thomas. Google Scholar
Lansing, C. R., &McConkie, G. (1994). A new method forspeech-reading research: Tracking observer’s eye movements.Journal of the Academy of Rehabilative Audiology,27, 25–43. Google Scholar
Leigh, R. J., &Zee, D. S. (1991). Oculomotor disorders. In R. H. S. Carpenter (Ed.),Eye movements (Vision and Visual Dysfunction, Vol. 8, pp. 297–319). London: Macmillan. Google Scholar
Luettin, J., Thacker, N. A., &Beet, S. W. (1996). Active shape models for visual speech feature extraction. In D. Stork & M. Hennecke (Eds.),Speechreading by humans and machines (NATO-ASI Series F: Computer and Systems Sciences, Vol. 150, pp. 383–390). Berlin: Springer-Verlag. Google Scholar
MacLeod, A., &Summerfield, Q. (1990). A procedure for measuring auditory and audiovisual speech-reception thresholds for sentences in noise:Rationale, evaluation, and recommendations for use.British Journal of Audiology,24, 29–43. ArticlePubMed Google Scholar
Massaro, D. W. (1987).Speech perception by ear and eye: A paradigm for psychological inquiry. Hillsdale, NJ: Erlbaum. Google Scholar
Massaro, D. W., Tsuzaki, M., Cohen, M. M., Gesi, A., &Heridia, R. (1993). Bimodal speech perception: An examination across languages.Journal of Phonetics,21, 445–478. Google Scholar
McGurk, H., &MacDonald, J. (1976). Hearing lips and seeing voices.Nature,264, 746–748. ArticlePubMed Google Scholar
Montgomery, A. A., &Jackson, P. L. (1983). Physical characteristics of the lips underlying vowel lipreading performance.Journal of the Acoustical Society of America,73, 2134–2144. ArticlePubMed Google Scholar
Moray, N. (1993). Designing for attention. In A. Baddeley & L. Weiskrantz (Eds.),Attention: Selection, awareness, and control. A tribute to Donald Broadbent (pp. 53–72). Oxford: Oxford University Press, Clarendon Press. Google Scholar
Munhall, K. G., Gribble, P., Sacco, L., &Ward, M. (1996). Temporal constraints on the McGurk effect.Perception & Psychophysics,58, 351–362. Google Scholar
Munhall, K. G., &Vatikiotis-Bateson, E. (1998). The moving face during speech communication. In R. Campbell, B. Dodd, & D. Burnham (Eds.),Hearing by eye: Part 2. Advances in the psychology of speechreading and auditory-visual speech (pp. 123–139). Sussex: Taylor & Francis, Psychology Press. Google Scholar
Petajan, E. D. (1985). Automatic lipreading to enhance speech recognition. In_Proceedings: ComputerVision and Pattern Recognition_ (pp. 40–47). San Francisco: IEEE Computer Society Press. Google Scholar
Polyak, S. L. (1941).The retina. Chicago: University of Chicago Press. Google Scholar
Posner, M. I. (1980). Orienting attention.Quarterly Journal of Experimental Psychology,32, 3–25. ArticlePubMed Google Scholar
Reisberg, D., Mclean, J., &Goldfield, A. (1987). Easy to hear but hard to understand. In B. Dodd & R. Campbell (Eds.),Hearing by eye: The psychology of lipreading (pp. 97–114). Hillsdale, NJ: Erlbaum. Google Scholar
Remez, R. E., Rubin, P. E., Berns, S. M., Pardo, J. S., &Lang, J. M. (1994). On the perceptual organization of speech.Psychological Review,101, 129–156. ArticlePubMed Google Scholar
Rosenblum, L. D., Johnson, J. A., &Saldana, H. M. (1996). Visual kinematic information for embellishing speech in noise.Journal of Speech & Hearing Research,39, 1159–1170. Google Scholar
Sekiyama, K., &Tohkura, Y. [I.] (1993). Inter-language differences in the influence of visual cues in speech perception.Journal of Phonetics,21, 427–444. Google Scholar
Sekiyama, K., Tohkura, Y. I., &Umeda, M. (1996). A few factors which affect the degree of incorporating lip-read information intospeech perception. In H. T. Bunnell & W. Idsardi (Eds.),Proceedings: ICSLP 96 (Vol. 3, pp. 1481–1484). Newcastle, DE: Citation Delaware. Google Scholar
Smeele, P. M. T. (1996). Psychology of human speechreading. In D. G. Stork & M. E. Hennecke (Eds.),Speechreading by humans and ma-chines (NATO-ASI Series F: Computer and Systems Sciences, Vol. 150, pp. 3–17). Berlin: Springer-Verlag. Google Scholar
Sumby, W. H., &Pollack, I. (1954). Visual contribution to speech intelligibility in noise.Journal of the Acoustical Society of America,26, 212–215. Article Google Scholar
Summerfield, Q. (1979). Use of visual information for phonetic perception.Phonetics,36, 314–331. Article Google Scholar
Summerfield, Q. (1987). Some preliminaries to a comprehensive account of audio-visual speech perception. In B. Dodd & R. Campbell (Eds.),Hearing by eye: The psychology ojlipreading (pp. 3–52). Hillsdale, NJ: Erlbaum. Google Scholar
Swisher, M. V, Christie, K., &Miller, S. L. (1989). The reception of signs in peripheral vision by deaf persons.Sign Language Studies,63, 99–125. Google Scholar
Vatikiotis-Bateson, E., Eigsti, I. M., &Yano, S. (1994). Listener eye movement behavior during audiovisual perception.Journal of the Acoustical Society of Japan,94-3, 679–680. Google Scholar
Vatikiotis-Bateson, E., Eigsti, I. M., Yano, S., &Munhall, K. G. (1996).Perceiver eye motion during audiovisual perception (TR Tech. Rep. No. TR-203, pp. 1–33). Kyoto: ATR Human Information Processing Research Laboratories. Google Scholar
Vatikiotis-Bateson, E., Munhall, K. G., Hirayama, M., Lee, Y. C, &Terzopoulos, D. (1996). The dynamics Of audiovisual behavior in speech. In D. Stork & M. Hennecke (Eds.),Speechreading by humans and machines (NATO-ASI Series, Series F, Computers and Systems Sciences, Vol. 150, pp. 221–232). Berlin: Springer-Verlag. Google Scholar
Vatikiotis-Bateson, E., &Yehia, H. C. (1996). Physiological modeling of facial motion during speech.Transactions of the Technical Committee on Psychological and Physiological Acoustics,H-96(65), 1–8. Google Scholar
Vatikiotis-Bateson, E., & Yehia, H. C. (1997). Unified model of audiblevisible speech production. In_EuroSpeech ’97: 5th European conference on speech communication and technology_.
Vitkovitch, M., &Barber, P. (1994). Effects of video frame rate on subjects’ ability toshadow one of two cometing verbalpassages.Journal of Speech & Hearing Research,37, 1204–1210. Google Scholar
Wolff, G. J., Prasad, K. V., Stork, D. G., &Hennecke, M. (1994). Lipreading by neural networks: Visual preprocessing, learning and sensory integration. In J. D. Cowan, G. Tesauro, & J. Alspector (Eds.),Advances in neural information processing systems 6 (pp. 1027–1034). San Francisco: Morgan Kaufmann. Google Scholar
Yamada, M. (1993).[Analysis of human visual information processing mechanisms using eye movement] (Tech. Rep. No. 17). Tokyo: Institute of TelevisionEngineers. Google Scholar
Yarbus, A. L. (1967).Eye movements and vision. New York: Plenum. Google Scholar
Yehia, H. C. Rubin, P. E., & Vatikiotis-Bateson, E. (in press). Quantitative association of acoustic, facial, and vocal-tract shapes.Speech Communication.