It’s in your eyes—using gaze-contingent stimuli to create truly interactive paradigms for social cognitive and affective neuroscience (original) (raw)
Journal Article
,
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
,
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
,
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
,
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
,
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
1Institute of Neurosciences and Medicine, Research Centre Juelich, Juelich, 2Department of Psychiatry, 3Department of Social Psychology, and 4Department of Neurology, University of Cologne, Cologne, Germany
Search for other works by this author on:
*These authors contributed equally to this work.
Received:
20 November 2009
Accepted:
17 February 2010
Cite
Marcus Wilms, Leonhard Schilbach, Ulrich Pfeiffer, Gary Bente, Gereon R. Fink, Kai Vogeley, It’s in your eyes—using gaze-contingent stimuli to create truly interactive paradigms for social cognitive and affective neuroscience, Social Cognitive and Affective Neuroscience, Volume 5, Issue 1, March 2010, Pages 98–107, https://doi.org/10.1093/scan/nsq024
Close
Navbar Search Filter Mobile Enter search term Search
Abstract
The field of social neuroscience has made remarkable progress in elucidating the neural mechanisms of social cognition. More recently, the need for new experimental approaches has been highlighted that allow studying social encounters in a truly interactive manner by establishing 'online’ reciprocity in social interaction. In this article, we present a newly developed adaptation of a method which uses eyetracking data obtained from participants in real time to control visual stimulation during functional magnetic resonance imaging, thus, providing an innovative tool to generate gaze-contingent stimuli in spite of the constraints of this experimental setting. We review results of two paradigms employing this technique and demonstrate how gaze data can be used to animate a virtual character whose behavior becomes 'responsive’ to being looked at allowing the participant to engage in 'online’ interaction with this virtual other in real-time. Possible applications of this setup are discussed highlighting the potential of this development as a new 'tool of the trade’ in social cognitive and affective neuroscience.
INTRODUCTION
Social neuroscience has helped to shed light upon the neural mechanisms underlying our ability to understand other minds under the headings of ‘theory of mind’ or ‘mentalizing’, commonly understood as the ability to represent other people’s mental states (Frith and Frith, 2003, 2008). An increasing number of functional neuroimaging studies suggests that two large-scale neural networks are involved, namely the so-called mirror neuron system, comprising essentially the parietal and premotor cortices and the so-called ‘social brain’, comprising essentially the medial prefrontal, the temporopolar, the temporoparietal cortices and the amygdala (Keysers and Gazzola, 2007; Lieberman, 2007).
Most of these studies, however, rely on paradigms in which participants are asked to merely observe others (‘offline’ mentalizing; Schilbach et al., 2006), while mentalizing during ‘online’ social interaction has only been studied by a minority of studies (e.g. Montague et al., 2002; Eisenberger et al., 2003), often making use of game theory paradigms from economics (e.g. Sanfey, 2007). Consequently, the need to develop ‘interactive mind’ paradigms that could provide a platform to systematically study the neural mechanisms of social interaction in an ecologically valid manner has been pointed out recently (Singer, 2006). ‘Online’ interaction crucially involves ‘closing the loop’ between interaction partners and establishing reciprocal relations where actions feed directly into the communication loop and elicit reactions which, in turn, may subsequently lead to reactions of the initiator and so forth. This has been referred to as adopting a ‘second-person-perspective’ (2PP; Reddy, 2003) which can be taken to suggest that awareness of mental states results from being psychologically engaged with someone and being an active participant of reciprocal interaction thereby establishing a subject-subject (‘Me–You’) rather than a subject–object (‘Me–She/He’) relationship. Paradigms that permit the systematic investigation of the reciprocity of interactions as well as the involvement of implicit and explicit processes will substantially enrich our knowledge of the neurobiology of social cognition (Frith and Frith, 2008).
The challenge for social neuroscience here will be twofold: a suitable experimental platform should allow real-time, ‘online’ interactions between participants and the social stimuli while also providing means for experimental control over changes of the social stimuli. Here, we suggest that measurements of participants’ gaze behavior inside the MR environment could be used to influence a virtual character’s gaze behavior making it ‘responsive’ to the participant’s gaze allowing to engage in ‘online’ interaction with the virtual other in real-time.
From a theoretical standpoint, gaze is known to be an important social cue in dyadic interaction indicating interest or disinterest, regulating intimacy levels, seeking feedback and expressing emotions (e.g. Argyle and Cook, 1976; Emery, 2000). In addition, gaze can also influence object perception (Becchio, et al., 2008) by means of establishing triadic relations between two observers and an object onto which the interactors can look ‘together’ and thereby establish ‘joint attention’ (Moore and Dunham, 1995). From a methodological standpoint, gaze behavior represents one of the few ways in which participants can interact with stimuli naturally in spite of the movement constraints when lying inside an MR scanner. Gaze is a socially most salient nonverbal behavior, which can be reliably measured even within an functional magnetic resonance imaging (fMRI) setting and can, therefore, be used as input for interactive paradigms in neuroimaging studies.
This leads to the further question of how gaze input can be used to automatically control contingent behavior of a social stimulus to create the illusion of real-time interaction in such a setting. Pre-programmed and strictly controlled visual presentation of nonverbal behaviors in general and of gaze cues in particular, can be established by using anthropomorphic virtual characters. Such computer-generated characters have been suggested as a valuable tool for social neuroscience (e.g. Sanchez-Vives and Slater, 2005) as they convey social information to human observers (e.g. Bente et al., 2001; Bente et al., 2002, 2008a; Bailenson et al., 2003) and cause reactions strikingly similar to those in real human interactions (e.g. Slater et al., 2006). An increasing number of behavioral as well as fMRI studies has now used such stimuli to study different aspects of social cognition including gaze perception (e.g. Pelphrey et al., 2005; Spiers et al., 2006; Schilbach et al., 2006; Bente et al., 2007, 2008b; Park et al., 2009; Kuzmanovic et al., 2009).
To establish a paradigm that actively engages participants in the 2PP as opposed to being a ‘detached’ observer of social stimuli (from a ‘third-person-perspective’; 3PP), we present a new technique that makes use of eyetracking data obtained from participants inside an MR scanner to control a virtual character’s gaze behavior in real-time making it ‘responsive’ to the human observer’s gaze (Schilbach et al., in press). In this setup the eye movements of the participant become a means to ‘probe’ the behavior of the virtual other similar to real-life situations. Importantly, this also seems compatible with an ‘enactive’ account of social cognition which understands social cognition as bodily experiences resulting from an organism’s adaptive actions upon salient and self-relevant aspects of the environment (e.g. Klin et al., 2003) that feed back into the social interaction process.
Consequently, our setup promises to allow the exploration of the neural basis of processes of interpersonal alignment and the reciprocity inherent to social interaction, i.e. whether and how social cues are detected as contingent upon one’s own behavior and how interaction partners initiate and respond to each other’s actions (Schilbach et al., in press). Both aspects seem highly relevant to make substantial progress in the field of social cognitive neuroscience and may lead to a reconsideration of the current emphasis on similarities between self- and other-related processing.
To implement these different eyetracking setups were tested in the fMRI setting to produce gaze-contingent stimuli. We review results of these different approaches which use eyetracking measurements overtly or covertly to drive MR-compatible experimental paradigms and underline the usability of this technique. Furthermore, we give examples for the applications of these interactive, eyetracking-based paradigms. Given the importance of gaze behavior during real-life social interaction, this approach, we suggest, provides a much needed, new ‘tool of the trade’ for the study of real-time ‘online’ interaction in social neuroscience.
METHODS
Interactive eyetracking setups
‘Interactive eyetracking’ relies on an MR-compatible eyetracking system that allows real-time data transmission to a visual stimulation controller. The controller receives the ongoing gaze data and adapts the visual stimulation according to preset task conditions and the volunteer's current gaze position on screen.
For stimulus delivery, different presentation devices were tested employing either a TFT screen or two different goggle systems. First, a custom-built, shielded TFT screen was used for the stimulus presentation at the rear end of the scanner (14° × 8° horizontal × vertical viewing angle, screen distance from volunteer’s eyes: 245 cm). Volunteers watched the stimuli via a mirror mounted on the head coil. Volunteers’ eye movements were monitored by means of an infrared camera (Resonance Technology, CA, USA). The camera and infra-red light source were mounted on the head coil using a custom-built gooseneck that allowed easy access to the volunteer’s eyes without interfering with the visual stimulation (setup A). Second, stimuli were presented using MR-compatible goggles. Volunteers’ eye movements were monitored by means of an infrared camera that was built into the goggles. In a 3T MR environment we used a VisuaStimTM system (30° × 22.5° horizontal × vertical viewing angle; Resonance Technology, CA, USA; setup B1) whereas in a 1.5 T environment we tested a Silent VisionTM (25.5° × 18° horizontal × vertical viewing angle; AvoTec, FL, USA; setup B2). The raw analog video signals of all setups were digitized at a frame rate of 60 Hz on a dedicated PC running a gaze extraction software (iViewXTM, SMI, Germany, and ClearviewTM, Tobii Technology AB, Sweden, respectively) which produced real-time gaze position output. Careful eyetracking calibration was performed prior to each training or data acquisition session in order to yield valid gaze positions in a stimulus-related coordinate system. Via a fast network connection, gaze position updates were transferred and, thus, made available to another PC running the software which controlled the stimulation paradigm (PresentationTM, http://www.neurobs.com).
Interactive eyetracking with overt feedback
This version was established so that study participants could engage in cognitive tasks using their eye movements only while receiving visual feedback. To this end we coupled the volunteer’s current gaze position to the location of a cursor-like object on the screen (hereafter: gaze cursor). Using their eyes, the volunteers could, thus, voluntarily move the gaze cursor according to the demands of the tasks. For the automatic detection of gaze fixations on screen targets in real-time the following computer-based algorithm was devised: Using PresentationTM software, gaze positions were transformed into stimulus screen coordinates (pixels). A continuously proceeding ‘sliding window’ average of the preceding 60 gaze positions was calculated throughout the whole stimulus presentation (Figure 1). In effect, the gaze cursor marked the volunteers' average gaze position within the preceding 1 s time window providing the observer with a smooth gaze-contingent visual stimulus to which the volunteers quickly adapted despite a brief temporal lag. In particular, this procedure lessened blinking artifacts, averaged out fixational eye movements (2–120 arcmin; Martinez-Conde et al., 2004), and attenuated the impact of erroneous gaze estimates caused by intermittent residual imaging artifacts in the eye video signal. Each ‘sliding window’ average was tested for being part of a coherent fixation period or not and was accepted by the algorithm as part of an ongoing fixation, if the standard deviation of the sliding window gaze elements was below a pre-specified threshold, in which case a counter was incremented. If the standard deviation criterion was not fulfilled, the counter was reset to zero. This procedure was repeated until a fixation period of a pre-specified length, i.e. a pre-specified number of consecutive sliding window averages, was detected. This procedure reliably recognized effective fixations from gaze behavior without prior knowledge of fixation coordinates. Fixations were subsequently tested for being within one of a set of predefined region-of-interests (ROIs) on the stimulus screen. If this was not the case, the algorithm searched for another fixation. This cycle was repeated until either a fixation was found that was within one of the predefined ROIs or the maximum duration of the current task was reached. Time stamps as well as coordinates of detected fixations were stored in a text file for offline data analysis.
Fig. 1
Flow diagram of gaze data processing. Raw gaze data is stored in a sliding window vector whose average value is shown as a gaze cursor in overt gaze feedback paradigms. A counter (c) increments for each consecutive updated sliding window data vector whose standard deviation is below a prespecified threshold (smax). The gaze status is identified as ‘fixation’ if a prespecified number (cmax) of consecutive sliding windows is reached. The average gaze position is thereafter tested for being within one of a given set of ROIs. The described procedure is typically run until a fixation was found within one of the given ROIs triggering a step forward in the experimental paradigm, e.g. the presentation of a new visual stimulus or the change of the gaze direction of a virtual character.
Interactive eyetracking with covert feedback
As during interactive eyetracking with overt feedback, here, participants engage in and ‘drive’ an experimental paradigm by looking at different locations on the screen. In this version of the setup, however, they do not receive visual feedback in form of a gaze cursor.
In conjunction with a virtual character whose gaze behavior could be made contingent upon fixations detected in ROIs this was done to generate an ecologically valid setting in which the gaze behavior of the virtual other could change in response to the human observer’s gaze position on the stimulus screen. In this setup the temporal delay between a relevant fixation and the reaction of the virtual character needed to be small to successfully induce a fluent experience of reciprocal exchange between the participant and the virtual character. Not providing continuous visual feedback via a gaze cursor precluded the possibility of participants being distracted by the gaze cursor but also to adapt to possible measurement errors, even if minimal in size. For example, participants’ head movements or variations in eye illumination could invalidate the initial eye tracking calibration and lead to displaced gaze coordinates. This type of technical problem could be met by defining larger ROIs that were less sensitive to distortions in gaze estimations. Despite minimal offset errors in the gaze coordinates, thus, meaningful reactions of the virtual character were still possible.
The tasks inside the MR environment
To make use of the overt interactive eyetracking mode we implemented two tasks resembling clinical bedside tests for visuo-spatial neglect. In a line bisection task participants had to bisect a horizontal line by fixating it centrally thereby moving a gaze cursor (a vertical line) on the screen into the desired position (Figure 2). In a target cancellation task volunteers had to search for and single out randomly distributed targets among distractors by fixating them one by one thereby moving a gaze cursor (a circle) over each of the detected targets (see online supplementary data for video of task performance). Participants were informed about the paradigm prior to entering the scanner room. Both overt interactive eyetracking tasks were tested with visual stimulation delivered via the TFT screen (setup A) and via goggles (setups B1 and B2). After careful calibration of the eyetracking, volunteers were allowed to get adjusted to the procedure and then went on to perform the task while lying inside the scanner. During this period we ran ‘dummy’ EPI sequences with MRI parameters identical to those in standard imaging experiments. We thereby introduced EPI artifacts in the eyetracking data to test that the devised algorithm would be able to successfully cope with the added noise.
Fig. 2
Overt gaze feedback tasks. (A) While being scanned using ‘dummy’ fMRI scans, a subject was instructed to bisect a horizontal line with a vertical line which is locked to her gaze. In effect, the subject performed the task by fixating the perceived line center. The screen positions of the horizontal lines were randomized. (B) Example of the spatial precision of one subject performing the bisection task. Each of the blue symbols represents one bisection position relative to the true line center. The red cross denotes the average and standard deviation of the spatial bisection error. (C) This histogram shows the frequency distribution of the number of line bisections that subjects were able to perform within one task/block length (21.9 s). On average over 18 subjects, each performing the tasks 15 times, 7.7 line bisections were successfully performed within one block length revealing a rather fluent task performance. (D) In a second task, the subject was asked to cancel targets (‘O’) among distractors by centering the black circle over each target until marked as cancelled. Since the circle's position was locked to the subject's gaze she only had to find and fixate the targets one by one. (E) The spatial precision of the same subject as in (B) performing the cancellation task. (F) On average over 18 subjects, 6.6 cancellations were successfully performed within one block.
In order to test the covert interactive eyetracking mode we made use of a task in which test subjects were asked to respond to or probe the gaze behavior shown by an anthropomorphic virtual character on screen (Figure 3; Schilbach et al., in press). Before participation test subjects were instructed that the gaze behavior shown by the virtual character on screen was actually controlled by a real person who was also participating in the experiment outside the scanner. Likewise, their own gaze behavior was said to be visualized for the other participant outside the MR scanner, so that both participants could engage in gaze-based ‘online’ interaction. During functional neuroimaging participants were instructed to direct the gaze of the other person towards one of three objects by looking at it. Alternatively, they were asked to respond to gaze shifts of the virtual character by either following or not following them to then look at an object. The gaze behavior of the other was made contingent to the participant’s gaze and systematically varied in a 2 × 2 factorial design (joint attention vs non-joint attention; self-initiation vs other-initiation; see online supplementary data for video of task performance). The neural correlates of task performance were investigated employing fMRI in 21 participants.
Fig. 3
Covert gaze feedback task. (A) Screenshot (as seen by participants) depicting anthropomorphic virtual character and three objects. (B) Illustration of gaze samples obtained for one exemplary participant during the experiment.
In this task, stimuli were presented to the participants lying inside the MR scanner using setup A. Due to the screen’s distance from the volunteers’ eyes and the corresponding narrow field of view; changes of the virtual character’s gaze behavior were easily observable while focusing on one of the three objects. Functional MRI (fMRI) data was acquired on a Siemens Trio 3T whole-body scanner (Erlangen, Germany) using blood-oxygen-level-dependent (BOLD) contrast (Gradient-echo EPI pulse sequence, TR = 2.304 s, slice thickness 3 mm, 38 axial slices, in-plane resolution 3 × 3 mm). Additional high-resolution anatomical images (voxel size 1 × 1 × 1 mm3) were acquired using a standard T1-weighted 3D MP-RAGE sequence. The neuroimaging data was preprocessed and analyzed using a general linear model (GLM) as implemented in SPM5 (for further details see: Schilbach et al., in press).
RESULTS
Interactive eyetracking under the constraints of fMRI was successfully installed in all setups. The quality depended on the length of the sliding window which in turn depended on the raw gaze data variance, the main source of which were residual imaging artifacts. Fixational eye movements were the other important source of raw gaze data jitter. The amplitude of such eye movements are in the range of 2–120 arcmin (Martinez-Conde et al., 2004) and their size in screen pixel coordinates scales with the eye's distance to the screen display. This rendered the fixational eye movements’ impact particularly disadvantageous in setup A where a screen distance of 245 cm translated the range of 2–120 arcmin to 2–120 pixels on the screen (given a 800 × 600 pixels resolution). On the other hand, setup A did also include important advantages regarding the handling of the eyetracking camera compared to setups using goggles. The distance as well as the angle with which the camera was positioned in relation to the test subject’s eyes region could be adjusted more easily when the eyetracking camera was mounted to the head coil as in the TFT-based setup. However, goggle systems that allowed a precise eyetracking with their built-in camera did hardly need any camera adjustments (e.g. setup B2). After the feasibility of all setups was secured we chose to use our 3 T MRI system and opted for the TFT-based visual stimulation setup for data acquisition. We nevertheless stress, that the other setups allowed running the paradigms as well.
Interactive eyetracking with overt feedback (line bisection and target cancellation)
During the task in which overt feedback was given, participants were able to use the gaze cursor which they could move across the screen in concordance with their eye movements to execute the task. For the bisection task this amounted to subjects completing an average of 7.7 line bisections in blocks of 21.9 s (n = 18 subjects), whereas during the cancellation task subjects were able to cancel an average of 6.6 targets within the same time (Figure 2). Subjects needed an average of 2.8 s to judge the center of a given line and position the gaze cursor in the respective position. An average additional 0.5 s was needed to search, find, and position the gaze cursor in cancellation tasks (average time between cancellations: 3.3 s). The time subjects needed to choose targets generally depended on the length of the sliding window. If the sliding window length was too small, increased residual gaze cursor jittering made it hard for volunteers to ‘focus’ a target, whereas too long a sliding window increased the temporal lag of the gaze cursor, which reduced the intuitive usability of the gaze feedback. Apart from this, the spatial precision of target choices was increased when the eye tracking calibration was optimal and subjects were well adjusted to the temporal lag of the gaze cursor (Figure 2).
Interactive eyetracking with covert feedback (joint attention)
Having been able to use overt gaze feedback successfully to drive an experiment, we went on to perform the joint attention task which included running regular fMRI measurements. During this task subjects were not given visual feedback in the form of a gaze cursor because we wanted to create a naturalistic as possible setup which allowed for an immersive experience during which participants could interact with the virtual other similarly to how one might interact with another person by means of gaze behavior in real life. In spite of the absence of continuous visual feedback participants were able to fulfill the task in which they had been asked to engage: they were able to establish ‘eye-contact’ with the virtual character and to respond to the virtual character either by following or not following its gaze to either fixate one of three visible objects ‘together with’ the virtual character (other-initiated joint attention: OTHER_JA) or not (other-initiated nonjoint attention: OTHER_NOJA; Figure 3A). Conversely, they were also able to establish ‘eye-contact’ and subsequently direct the virtual character’s gaze towards one of the three objects (self-initiated joint attention: SELF_JA). In an equal number of occasions subjects were unable to do so as the character would ‘react’ by fixating an object other than the one chosen by the participant (self-initiated nonjoint attention: SELF-NOJA). On average, this procedure amounted to approximately four object fixations per ‘interaction segment’ (i.e. a block of 18 s duration) for all conditions (SELF_JA: 4.10 (n = 21, s.d. = 0.68), SELF_NOJA: 4.06 (n = 21, s.d. = 0.79), OTHER_JA: 3.96 (n = 21, s.d. = 0.88), OTHER_NOJA: 4.03 (n = 21, s.d. = 0.88); see Figure 3B for exemplary gaze data; for more details see Schilbach et al., in press). FMRI results demonstrated, firstly, that interpersonal gaze coordination and ‘joint attention’ (main effect of JA) resulted in a differential increase of neural activity in the medial prefrontal cortex (MPFC) and posterior cingulate cortex (PCC) as well as the anterior temporal poles bilaterally. While this activation pattern bears some resemblance with the ‘default mode of brain function’ (Raichle et al., 2001; Schilbach et al., 2008), activations in ventral and dorsal MPFC—at voxel-level correction for multiple comparisons—have been related to outcome monitoring and the understanding of communicative intent (Amodio and Frith, 2006) as well as representations of triadic relations (Saxe, 2006). Conversely, looking at an object different from the one attended by the virtual character—regardless of whether or not this was self-initiated (main effect of NOJA)—recruited a bilateral fronto-parietal network known to be involved in attention and eye-movement control (Schilbach et al., in press; Figure 4).
Fig. 4
Neural correlates of joint attention task. (A) Differential increase of neural activity in MPFC, PCC as well as ventral striatum and anterior temporal poles (latter not illustrated here) for main effect of joint attention. (B) Differential increase of neural activity in medial and lateral parietal as well as frontal cortex bilaterally for main effect of nonjoint attention (taken from: Schilbach et al., in press).
DISCUSSION
Here, we have presented a method by which eyetracking data obtained from study participants lying inside an MR scanner can be processed in real-time in order to directly influence visual stimulus material in spite of the electromagnetic noise associated with fMRI measurements. This can be realized by using a form of overt gaze feedback, i.e. a gaze cursor which subjects can control with their eye movements to carry out a task. Alternatively, this can be done by means of covert gaze feedback where gaze data is used to systematically manipulate and drive the visual stimulation unbeknownst to the participant. In combination with the presentation of anthropomorphic virtual characters whose behavior can be made responsive to the participants’ fixations, the latter technique can be used to allow participants to engage in reciprocal ‘online’ interaction with a virtual other similar to instances of interpersonal gaze coordination in real-life social encounters. Even though similar approaches of using gaze-contingent stimuli have been used in other areas of research (e.g. Duchowski et al., 2004), the development of an MR-compatible version of the technique as presented here is of crucial importance for social cognitive and affective neuroscience as it allows to experimentally target the neural underpinnings of processes during ‘online’ interaction which have so far been largely inaccessible due to the technical constraints of the MR environment.
Different possible paradigms come to mind which could benefit from making use of the here described method. Given the scope of this article, we will limit our description to interactive paradigms in which gaze behavior is exchanged between a human observer and a virtual character. We will focus here on dyadic interaction between two interactors (‘Me–You’), but also on triadic interaction where two interactors relate to an object in the environment (‘Me–You–This’; Saxe, 2006).
Within dyadic interaction gaze is known to have important regulatory functions impacting on a wide range of cognitive, affective and motivational processes (Argyle and Cook, 1976; Emery, 2000). Furthermore, gaze is known to influence our social perception and evaluation of others (e.g. Macrae et al., 2002; Mason et al., 2005) as it conveys the direction of an agent’s attention and has been suggested to be closely related to mentalizing, i.e. our ability to understand other people’s mental states (Nummenmaa and Calder, 2009). Importantly, gaze is also known to ‘connect’ human beings in everyday life situations by means of a ‘communication loop’ in which interactors impact reciprocally on each others’ behavior (e.g. Frith, 2007, p. 175). This procedural dimension of ‘social gaze’ in ‘online’ interaction has only recently begun to be systematically investigated (e.g. Senju and Csibra, 2008) and promises to allow radically new insights into the temporal dynamics of implicit interpersonal ‘alignment’.
As many previous studies concerning the social effects of gaze on person perception have used static, non-interactive stimuli, it may be important to revisit these paradigms by making use of this new technique to validate whether the findings actually result from social communicative processes or not. Here, it is important to note, however, that our setup in its current version does not allow to investigate real social interaction (as in the setup used by Montague et al., 2002), but uses anthropomorphic virtual characters in conjunction with a cover story to generate the impression of interacting with a ‘mindful’ agent. While this can be seen as a limitation of our setup, it is important to note that using gaze feedback has the important benefit of enabling the systematic study of a major component of real interpersonal interaction as it provides a naturalistic way to engage participants. Furthermore, future research could explicitly address how variations of the temporal and stochastic characteristics of a virtual character’s gaze behavior made contingent upon the human observer’s gaze impact on a human observer’s perception of the nature of the agent (‘social’ Turing test).
Apart from aspects related to dyadic interaction, gaze is also known to contribute to the establishment of triadic relations between two interactors who can look at an object together and engage in (gaze-based) joint attention (Moore and Dunham, 1995). Apart from the convergence of gaze directions, this, importantly, also requires mutual awareness of being intentionally directed towards the same aspect of the world which may result directly from the process of interaction. Therefore, joint attention can be construed as an interactively constituted phenomenon whose different facets can only be explored by making use of an interactive paradigm (e.g. Schilbach et al., in press). Interestingly, it has been suggested that being actively engaged in triadic interaction may have an impact both on the perception of the other person (e.g. his/her trustworthiness and attractiveness) as well as on the perception of an object (e.g. its value) that may be jointly attended (Heider, 1958).
To the best of our knowledge, there are no neuroimaging studies which have targeted the neural correlates of the perception of jointly attended objects. Such investigations might be extremely informative, however, by allowing the empirical investigation of the neural correlates of different formats or varying degrees of shared intentionality and could help to investigate the complex relationship of implicit and explicit processes involved in social cognition (Frith and Frith, 2008). Further applications of the method could include investigations of how interactive gaze cues shown by a virtual character impact on object-related decision-making or memory performance. Finally, gaze-based triadic interaction could also help to disentangle differences between ‘online’ and ‘offline’ social cognition, i.e. social cognition from a second- or third-person-perspective, by realizing interactions between two virtual agents and a human observer while introducing systematic differences in social responsiveness of the agents making them more or less likely to actually engage with the participant (Eisenberger et al., 2003).
Apart from investigations of dyadic and triadic interaction in healthy adults, we also see great potential in using the above described method to investigate social cognition in development and diagnostic groups. In spite of the importance of joint attention in ontogeny the neural correlates of this significant phenomenon are only incompletely understood. Given that our paradigm does not rely upon verbal information and higher-order reasoning about others’ mental states, but relies upon naturally occurring social behavior, it might prove to be particularly helpful for the study of the neurofunctional substrates of the development of social cognitive and perceptual abilities during ontogeny.
Specific alterations of social cognition are known to be characteristic of psychiatric disorders such as autism and schizophrenia. In the former case, a dissociation between implicit and explicit processes underlying social cognition has recently been emphasized (Senju et al., 2009). Also, it has been suggested that autistic individuals might be more sensitive to perfect, non-social as compared to imperfect, social contingencies in the environment (Gergely, 2001; Klin et al., 2009). We suggest that the investigation of the neural mechanisms underlying these clinically relevant differences in high-functioning autism will benefit substantially from the method described here (see also Boraston and Blakemore, 2007).
CONCLUSIONS
Taken together, we have shown that the interactive eyetracking method here presented allows to generate gaze-contingent stimuli during fMRI in spite of the electromagnetic noise introduced by such measurements. Used in conjunction with anthropomorphic virtual characters whose behavior can be made ‘responsive’ to the participant’s current gaze position, this method has the potential to substantially increase our knowledge of the neural mechanisms underlying social cognition by making psychological processes accessible for empirical investigation that are involved in the interpersonal coordination of gaze behavior, both in dyadic and triadic interaction. Making use of this new ‘tool of the trade’, we suggest, could open up an entire new avenue of research in social cognitive and affective neuroscience.
We are grateful to our colleagues from the MR group at the Institute for Neurosciences and Medicine, in particular Ralph Weidner and Sharam Mirzazade, for their assistance in acquiring eyetracking data inside the scanner. The study was supported by a grant of the German Ministery of Research and Education (BMBF) on ‘social gaze’ (G.B. and K.V.). L.S. was also supported by the Koeln Fortune Program of the Faculty of Medicine, University of Cologne.
REFERENCES
Meeting of minds: the medial frontal cortex and social cognition
,
Nature Review Neuroscience
,
2006
, vol.
7
4
(pg.
268
-
77
)
,
Gaze and Mutual gaze
,
1976
Cambridge
Cambridge University Press
Interpersonal distance in immersive virtual environments
,
Personality and Social Psychology Bulletin
,
2003
, vol.
29
7
(pg.
819
-
33
)
How the gaze of others influences object processing
,
Trends in Cognitive Science
,
2008
, vol.
12
7
(pg.
254
-
8
)
Computer animated movement and person perception: methodological advances in nonverbal behavior research
,
Journal of Nonverbal Behavior
,
2001
, vol.
25
3
(pg.
151
-
66
)
Virtual gestures: Analyzing social presence effects of computer-mediated and computer-generated nonverbal behaviour
,
Proceedings of the Fifth Annual International Workshop PRESENCE, October, 2002, Porto, Portugal
,
2002
Effects of simulated gaze on social presence, person perception and personality attribution in avatar-mediated communication
,
Proceedings of the 10th Annual International Workshop on Presence 2007
,
2007
Is there anybody out there? Analyzing the effects of embodiment and nonverbal behavior in avatar-mediated communication
,
Mediated Interpersonal Communication
,
2008
Mahwah, NJ
Lawrence Erlbaum Associates
(pg.
131
-
57
)
Deciphering the secret code: A new methodology for the cross-cultural analysis of nonverbal behavior
,
Behavior Research Methods
,
2008
, vol.
40
1
(pg.
269
-
77
)
The application of eye-tracking technology in the study of autism
,
Journal of Physiology
,
2007
, vol.
581
Pt 3
(pg.
893
-
8
)
Gaze-contingent displays: a review
,
Cyberpsychol Behav
,
2004
, vol.
7
6
(pg.
621
-
34
)
Does rejection hurt? An FMRI study of social exclusion
,
Science
,
2003
, vol.
302
5643
(pg.
290
-
2
)
The eyes have it: the neuroethology, function and evolution of social gaze
,
Neuroscience Biobehavior Review
,
2000
, vol.
24
6
(pg.
581
-
604
)
,
Making up the mind
,
2007
Blackwell Publishing
Implicit and explicit processes in social cognition
,
Neuron
,
2008
, vol.
60
3
(pg.
503
-
10
)
Development and neurophysiology of mentalizing
,
Philosophical Transactions of the Royal Society London B: Biological Sciences
,
2003
, vol.
358
1431
(pg.
459
-
73
)
The obscure object of desire: ‘nearly, but clearly not, like me': contingency preference in normal children versus children with autism
,
Bull Menninger Clin
,
2001
, vol.
65
3
(pg.
411
-
26
)
,
The psychology of interpersonal relations
,
1958
New York
Wiley
Integrating simulation and theory of mind: from self to social cognition
,
Trends in Cognitive Science
,
2007
, vol.
11
5
(pg.
194
-
6
)
The enactive mind, or from actions to cognition: lessons from autism
,
Philosophical Transactions of the Royal Society London B: Biological Sciences
,
2003
, vol.
358
1430
(pg.
345
-
60
)
Two-year-olds with autism orient to non-social contingencies rather than biological motion
,
Nature
,
2009
, vol.
459
7244
(pg.
257
-
61
)
et al.
Duration matters: dissociating neural correlates of detection and evaluation of social gaze
,
Neuroimage
,
2009
, vol.
46
4
(pg.
1154
-
63
)
Social cognitive neuroscience: a review of core processes
,
Annual Review of Psychology
,
2007
, vol.
58
(pg.
259
-
89
)
Are you looking at me? Eye gaze and person perception
,
Psycholohical Science
,
2002
, vol.
13
5
(pg.
460
-
4
)
The role of fixational eye movements in visual perception
,
Nature Review of Neuroscience
,
2004
, vol.
5
3
(pg.
229
-
40
)
The look of love: gaze shifts and person perception
,
Psychological Science
,
2005
, vol.
16
3
(pg.
236
-
9
)
et al.
Hyperscanning: simultaneous fMRI during linked social interactions
,
Neuroimage
,
2002
, vol.
16
4
(pg.
1159
-
64
)
,
Joint Attention: Its Origin and Role in Development
,
1995
Hillsdale
Lawrence Erlbaum
Neural mechanisms of social attention
,
Trends in Cognitive Science
,
2009
, vol.
13
3
(pg.
135
-
43
)
et al.
Neural basis of attributional style in schizophrenia
,
Neuroscience Letters
,
2009
, vol.
459
1
(pg.
35
-
40
)
Neural basis of eye gaze processing deficits in autism
,
Brain
,
2005
, vol.
128
(pg.
1038
-
48
)
A default mode of brain function
,
Proceedings of National Academy of Sciences United States of America
,
2001
, vol.
98
2
(pg.
676
-
82
)
On being the object of attention: implications for self-other consciousness
,
Trends in Cognitive Sciences
,
2003
, vol.
7
9
(pg.
397
-
402
)
From presence to consciousness through virtual reality
,
Nature Review of Neurosciences
,
2005
, vol.
6
4
(pg.
332
-
9
)
Uniquely human cognition
,
Current Opinion of Neurobiology
,
2006
, vol.
16
(pg.
235
-
9
)
Social decision-making: insights from game theory and neuroscience
,
Science
,
2007
, vol.
318
5850
(pg.
598
-
602
)
et al.
Being with virtual others: Neural correlates of social interaction
,
Neuropsychologia
,
2006
, vol.
44
5
(pg.
718
-
30
)
Minds at rest? Social cognition as the default mode of cognizing and its putative relationship to the “default system” of the brain
,
Conscious Cognition
,
2008
, vol.
17
2
(pg.
457
-
67
)
et al.
Minds made for sharing. Initiating joint attention recruits reward-related neurocircuitry
,
Journal of Cognitive Neuroscience
[Epub-ahead of print; 25 November 2009]; doi: 10.1162/jocn.2009.21401
Gaze following in human infants depends on communicative signals
,
Current Biology
,
2008
, vol.
18
9
(pg.
668
-
71
)
Mindblind eyes: an absence of spontaneous theory of mind in Asperger syndrome
,
Science
,
2009
, vol.
325
5942
(pg.
883
-
5
)
The neuronal basis and ontogeny of empathy and mind reading: review of literature and implications for future research
,
Neurosci Biobehaviour Review
,
2006
, vol.
30
6
(pg.
855
-
63
)
An experimental study on fear of public speaking using a virtual environment
,
Cyberpsychol Behaviour
,
2006
, vol.
9
5
(pg.
627
-
33
)
Spontaneous mentalizing during an interactive real world task: an fMRI study
,
Neuropsychologia
,
2006
, vol.
44
10
(pg.
1674
-
82
)
Author notes
*These authors contributed equally to this work.
© The Author(s) (2010). Published by Oxford University Press.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com
Supplementary data
Citations
Views
Altmetric
Metrics
Total Views 4,157
3,228 Pageviews
929 PDF Downloads
Since 12/1/2016
Month: | Total Views: |
---|---|
December 2016 | 4 |
January 2017 | 21 |
February 2017 | 32 |
March 2017 | 23 |
April 2017 | 15 |
May 2017 | 17 |
June 2017 | 12 |
July 2017 | 12 |
August 2017 | 39 |
September 2017 | 26 |
October 2017 | 30 |
November 2017 | 19 |
December 2017 | 77 |
January 2018 | 71 |
February 2018 | 68 |
March 2018 | 106 |
April 2018 | 72 |
May 2018 | 79 |
June 2018 | 59 |
July 2018 | 87 |
August 2018 | 70 |
September 2018 | 60 |
October 2018 | 72 |
November 2018 | 60 |
December 2018 | 67 |
January 2019 | 72 |
February 2019 | 79 |
March 2019 | 64 |
April 2019 | 77 |
May 2019 | 52 |
June 2019 | 48 |
July 2019 | 46 |
August 2019 | 57 |
September 2019 | 66 |
October 2019 | 54 |
November 2019 | 56 |
December 2019 | 33 |
January 2020 | 59 |
February 2020 | 47 |
March 2020 | 37 |
April 2020 | 57 |
May 2020 | 29 |
June 2020 | 35 |
July 2020 | 39 |
August 2020 | 38 |
September 2020 | 42 |
October 2020 | 64 |
November 2020 | 61 |
December 2020 | 35 |
January 2021 | 57 |
February 2021 | 77 |
March 2021 | 54 |
April 2021 | 56 |
May 2021 | 52 |
June 2021 | 40 |
July 2021 | 36 |
August 2021 | 40 |
September 2021 | 41 |
October 2021 | 53 |
November 2021 | 30 |
December 2021 | 31 |
January 2022 | 29 |
February 2022 | 34 |
March 2022 | 40 |
April 2022 | 38 |
May 2022 | 25 |
June 2022 | 16 |
July 2022 | 33 |
August 2022 | 30 |
September 2022 | 50 |
October 2022 | 40 |
November 2022 | 43 |
December 2022 | 40 |
January 2023 | 55 |
February 2023 | 29 |
March 2023 | 34 |
April 2023 | 31 |
May 2023 | 91 |
June 2023 | 18 |
July 2023 | 17 |
August 2023 | 39 |
September 2023 | 21 |
October 2023 | 28 |
November 2023 | 24 |
December 2023 | 32 |
January 2024 | 59 |
February 2024 | 28 |
March 2024 | 15 |
April 2024 | 35 |
May 2024 | 40 |
June 2024 | 18 |
July 2024 | 22 |
August 2024 | 31 |
September 2024 | 15 |
October 2024 | 34 |
November 2024 | 11 |
Citations
117 Web of Science
×
Email alerts
Citing articles via
More from Oxford Academic