Crossmodal integration of audio/visual information is vital for recognition, interpretation and appropriate reaction to social signals. Here we examined how rhesus macaques process bimodal species-specific vocalizations by eye tracking, using an unconstrained preferential looking paradigm. Six adult rhesus monkeys (3M, 3F) were presented two side-by-side videos of unknown male conspecifics emitting different vocalizations, accompanied by the audio signal corresponding to one of the videos. The percentage of time animals looked to each video was used to assess crossmodal integration ability and the percentages of time spent looking at each of the six a priori ROIs (eyes, mouth, and rest of each video) were used to characterize scanning patterns. Animals looked more to the congruent video, confirming reports that rhesus monkeys spontaneously integrate conspecific vocalizations. Scanning patterns showed that monkeys preferentially attended to the eyes and mouth of the stimuli, with subtle differences between males and females such that females showed a tendency to differentiate the eye and mouth regions more than males. These results were similar to studies in humans indicating that when asked to assess emotion-related aspects of visual speech, people preferentially attend to the eyes. Thus, the tendency for female monkeys to show a greater differentiation between the eye and mouth regions than males may indicate that female monkeys were slightly more sensitive to the socio-emotional content of complex signals than male monkeys. The current results emphasize the importance of considering both the sex of the observer and individual variability in passive viewing behavior in nonhuman primate research.
Citation: Payne C, Bachevalier J (2013) Crossmodal Integration of Conspecific Vocalizations in Rhesus Macaques. PLoS ONE 8(11): e81825. doi:10.1371/journal.pone.0081825
Editor: Elsa Addessi, CNR, Italy
Received: April 19, 2012; Accepted: October 28, 2013; Published: November 13, 2013
Copyright: © 2013 Payne, Bachevalier. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: This work has been supported by National Institute of Child Health and Human Development 35471; National Institute Mental Health 58846; Yerkes Base Grant NIH RR00165 (currently supported by the Office of Research Infrastructure Programs/OD P51OD11132); Center for Behavioral Neuroscience grant NSF IBN-9876754; The Robert W. Woodruff Health Sciences Center Fund, Emory University; National Institute of Mental Health T32-MH0732505; Autism Speaks Mentor-Based Predoctoral Fellowship Grant: 1657. The funders had no role in the study design, data collection and analysis, decision to publish or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Successful integration into complex social environments requires humans and nonhuman primates to recognize, manipulate, and behave according to the immediate social context. Key elements of this task are building representations of relations between self and others, and flexibly using these representations to guide social behavior [1,2]. This set of skills relies upon the ability to distinguish and interpret social cues that are often broadcast over multiple sensory modalities. Hence, crossmodal integration has become a crucial component of social success in primates.
The remarkable behavioral [3-6] similarities between humans and nonhuman primates include the use of species-specific facial expressions and vocalization [7-9]. For both species, decoding the specific “message” of a social display relies on crossmodal integration. The rhesus communicative system is comprised of a small repertoire of relatively fixed calls characterized with distinct facial expressions, postures, and gestures and associated with particular social contexts. This repertoire has been successfully used to explore the evolutionary basis and neural mechanisms of visual speech perception (reviewed by ).
Recent studies have demonstrated that rhesus macaques spontaneously recognize the correspondence between facial and vocal expressions . When pairs of videos depicting two different conspecific vocalizations (i.e., coo and threat) are presented simultaneously with the auditory track matching one of the videos, rhesus macaques look longer to the congruent stimulus video. This is interpreted as spontaneous integration of the auditory and visual components of the stimuli. This paradigm, however, does not rule out the possibility that monkeys merely rely upon the temporal coincidence of facial movements with the onset of the vocal track. A subsequent electrophysiological experiment using the same videos presented sequentially and including a non-biological, mechanical control that mimicked the mouth movements of the videos (in space and time) indicates that integration of the bimodal vocalizations is not dependent upon temporal coincidence . However, given that the videos in the latter experiment were presented individually, the possibility remains that the preference for congruence observed in the preferential viewing paradigm is attributable to the mechanical or temporal coincidence of the auditory and visual components of the stimulus videos. The mechanisms underlying this spontaneous preference for congruence have yet to be systematically explored; and little is known about the visual scanning strategies used by monkeys during crossmodal integration. It has been demonstrated that human subjects modify their scanning strategies of audiovisual stimuli based on the information they are instructed to extract and the efficacy of the social signals [13-16]. It has also been suggested that men and women are differentially sensitive to the emotional content of audiovisual social communication , which may manifest as sexual dimorphic scanning strategies.
To date, the only investigation to monitor how monkeys look at socially salient bimodal stimuli was designed to explore the evolutionary basis for humans’ use of facial cues to enhance speech comprehension . This report highlighted the importance of the eye region to rhesus monkeys, but did not directly identify the facial cues needed to support a preference for congruence. Nor did this report assess sex differences in the way male and female rhesus macaques process socio-emotional stimuli. Accordingly, the goals of the present investigation were to assess integration ability in surrogate nursery-reared male and female rhesus macaques using a preferential viewing paradigm; determine whether spontaneous integration ability is solely dependent upon temporal or mechanical coincidence of the auditory and visual components of species-typical vocalizations using an ethologically relevant mechanical control; and characterize the scanning strategies during the preferential viewing paradigm to determine what features the male and female rhesus macaques use to process the stimuli using eye-tracking technology.
All procedures were approved by the Animal Care and Use Committee of the University of Texas Health Science Center at Houston in Houston, TX and of Emory University in Atlanta, GA and carried out in accordance with the National Institutes of Health Guide for the Care and Use of Laboratory Animals. Power analyses were completed to determine that a minimum of 5 trials were required to detect large effects at 80% power in a cohort of monkeys with 3 males and 3 females.
Six adult rhesus monkeys (Macaca mulatta) aged 4-6 years (3 males, 3 females) were used in this investigation. Animals were surrogate-peer reared in a socially enriched environment that promoted species-specific social skills and alleviated psychological stress [19-21]. Surrogate-peer rearing involved individual housing in size-appropriate wire cages that allowed physical contact with animals in neighboring cage(s), as well as visual, auditory, and olfactory contact with all other infants in the nursery. Each infant was provided a synthetic plush surrogate and cotton towels for contact comfort. The infants received daily social interaction with age- and sex-matched peers as well as with human caregivers, and had repeated assessments of memory, emotional reactivity, social behavior, and reward appraisal throughout their lives. These animals served as sham-operated controls in a program of experiments designed to characterize the functional and neuroanatomical development of hippocampus, amygdala and orbital frontal cortex. Accordingly, they received sham operations at 10-12 days of age, which included small bilateral craniotomies with no penetration of the dura layer (for details, see 19) and underwent multiple magnetic resonance imaging (MRI) scans to assess gross neural development between 2 weeks and 2.5 years of age . All neuroimaging and surgical procedures were performed under deep anesthesia (Isoflurane, 1-2%) and using aseptic procedures. Animals received pre- and post-surgical treatments to minimize risk of infection (Cephazolin, 25 mg/kg, per os) and control swelling (dexamethazone sodium phosphate, 0.4 mg/kg, s.c.). Topical antibiotic ointment (bacitracin-neomycin-polymyxin) was applied daily and acetaminophen (10mg/kg, p.o.) was given four times a day to relieve pain.
Crossmodal Integration Task
A preferential viewing paradigm similar to that used by Ghazanfar and Logothetis  was selected in the present investigation.
Testing was completed in a sound-attenuated room. Monkeys were seated in a primate chair 2-feet from of a 24-inch, flat panel LCD monitor with attached speaker and small eye-tracking camera (60 Hz; ISCAN, Inc.; Woburn, MA). Head movements were gently minimized with a restraint device attached to the primate chair. Ambient white noise was played to further dampen unrelated noises and a curtain concealed all additional equipment.
Animals were presented two side-by-side digital 2-sec videos of the facial gestures associated with species-typical calls (coo, grunt, scream and threat). The videos were those used by Ghazanfar and Logothetis  and depicted two unknown rhesus monkeys (stimulus animals) emitting the vocalizations. One stimulus animal generated the coo and threat vocalizations and the other stimulus animal generated the grunt and scream vocalizations (see Figure 1). Videos were 640 x 480 pixels and spaced apart maximally (200 pixels) on a solid black background. The sound track corresponding with one of the presented facial gestures was heard through the speaker centered beneath the monitor. The auditory and visual components were played in a continuous loop for 10 sec (5 repetitions). Stimulus presentation was controlled using the Presentation software package (Neurobehavioral Systems, Inc; Albany, CA).
Screen shots of coo-grunt (A) and scream-threat (B) pairings with borders of eye and mouth ROIs. In (A), the audible vocalization was a “coo” and in (B), the audible vocalization was a “threat”. ROIs were determined such that the entire region was included throughout the entire video, resulting in slightly extended ROIs in the still representation of the videos. Stimulus sets were comprised of all possible combinations of videos. Labels were not part of stimuli.
The auditory component and the left-right position of the two facial gestures were counterbalanced. Stimuli were presented under two different conditions: Synchronized and Desynchronized. The Synchronized condition was used as the standard for integration assessment and were constructed such that the onsets of the auditory and visual components were simultaneous. A total of eight trials in the Synchronized condition were administered across four testing sessions (2 trials/day). The Desynchronized condition was employed to assess whether integration ability relied only upon the mechanical properties of the stimuli (i.e. the coincidence of mouth movements with the auditory component). Trials in the Desynchronized condition were constructed such that the onset of the auditory component was delayed 330 - 430 msec from the onset of the visual component, a delay range that has been shown to disrupt the perception of the stimuli as a single event  and resulted in no overlap between the mouth movements and sound. A total of eight trials in the Desynchronized condition were administered across two testing sessions (4 trials/day).
In a given trial, there was one congruent video (i.e., depicted the facial gestures that matched the audio component) and one incongruent video (i.e., facial gestures did not match the audio track). Crossmodal integration was determined by comparing the percent looking time to each video to the chance level of 50%. Integration of the audio and visual components was inferred when monkeys showed a preference for one of the video clips (i.e., looked significantly more than chance to either the congruent or incongruent stimulus video). Accordingly, an inability to integrate the complex social signals would be demonstrated by monkeys exhibiting equal looking times to each video in the pair.
Scanning Pattern Characterization.
Percentages of looking time to a priori regions of interest (ROIs) of the videos were recorded. Static ROIs of the eyes and mouth were created with the ISCAN P.O.R. Fixation Analysis software (v1.2, ISCAN, Inc., Figure 1) such that each ROI encapsulated the entire feature of interest throughout the entire 2-sec video. The region of the video not included in either the “eyes” or “mouth” ROI was analyzed as the third ROI labeled “other”. There were six ROIs in each trial: eyes, mouth, and other for each of the two stimulus videos. Scanning patterns were characterized by comparing the amount of time animals spent looking at each ROI, which was calculated from the summation of the fixation durations in a given ROI. A fixation was defined as the eye gaze coordinates remaining within 1° x 1° visual angle for at least 50 msec. Fixations were categorized by ROI using the ISCAN P.O.R. Fixation Analysis Software, and variability in looking time across trials and animals was accounted for by expressing looking to each ROI as a percentage of total looking ((ROI/Total)*100).
All measures were normally distributed (Shapiro-Wilk W = 0.799-1.000, p = 0.112-0.973). Integration abilities were assessed separately for the Synchronized and Desynchronized conditions by comparing the percentages of looking to the congruent stimuli to the chance level of 50% using a one-sample t-test. Repeated measures ANOVA were used to evaluate sex differences and to compare the integration abilities across conditions. Scanning patterns of the ROIs of each stimulus video in a trial were analyzed using repeated measures MANOVA (stimulus video x ROI x sex) with simple interactions and simple comparisons used to conduct planned comparisons of the relative looking to individual ROIs across stimulus video and sex. The assumption of equality of variances was met for all analyses (Levene’s: F(1,4) = 0.007-7.357, p = 0.053-0.939) except for two measures in the analysis of the congruent and incongruent stimulus videos across all trials (Levene’s: F(1,4) = 8.952 - 9.336, p = 0.038-0.040). Natural log transformations were used to correct for the violations.
Overall Integration and Scanning Patterns
In the Synchronized condition (Figure 2), animals exhibited spontaneous integration of complex crossmodal social signals by looking significantly more than chance to the congruent stimulus video (t(5) = 2.941, p = 0.032). Qualitatively this effect appears to be driven by the behavior of the females (see Figure 2, open symbols), but this apparent sex difference was not statistically significant (F(1,4) = 2.186, p = 0.213). In the Desynchronized condition, animals did not show a preference for congruence (t(5) = -1.115, p = 0.316; Figure 2), and males and females did not differ (F(1,4) = 0.060, p = 0.819).
Percentages of looking time (± s.e.m.) to the congruent stimulus video in the Synchronized and Desynchronized conditions. The dashed line represents chance level of 50%. Values greater than 50% represent greater looking to the congruent video and values less than 50% represent greater looking to the incongruent video. Symbols represent individual data points for males (filled) and females (open). (*) p < 0.05.
Scanning Pattern Characterization.
Figure 3 illustrates notable differences in monkeys’ exploration of the congruent and incongruent stimulus videos in the Synchronized condition. On the congruent stimulus video (Figure 3A), they spent more time looking to the eye region than the mouth region (F(1,4) = 69.115, p = 0.001), and looked longer to the eye and mouth regions than to the rest of the video (eyes > other F(1,4) = 45.672, p = 0.003; mouth > other F(1,4) = 21.927, p = 0.009). There was a weak trend for a sex difference in the relative looking to the eye and mouth regions (ROI x Sex interaction: F(1,4) = 5.262, p = 0.084), with females exhibiting a larger differentiation than males (Figure 3A inset). On the incongruent stimulus video (Figure 3B), monkeys spent comparable amounts of time looking at the eyes and either the mouth or rest of the stimulus video (eyes = mouth: F(1,4) = 0.001, p = 0.972; eyes = other: F(1,4) = 2.544, p = 0.186) but looked more to the mouth than the rest of the stimulus video (mouth > other: F(1,4) = 9.558, p = 0.037). No sex differences were observed for the incongruent stimulus video.
Percentages of looking time (± s.e.m.) to the eyes (e), mouth (m), and other (o) of the congruent stimulus video (A), the incongruent stimulus video (B). Inset in A: Percentages of looking time (± s.e.m.) to the eyes (diamonds/solid line) and mouth (circles/dashed line) of the congruent stimulus video in for males and females. (*) p ≤ 0.05.
There were no interactions of Condition with either ROI (F(2,8) = 0.121, p = 0.887) or Stimulus Video and ROI (F(2,8) = 0.086, p = 0.918); neither were there interactions between these factors and Sex (Condition x ROI x Sex: (F(2,8) = 0.004, p = 0.996; Condition x Stimulus Video x ROI x Sex: F(2,8) = 1.904, p = 0.211). This indicates that scanning of trials in the Desynchronized condition did not differ from that of trials in the Synchronized condition. Planned simple effect comparisons of Condition at the individual ROIs confirmed that animals spent comparable proportions of time looking at the eyes (congruent: F(1,4) = 1.550, p = 0.281; incongruent: F(1,4) = 2.518, p = 0.188), mouth (congruent: F(1,4) = 0.280, p = 0.625; incongruent: F(1,4) = 2.406, p = 0.196) and rest of the stimulus videos (congruent: F(1,4) = 0.365, p = 0.578; incongruent: F(1,4) = 0.014, p = 0.911) of the Synchronized and Desynchronized conditions.
The results confirm previous findings that rhesus macaques spontaneously integrate the auditory and visual components of complex social cues emitted by novel conspecific males . They further suggested that these abilities might be influenced by, but perhaps not dependent upon, the mechanical properties of stimuli. Finally, monkeys looked at the eyes of the congruent stimulus video more than other facial cues, with females showing a slightly larger differentiation between eyes and mouth than males.
Before discussing the implications of these results, it is important to acknowledge the impact of individual variability on the current findings. This investigation employed an experimental design that assesses the animals’ spontaneous looking behavior. Therefore, unlike more cognitive crossmodal matching tasks that require responders to determine the inter-sensory relatedness of two stimuli in order to receive a reward, there is no right or wrong video in a preferential viewing paradigm. Inferences were based on where the animals “prefer” to look, which could vary substantially across animals. For example, the female represented by the open triangle demonstrated a preference for congruence in the Synchronized condition but looked more to the incongruent stimulus video in the Desynchronized condition. Assessment of scanning patterns of this animal revealed that it looked most to the eye region of the congruent video, but in the incongruent video, it fixated most on the mouth region. Comparatively, the female represented by the open circle demonstrated a clear preference for the congruent video in the Synchronized condition but looked more equally to the videos in the Desynchronized condition; and this animal’s scanning patterns across the congruent and incongruent videos were strikingly similar to each other, with a strong preference of the eye region in both videos.
This variability should be considered when interpreting the lack of a preference for congruence in the Desynchronized condition. Studies employing non-social control conditions have previously shown that integration ability does not rely solely on the mechanical properties of the stimuli . This brings to question whether the lack of preference observed in the Desynchronized condition of the current investigation was due to the social complexity of the stimuli. As illustrated in Figure 2, in the Desynchronized condition, two animals looked slightly more towards the congruent video, whereas two animals looked slightly more and two animals looked substantially more towards the incongruent video. The social complexity of the stimuli makes it difficult to interpret how the Desynchronized videos were processed. One reasonable explanation for the variability seen across animals is that different animals focus on different aspects of the stimuli (e.g., social content or mechanical properties). Thus, although the lack of significant preference in the Desynchronized condition could indicate that rhesus macaques relied on the temporal coincidence of the auditory and visual components for integration into a single construct, contradictory previous findings  combined with the individual variability and lack of differences in scanning patterns across the Synchronized and Desynchronized conditions observed in the current study suggests that further analysis is needed.
Viewing of Eye Regions
Characterization of the scanning patterns indicated that rhesus monkeys attended to the eye regions of the stimulus animals as they evaluated the dynamic, bimodal vocalizations. This interest in the eye region adds to a number of previous studies reporting that both humans and monkeys preferentially investigate the eye regions of conspecifics presented either in static images [24-34] or dynamic, naturalistic videos [18,35-37]. Both humans and rhesus monkeys broadcast important socio-emotional information through their eyes (e.g., their emotional or mental state, social intentions, or focus of their attention), thus attending to the eye region provides the observer with a wealth of socially relevant information .
Interestingly, males and females exhibited subtle differences in their looking of the eye and mouth regions of the congruent stimulus video, with females showing a slightly greater differentiation between the regions than males. Although differential scanning by males and females has not been empirically investigated in monkeys, previous studies have shown that humans modify their gaze behavior based on the information they intend to extract. Thus, when instructed to focus on emotion-related cues (e.g., prosody) or make social judgments, human subjects look more to the eye region than the mouth region [13,14]. However, when attending to speech-specific aspects of the communication signal (e.g., phonetic details in high levels of ambient noise), they focus significantly more to the mouth region [15,16]. Interestingly, when allowed to passively view videos of vocalizing actors, human subjects also preferentially attend to the eye regions [36,37]. It can thereby be inferred that, during passive viewing, humans preferentially attend to the socio-emotional aspects of the stimuli. By extension, the present findings suggest that monkeys attended to the socio-emotional aspects of the stimuli. The results further suggest that female monkeys may be slightly more sensitive to the socio-emotional content of complex signals than male monkeys. Although further studies are clearly needed to better understand the significance of this sex difference, the data parallel recent findings in humans indicating that women recognize crossmodal emotional expressions of fear and disgust strikingly better than men .
Humans and nonhuman primates live in complex social environments where social signals are primarily transmitted via faces and vocalizations. The ability to process audiovisual information is necessary for the recognition of individuals and their emotional states. Rhesus macaques possess the ability to integrate the audio and visual components of species-specific vocalizations, and females may be slightly more attuned to the socio-emotional aspects of complex, species-specific social signals. The current results emphasize that subsequent investigations in nonhuman primates should take into account the sex of the observer, as well as considerable individual variability in passive viewing behavior.
Characterization of these types of naturally occurring behavioral differences in normal subjects and the identification of the neural substrates of those differences are particularly important for research on disorders characterized by deficits in emotional crossmodal integration, such as autism spectrum disorder [39-42], pervasive developmental disorder [43,44]; and schizophrenia [45-47]. Only a few functional neuroimaging studies in humans have begun to identify neuroanatomical correlates of emotional crossmodal integration and have shown greater responses to bimodal emotional expressions (face and voice) than unimodal emotional expressions in the amygdala , medial temporal gyrus, anterior fusiform gyrus , and posterior superior temporal gyrus (), as well as the thalamus . None have documented sex differences in activation patterns. Although several investigations have empirically demonstrated emotional crossmodal integration abilities in nonhuman primates (e.g. [51-53]), to date, the neural substrates of these abilities in monkeys have yet to be investigated.
We thank asif Ghazanfar, Princeton University Neuroscience Institute Departments of Psychology and Ecology & Evolutionary Biology, for generously allowing us to use his stimuli, and Nancy Bliwise, Emory University Psychology Department, for her guidance on data analysis. We also thank Lisa Parr, Yerkes National Primate Research Center, and Harold Gouzoules, Emory University Psychology Department, for their contributions to the experimental design.
Conceived and designed the experiments: CP JB. Performed the experiments: CP. Analyzed the data: CP. Contributed reagents/materials/analysis tools: CP JB. Wrote the manuscript: CP JB.
- 1. Baron-Cohen S, Ring HA, Wheelwright S, Bullmore ET, Brammer MJ et al. (1999) Social intelligence in the normal and autistic brain: an fMRI study. Eur_J Neurosci 11: 1891-1898. doi:10.1046/j.1460-9568.1999.00621.x. PubMed: 10336657.
- 2. Adolphs R, Sears L, Piven J (2001) Abnormal processing of social information from faces in autism. J Cogn Neurosci 13: 232-240. doi:10.1162/089892901564289. PubMed: 11244548.
- 3. Brothers L (1995) Neurophysiology of the perception of intention by primates. In: M. Gazzaniga. The Cognitive Neurosciences. Cambridge, MA: The MIT Press. p. 1107.
- 4. Byrne RW, Whiten A (1988) Machiavellian intelligence: social expertise and the evolution of intellect in monkeys, apes and humans. Oxford: Clarendon Press.
- 5. Cheney DL, Seyfarth RM (1990) How Monkeys See the World. Chicago: University of Chicago Press.
- 6. de Waal FBM (1989) Peacemaking among primates. Boston: Harvard University Press.
- 7. Hauser MD (1993) Right hemisphere dominance for the production of facial expression in monkeys. Science 261: 475-477. doi:10.1126/science.8332914. PubMed: 8332914.
- 8. Hinde R, Roswell T (1962) Communication by postures and facial expressions in the rhesus monkey (Macaca mulatta). Proc Roy Soc Lon B 138. doi: 10.1111/j.1469-7998.1962.tb05684.x
- 9. Partan SR (2004) Multisensory animal communication. In: G. CalvertC. SpenceB. Stein. The Handbook of Multisensory Processes. Cambridge, MA: The MIT Press. pp. 225-242.
- 10. Romanski LM, Ghazanfar AA (2010) The primate frontal and temporal lobes and their role in multisensory vocal communication. In: M. PlattA. Ghazanfar. Primate Neuroethology. New York: Oxford University Press. pp. 500-524.
- 11. Ghazanfar AA, Logothetis NK (2003) Neuroperception: facial expressions linked to monkey calls. Nature 423: 937-938. doi:10.1038/423937a. PubMed: 12827188.
- 12. Ghazanfar AA, Maier JX, Hoffman KL, Logothetis NK (2005) Multisensory integration of dynamic faces and voices in rhesus monkey auditory cortex. J Neurosci 25: 5004-5012. doi:10.1523/JNEUROSCI.0799-05.2005. PubMed: 15901781.
- 13. Buchan JN, Paré M, Munhall KG (2008) The effect of varying talker identity and listening conditions on gaze behavior during audiovisual speech perception. Brain Res 1242: 162-171. doi:10.1016/j.brainres.2008.06.083. PubMed: 18621032.
- 14. Lansing CR, McConkie GW (1999) Attention to facial regions in segmental and prosodic visual speech perception tasks. J Speech Lang Hear Res 42: 526-539. PubMed: 10391620.
- 15. Lansing CR, McConkie GW (2003) Word identification and eye fixation locations in visual and visual-plus-auditory presentations of spoken sentences. Percept Psychophys 65: 536-552. doi:10.3758/BF03194581. PubMed: 12812277.
- 16. Vatikiotis-Bateson E, Eigsti IM, Yano S, Munhall KG (1998) Eye movement of perceivers during audiovisual speech perception. Percept Psychophys 60: 926-940. doi:10.3758/BF03211929. PubMed: 9718953.
- 17. Collignon O, Girard S, Gosselin F, Saint-Amour D, Lepore F et al. (2010) Women process multisensory emotion expressions more efficiently than men. Neuropsychologia 48: 220-225. doi:10.1016/j.neuropsychologia.2009.09.007. PubMed: 19761782.
- 18. Ghazanfar AA, Nielsen K, Logothetis NK (2006) Eye movements of monkey observers viewing vocalizing conspecifics. Cognition 101: 515-529. doi:10.1016/j.cognition.2005.12.007. PubMed: 16448641.
- 19. Goursaud AP, Bachevalier J (2007) Social attachment in juvenile monkeys with neonatal lesion of the hippocampus, amygdala and orbital frontal cortex. Behav Brain Res 176: 75-93. doi:10.1016/j.bbr.2006.09.020. PubMed: 17084912.
- 20. Roma PG, Champoux M, Suomi SJ (2006) Environmental control, social context, and individual differences in behavioral and cortisol responses to novelty in infant rhesus monkeys. Child Dev 77: 118-131. doi:10.1111/j.1467-8624.2006.00860.x. PubMed: 16460529.
- 21. Sackett GP, Ruppenthal GC, Davis AE (2002) Survival, growth, health, and reproduction following nursery rearing compared with mother rearing in pigtailed monkeys (Macaca nemestrina). Am J Primatol 56: 165-183. doi:10.1002/ajp.1072.abs. PubMed: 11857653.
- 22. Payne C, Machado CJ, Bliwise NG, Bachevalier J (2010) Maturation of the hippocampal formation and amygdala in Macaca mulatta: a volumetric magnetic resonance imaging study. Hippocampus 20: 922-935. doi:10.1002/hipo.20688. PubMed: 19739247.
- 23. Dixon NF, Spitz L (1980) The detection of auditory visual desynchrony. Perception 9: 719-721. doi:10.1068/p090719. PubMed: 7220244.
- 24. Adolphs R, Gosselin F, Buchanan TW, Tranel D, Schyns P et al. (2005) A mechanism for impaired fear recognition after amygdala damage. Nature 433: 68-72. doi:10.1038/nature03086. PubMed: 15635411.
- 25. Farzin F, Rivera SM, Hessl D (2009) Brief report: Visual processing of faces in individuals with fragile X syndrome: an eye tracking study. J Autism Dev Disord 39: 946-952. doi:10.1007/s10803-009-0744-1. PubMed: 19399604.
- 26. Gamer M, Büchel C (2009) Amygdala activation predicts gaze toward fearful eyes. J Neurosci 29: 9123-9126. doi:10.1523/JNEUROSCI.1883-09.2009. PubMed: 19605649.
- 27. Gothard KM, Brooks KN, Peterson MA (2009) Multiple perceptual strategies used by macaque monkeys for face recognition. Anim Cogn 12: 155-167. doi:10.1007/s10071-008-0179-7. PubMed: 18787848.
- 28. Gothard KM, Erickson CA, Amaral DG (2004) How do rhesus monkeys (Macaca mulatta) scan faces in a visual paired comparison task? Anim Cogn 7: 25-36. doi:10.1007/s10071-003-0179-6. PubMed: 14745584.
- 29. Guo K, Robertson RG, Mahmoodi S, Tadmor Y, Young MP (2003) How do monkeys view faces?--A study of eye movements. Exp Brain Res 150: 363-374. PubMed: 12707744.
- 30. Kennedy DP, Adolphs R (2010) Impaired fixation to eyes following amygdala damage arises from abnormal bottom-up attention. Neuropsychologia 48: 3392-3398. doi:10.1016/j.neuropsychologia.2010.06.025. PubMed: 20600184.
- 31. Machado CJ, Nelson EE (2011) Eye-tracking with nonhuman primates is now more accessible than ever before. Am J Primatol 73: 562-569. doi:10.1002/ajp.20928. PubMed: 21319204.
- 32. Pelphrey KA, Sasson NJ, Reznick JS, Paul G, Goldman BD et al. (2002) Visual scanning of faces in autism. J Autism Dev_Disord 32: 249-261. doi:10.1023/A:1016374617369. PubMed: 12199131.
- 33. Riby DM, Hancock PJ (2008) Viewing it differently: social scene perception in Williams syndrome and autism. Neuropsychologia 46: 2855-2860. doi:10.1016/j.neuropsychologia.2008.05.003. PubMed: 18561959.
- 34. Spezio ML, Huang PY, Castelli F, Adolphs R (2007) Amygdala damage impairs eye contact during conversations with real people. J Neurosci 27: 3994-3997. doi:10.1523/JNEUROSCI.3789-06.2007. PubMed: 17428974.
- 35. Buchan JN, Paré M, Munhall KG (2007) Spatial statistics of gaze fixations during dynamic face processing. Soc Neurosci 2: 1-13. doi:10.1080/17470910601043644. PubMed: 18633803.
- 36. Everdell IT, Marsh HO, Yurick MD, Munhall KG, Paré M (2007) Gaze behaviour in audiovisual speech perception: asymmetrical distribution of face-directed fixations. Perception 36: 1535-1545. doi:10.1068/p5852. PubMed: 18265836.
- 37. Klin A, Jones W, Schultz R, Volkmar F, Cohen D (2002) Visual fixation patterns during viewing of naturalistic social situations as predictors of social competence in individuals with autism. Arch Gen Psychiatry 59: 809-816. doi:10.1001/archpsyc.59.9.809. PubMed: 12215080.
- 38. Emery NJ (2000) The eyes have it: the neuroethology, function and evolution of social gaze. Neurosci Biobehav Rev 24: 581-604. doi:10.1016/S0149-7634(00)00025-7. PubMed: 10940436.
- 39. Hobson RP, Ouston J, Lee A (1988) Emotion recognition in autism: coordinating faces and voices. Psychol_Med 18: 911-923. PubMed: 3270834.
- 40. Loveland K (2005) Social-emotional impairment and self-regulation in autism spectrum disorders. In: J. NadelD. Muir. Typical and impaired emotional development. Oxford: Oxford University Press. pp. 365-382.
- 41. Loveland K, Pearson D, Reddoch S (2005) Anxiety symptoms in children and adolescents with autism. International Meeting for Autism. Research. Boston, MA.
- 42. Loveland K, Tunali-Kotoski B, Chen R, Brelsford K, Ortegon J et al. (1995) Intermodal perception of affect in persons with autism or Down syndrome. Dev Psychopathol 7. doi: 10.1017/s095457940000660x
- 43. Magnée MJ, de Gelder B, van Engeland H, Kemner C (2007) Facial electromyographic responses to emotional information from faces and voices in individuals with pervasive developmental disorder. J Child Psychol Psychiatry 48: 1122-1130. doi:10.1111/j.1469-7610.2007.01779.x. PubMed: 17995488.
- 44. Magnée MJ, de Gelder B, van Engeland H, Kemner C (2008) Audiovisual speech integration in pervasive developmental disorder: evidence from event-related potentials. J Child Psychol Psychiatry 49: 995-1000. doi:10.1111/j.1469-7610.2008.01902.x. PubMed: 18492039.
- 45. de Gelder B, Vroomen J, de Jong SJ, Masthoff ED, Trompenaars FJ et al. (2005) Multisensory integration of emotional faces and voices in schizophrenics. Schizophr Res 72: 195-203. doi:10.1016/j.schres.2004.02.013. PubMed: 15560964.
- 46. de Jong JJ, Hodiamont PP, de Gelder B (2010) Modality-specific attention and multisensory integration of emotions in schizophrenia: reduced regulatory effects. Schizophr Res 122: 136-143. doi:10.1016/j.schres.2010.04.010. PubMed: 20554159.
- 47. de Jong JJ, Hodiamont PP, Van den Stock J, de Gelder B (2009) Audiovisual emotion recognition in schizophrenia: reduced integration of facial and vocal affect. Schizophr Res 107: 286-293. doi:10.1016/j.schres.2008.10.001. PubMed: 18986799.
- 48. Dolan RJ, Morris JS, de Gelder B (2001) Crossmodal binding of fear in voice and face. Proc_Natl_Acad_Sci_U_S_A 98: 10006-10010. PubMed: 11493699.
- 49. Pourtois G, de Gelder B, Bol A, Crommelinck M (2005) Perception of facial expressions and voices and of their combination in the human brain. Cortex 41: 49-59. doi:10.1016/S0010-9452(08)70177-1. PubMed: 15633706.
- 50. Kreifelts B, Ethofer T, Grodd W, Erb M, Wildgruber D (2007) Audiovisual integration of emotional signals in voice and face: an event-related fMRI study. NeuroImage 37: 1445-1456. doi:10.1016/j.neuroimage.2007.06.020. PubMed: 17659885.
- 51. Izumi A, Kojima S (2004) Matching vocalizations to vocalizing faces in a chimpanzee (Pan troglodytes). Anim Cogn 7: 179-184. PubMed: 15015035.
- 52. Martinez L, Matsuzawa T (2009) Auditory-visual intermodal matching based on individual recognition in a chimpanzee (Pan troglodytes). Anim Cogn 12 Suppl 1: S71-S85. doi:10.1007/s10071-009-0269-1. PubMed: 19701656.
- 53. Parr LA (2004) Perceptual biases for multimodal cues in chimpanzee (Pan troglodytes) affect recognition. Anim Cogn 7: 171-178. PubMed: 14997361.