Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The resolution of face perception varies systematically across the visual field

  • Anisa Y. Morsi ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    annie.morsi.18@ucl.ac.uk

    Affiliation Experimental Psychology, University College London, London, United Kingdom

  • Valérie Goffaux,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliations Psychological Sciences Research Institute, UCLouvain, Ottignies-Louvain-la-Neuve, Belgium, Institute of Neuroscience, UCLouvain, Ottignies-Louvain-la-Neuve, Belgium

  • John A. Greenwood

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Resources, Software, Supervision, Visualization, Writing – review & editing

    Affiliation Experimental Psychology, University College London, London, United Kingdom

Abstract

Visual abilities tend to vary predictably across the visual field–for simple low-level stimuli, visibility is better along the horizontal vs. vertical meridian and in the lower vs. upper visual field. In contrast, face perception abilities have been reported to show either distinct or entirely idiosyncratic patterns of variation in peripheral vision, suggesting a dissociation between the spatial properties of low- and higher-level vision. To assess this link more clearly, we extended methods used in low-level vision to develop an acuity test for face perception, measuring the smallest size at which facial gender can be reliably judged in peripheral vision. In 3 experiments, we show the characteristic inversion effect, with better acuity for upright faces than inverted, demonstrating the engagement of high-level face-selective processes in peripheral vision. We also observe a clear advantage for gender acuity on the horizontal vs. vertical meridian and a smaller-but-consistent lower- vs. upper-field advantage. These visual field variations match those of low-level vision, indicating that higher-level face processing abilities either inherit or actively maintain the characteristic patterns of spatial selectivity found in early vision. The commonality of these spatial variations throughout the visual hierarchy means that the location of faces in our visual field systematically influences our perception of them.

Introduction

Vision varies across the visual field. The recognition of simple low-level stimuli (ranging from lines to letters) varies systematically, becoming worse with increasing distance from fixation [1] and at specific angular locations around fixation [2]. In contrast, the perception of high-level stimuli such as faces has been found to vary across the visual field in a unique or idiosyncratic fashion [35]. These distinct variations are consistent with the view that faces are “special” in the visual system [6, 7], and suggest that variations in high-level processing may arise independently from those of low-level vision. Comparison of these low- and high-level visual abilities is difficult due to differences in the methodology used to measure them, however. Here, we aligned the methodology used to investigate anisotropies for low-level stimuli and faces by measuring the spatial resolution of face perception around the visual field.

Low-level properties such as visual acuity (spatial resolution) not only decline with eccentricity [1] but also vary by location with eccentricity held constant. Acuity is typically better along the horizontal meridian vs. the vertical [810] and in the lower vs. the upper visual field [8, 9]. These horizontal-vertical and upper-lower anisotropies consistently emerge for many elements of vision, including orientation discrimination and contrast sensitivity [1113], and have been linked with the retinotopic organisation of the visual system [14]. At the retinal level, the density of retinal ganglion cells is higher along the horizontal vs. vertical meridian [15, 16]. In early visual cortex (V1-V3), smaller population receptive fields (pRFs) have been found along the horizontal vs. vertical meridian and the lower vs. upper field, highlighting variations in visual field sampling [17, 18]. Higher cell densities and smaller pRF sizes have been associated with better acuity [19]. In this way, low-level vision is fundamentally influenced by location.

Variations in face recognition appear to differ substantially from those of low-level vision. Some studies report faster recognition of face gender in the upper vs. the lower field [5, 20]. Others report no significant horizontal-vertical or upper-lower differences in discriminating facial identity, instead observing an advantage for the left vs. right visual field [4], consistent with left hemifield biases in face perception [21, 22]. Finally, biases in the apparent gender, age, and identity of morphed faces have been found to vary across the visual field in an entirely idiosyncratic manner across participants [3, 23]. These distinct and/or idiosyncratic patterns suggest a dissociation in the mechanisms driving the visual field variations in low-level vision and face perception.

This dissociation may not be surprising given evidence that faces undergo distinct forms of processing [6]. Relative to other objects, face recognition is disproportionately impaired for upside-down vs. upright faces [24]. This inversion effect is driven by increased sensitivity to the spatial relationships between features (configural processing) within upright faces [7, 25, 26]. Neuroimaging has identified ventral occipitotemporal brain regions dedicated to face processing, including the fusiform face area (FFA) which shows greater activation for upright vs. inverted faces [27, 28]. These higher-level face-selective regions nonetheless show retinotopic sensitivity, with smaller receptive fields in the fovea vs. the periphery [29, 30]. Given the dissociations between low-level vision and face perception however, it is unclear how this selectivity is linked to that of earlier stages in the visual hierarchy.

A major challenge in comparing variations in low- and high-level vision derives from methodological differences. While measurements of low-level vision focus on spatial properties such as acuity, face recognition is usually measured via appearance-based judgements of fixed-size faces [35]. To align these approaches, we developed an acuity test for faces, measuring the smallest size necessary to judge gender at each visual field location. If face processing systems share the spatial properties of early visual cortex, anisotropies similar to those found for low-level vision should emerge for gender acuity. This pattern could arise either because face recognition systems inherit these variations from earlier stages or because face-selective brain regions actively maintain the same anisotropies. Alternatively, gender acuity could show either idiosyncratic [3, 23] or systematic patterns of variation that are unlike those of low-level vision [4, 5]. The latter outcomes would suggest that face perception involves distinct mechanisms that do not inherit the spatial selectivity of earlier brain regions. To determine the engagement of face-selective processes in our task, we measured gender acuity with both upright and inverted faces. If the task were solely limited by low-level acuity, the recognition of upright and inverted faces should not differ. Face-selective processes would instead be revealed by the characteristic advantage for upright faces [24].

Experiment 1

Method

Participants.

14 participants (13 female, 1 male, Mage = 24.9 years) took part, including authors AYM and JAG; the rest were naïve. All had normal or corrected-to-normal foveal vision of at least 20/20, assessed using a Snellen chart. 9 were right-eye dominant, determined using the Crider ring test [31]. This sample size was derived from previous studies with similar designs [11]. All three experiments were approved by the Research Ethics Committee for Experimental Psychology at University College London and all participants gave written informed consent before testing began.

Apparatus.

The experiment was programmed in MATLAB (MathWorks, Inc) and conducted on an Apple iMac running PsychToolbox [3234]. Stimuli were viewed binocularly on a Cambridge Research Systems Display++ monitor with 2560 x 1440 resolution and 120 Hz refresh rate. The monitor was gamma corrected and linearised through software to have a minimum luminance of 0.16 cd/m2 and a maximum of 143 cd/m2. Participants were seated at a 50cm viewing distance, with head movements minimised using forehead and chin rests. The experiment took place in a dark room, and responses were recorded with a keypad.

Stimuli.

8 male and 8 female faces were selected from a bank of faces created by Laguesse and colleagues [35]. Because our task involved a binary gender judgement, we sought to avoid ambiguity by selecting faces at each end of the gender spectrum. This was ensured by taking faces that had received ratings above 8 out of 10 for either maleness or femaleness in a separate study (where faces were presented in a similar fashion to the current study–greyscale and within an oval aperture). All faces were grayscale, front-facing, and had a neutral expression (Fig 1B). Upright and inverted faces were included to determine whether face-specific configural mechanisms were engaged, and to assess whether visual field variations would differ with inversion. Using Adobe Photoshop CS6, each face was edited into an egg-shaped aperture measuring 657 x 877 pixels (at its widest and highest point, respectively) so that the only differences between images were due to internal features and not outer face shape (e.g. jawline). Pilot testing was conducted to ensure that gender-recognition performance was broadly similar across the face set (to minimise the possibility that distinctive features in individual faces could drive performance). The faces were set to have the same mean luminance as the monitor, with matched root-mean square (RMS) contrast values of 0.68. This ensured that overall luminance or contrast values could not be used as cues to gender.

thumbnail
Fig 1.

(A) The 8 polar angles tested, starting with 0° in the right visual field and preceding counterclockwise in 45° increments. The horizontal and vertical meridians are represented by pink and purple dashed lines, respectively. (B) Examples of female (left) and male (right) faces. (C) Experimental paradigm. A Gaussian fixation point first appeared, then a face was presented for 500 ms at one of the eight possible locations, selected randomly (shown here at 180°). Each face was followed by a mask which remained on screen until a keyboard response was made. Faces varied in size from trial to trial according to an adaptive QUEST+ procedure.

https://doi.org/10.1371/journal.pone.0303400.g001

Procedure.

Participants were instructed to fixate on a white two-dimensional Gaussian element (standard deviation of 13.8 minutes of arc) in the centre of the screen. During each trial (Fig 1C), a face was presented for 500 ms, with the image centre located at 10° eccentricity and at one of 8 possible angles (Fig 1A). The face was immediately followed by a 1/f noise egg-shaped mask, which broadly matches the spatial frequency content of faces and natural scenes [36]. The size of the mask varied trial-by-trial to match the size of the face just shown, remaining on screen until participants made their response. We used a single interval 2-alternative forced choice (2AFC) response method, with participants reporting the face as either male or female using a numeric keypad. Audio feedback was provided after each response.

Before experimental trials began, participants completed a shorter set of 72 practice trials to become accustomed to the task. For the practice trials, faces were presented at fixed sizes of 600, 400 and 200 pixels (face size refers to vertical height, with width scaled proportionately), with 9 trials at each location. Participants were required to be at least 90% correct in order to continue. During the experimental trials an adaptive QUEST procedure [37] varied face sizes presented at each location according to the participant’s responses, set to converge on the size at which 75% of responses were correct. Within each block of trials, QUEST estimates were computed separately for each location. Faces were presented at sizes within ±1/3 of the QUEST threshold estimate on each trial (minimum 5 and maximum 640 pixels). This “jitter” allowed us to collect data for a range of sizes, which improved the subsequent fitting of psychometric functions to the data [38].

Each experimental block contained 50 faces shown at each of the 8 locations (with independent QUEST procedures) to give 400 trials in total. Each face was shown an equal number of times, in a randomised order, with the location it appeared at also randomised. Upright and inverted faces were presented in alternate blocks. The experiment consisted of 1–2 practice blocks, followed by 8 experimental blocks (4 repeats for both upright and inverted faces) to give 3200 trials in total. During analysis, we fit psychometric functions to the combined data from these 4 repeats (separately for each location and inversion condition).

Analyses.

Responses were first sorted by face size (in pixels) and collated in 20-pixel bins (e.g. faces of 8, 15 and 18 pixels would fall in the same bin). The proportion of correct responses was then calculated for each face-size bin. Cumulative Gaussian functions (Fig 2) were fit to these data using 3 free parameters for the mean, variance and lapse rate [39]. The lapse rate was set to a maximum of 0 by default. For some participants whose responses did not reach ceiling at the largest face sizes, the maximum allowable lapse rate was increased first to 0.05 and then 0.1 in order to improve curve fitting (required for 7 participants in Experiment 1, 3 in Experiment 2, and none in Experiment 3). Importantly, this was applied on an individual basis (equally across all conditions), meaning that within-subjects variations across locations or conditions cannot be attributed to this factor. Gender acuity thresholds for each location were taken as the size at which 75% accuracy was reached, then converted from pixels to degrees of visual angle.

thumbnail
Fig 2. Psychometric functions for a single participant, showing the proportion of correct gender judgements for different sized upright faces at each of the 8 visual field locations (labelled at the top of each graph, where 0 = rightwards, 90 = upwards, etc.).

Performance improves monotonically as a function of face size. Dashed grey lines plot thresholds for gender acuity (the size at which 75% accuracy was reached).

https://doi.org/10.1371/journal.pone.0303400.g002

Statistical analyses were carried out using a 3-way mixed effects analysis of variance (ANOVA), with participant as a between-subjects random factor and inversion (upright, inverted) and location (0, 35, 90, 135, 180, 225, 270, 315°) as within-subjects fixed factors. A priori comparisons took the form of repeated-measures t-tests, comparing thresholds between the horizontal (0 and 180°) vs. vertical (90 and 270°) meridians, the upper (90°) and lower (270°) field, and the left (180°) and right (0°) locations, for upright and inverted faces separately.

Results

Mean gender acuity thresholds across participants are plotted as both a bar chart and according to the polar angle of each of the 8 locations in Fig 3. Smaller values represent better gender acuity. Mean gender acuity thresholds were worse for inverted compared to upright faces overall, indicating that inversion disrupted the ability to judge gender at all locations. There was a sizeable difference in gender acuity according to location; for upright faces, there was a range of almost 2° of visual angle between the smallest threshold value in the lower field (270°) and the largest in the upper field (90°). Thresholds were smaller along the horizontal (0, 180°) as opposed to the vertical (90, 270°) meridian. Thresholds at the diagonal locations varied inconsistently, with smaller thresholds in the top left (135°) vs. top right (45°) location, but larger thresholds in the bottom left (225°) vs. bottom right (315°) location.

thumbnail
Fig 3.

Mean gender acuity thresholds (in degrees of visual angle) measured in Experiment 1, plotted in two ways. Firstly, as a bar graph (A) with each angular location indicated via colour (see legend). Individual data points represent thresholds for each participant. Secondly, as a polar angle plot (B), where 0° is at the right and angles proceed counterclockwise in 45° jumps. Upright faces are shown in purple and inverted in pink. Shaded regions denote ± 1 SEM.

https://doi.org/10.1371/journal.pone.0303400.g003

The ANOVA revealed a main effect of location, F(7,91) = 3.41, p = .003, d = 0.21, confirming that the location of faces in the visual field influenced gender perception. Planned contrasts revealed that thresholds were significantly smaller along the horizontal (0° and 180° averaged) compared to the vertical (90° and 270° averaged) meridian for both upright, t(13) = -2.84, p = .014, and inverted faces, t(13) = -2.21, p = .046. Thresholds were also smaller in the lower compared to the upper field for upright faces, t(13) = 2.68, p = .019, although not for inverted faces, t(13) = -0.10, p = .923. There was no difference between thresholds at the left and right locations for upright, t(13) = -1.79, p = .096, or inverted faces, t(13) = -0.61, p = .551. In other words, we observe both horizontal-vertical and upper-lower anisotropies for gender acuity, though performance did not differ between left and right hemifields.

The error bars in Fig 3 indicate that there was considerable between-participants variability in gender acuity thresholds, with the ANOVA showing a main effect of participant, F(7,91) = 2.98, p = .029, d = 0.75. However, there was no interaction between location and participant, F(91,91) = 1.07, p = .383, d = 0.52, indicating that individuals varied in their overall threshold magnitude rather than exhibiting idiosyncratic variations across the 8 locations. We note in particular that a subset of participants showed thresholds that were considerably smaller (with values ~0.5°) than the rest of the group.

The presence of an inversion effect was supported by a main effect of inversion, F(1,13) = 6.61, p = .023, d = 0.34, showing that thresholds were significantly higher for inverted compared to upright faces. Therefore, the processing of configural information in upright faces appears to have benefitted performance in our task. There was no interaction between inversion and location, F(7,91) = 0.91, p = .506, d = 0.07, indicating that inversion disrupted gender perception to a similar extent across the visual field. There was, however, a significant interaction between inversion and participant, F(13,91) = 10.55, p = < .001, d = 0.60. To investigate this interaction further, we calculated mean face inversion effect (FIE) values across participants by subtracting upright from inverted thresholds. Upon closer analysis the inversion-participant interaction appeared to be driven by two participants with very large FIEs, and indeed we found that removing their data from the analysis eliminated the interaction.

Experiment 2

Experiment 1 demonstrates that face recognition acuity varies in the same way as low-level vision. We had two concerns, however. First, the low thresholds of some participants (as small as 0.5°) suggests they may have occasionally fixated the faces. Second, the eyes of our upright face stimuli were closer to fixation in the lower vs. upper field, which could have driven the upper-lower difference, given the importance of the eyes in gender perception [40, 41]. In Experiment 2 we sought to validate our findings by adding eye-tracking and fixing eye position across locations.

Method

Participants.

14 participants (12 female, 2 male, Mage = 23.6 years) took part, including author AYM; the rest were naïve and newly recruited. All had normal or corrected-to-normal vision. 7 were right-eye dominant.

Apparatus.

In addition to the setup in Experiment 1, we used an EyeLink 1000 (SR Research, Mississauga, ON, Canada) to monitor fixation during trials.

Stimuli.

The same face stimuli from Experiment 1 were used. Stimulus locations were centred on the eyes themselves, so that the centre of the eyes was always 10° from fixation regardless of face size, angular location or inversion (Fig 4). The position of each face in the egg aperture was also shifted and/or rotated slightly to ensure that the eyes fell at the same point within the aperture.

thumbnail
Fig 4.

In Experiment 1 (left panels), faces were presented 10° from fixation according to the centre of the face. In Experiment 2 (right panels), the centre of the eyes was always 10° from fixation regardless of face size, angular location or inversion.

https://doi.org/10.1371/journal.pone.0303400.g004

Procedure.

Following calibration of the EyeLink to track their left eye, participants were required to fixate the Gaussian element (with an allowable error of 1.5° radius) in order for each trial to start. Trials in which fixation diverged from this region were cancelled and repeated at the end of the block. Participants first completed the practice block(s), as in Experiment 1. Experimental blocks were split according to whether faces were upright or inverted and locations were cardinal (0, 90, 180, 270°) or diagonal (45, 135, 225, 315°) angles, resulting in 4 blocks: upright cardinal, upright diagonal, inverted cardinal and inverted diagonal. These split blocks were introduced so that the blocks would not be too long, as the eye tracking increased the duration of data collection depending on the amount of cancelled trials and recalibration needed. Data were collected over 4 hour-long sessions, with each of the 4 conditions repeated once per session. This gave a total of 16 blocks and 4096 trials per participant.

Results

Mean gender acuity thresholds are plotted in Fig 5. Again, smaller values represent better gender acuity. Compared to Experiment 1, thresholds in Experiment 2 were higher overall and had reduced variability across participants (particularly in the inverted condition), suggesting that eye tracking successfully stopped participants from looking directly at faces. Indeed, the smallest-measured threshold in Experiment 2 was 1.31°, compared with 0.27° in Experiment 1.

thumbnail
Fig 5.

Mean gender acuity thresholds from Experiment 2, first shown as a bar graph (A) with each location indicated via colour (see legend). Individual data points represent thresholds for each participant. Mean thresholds are also visualised in a polar angle plot (B), with 0° at the right and angles increasing counterclockwise by 45° each time. Upright faces are shown in purple and inverted in pink. Shaded regions represent ± 1 SEM. (C) Bar charts comparing the horizontal-vertical difference and upper-lower difference in Experiments 1 and 2. Data are plotted for upright faces only. Horizontal refers to thresholds averaged across 0° and 180° locations, with vertical the average of 90° and 270°. Upper represents 90° and lower 270°. Individual data points represent thresholds for each participant. Significant differences (p < .05) are marked with an asterisk.

https://doi.org/10.1371/journal.pone.0303400.g005

The ANOVA revealed a main effect of location, F(7,91) = 5.55, p = < .001, d = 0.30, indicating that gender acuity was influenced by the location of faces. There was a clear horizontal-vertical difference, with planned comparisons revealing that thresholds were significantly smaller along the horizontal (0° and 180° averaged) compared to the vertical (90° and 270° averaged) meridian, for both upright, t(13) = -6.16, p < .001, and inverted faces, t(13) = -3.00, p = .010. However, although thresholds were smaller in the lower compared to the upper field, the difference was not significant for either upright, t(13) = 1.19, p = .256, or inverted faces, t(13) = 0.38, p = .713. Similarly, thresholds did not differ between the left and right locations for either upright, t(13) = -0.10, p = .926, or inverted faces, t(13) = 0.39, p = .704.

The engagement of face-selective processes can again be seen with the inversion effect in the mean data (Fig 5A and 5B), with higher thresholds and therefore reduced ability to perceive gender (i.e. larger faces needed) for inverted compared to upright faces overall. This was confirmed by a significant main effect of inversion, F(1,13) = 17.93, p = .001, d = 0.58. There was no interaction between inversion and location, F(7,91) = 1.29, p = .265, d = 0.09, indicating that inversion effects did not differ significantly across the visual field.

Gender recognition abilities again differed between individuals, which was confirmed by a main effect of participant, F(7,91) = 11.52, p < .001, d = 0.93. There was however no interaction between location and participant, F(91,91) = 0.94, p = .623, d = 0.48, indicating that location-based variations in gender perception are not specific to the individual. In other words, face perception differed across the visual field in a characteristic pattern shared across individuals.

Unlike the previous experiment there was no interaction between inversion and participant, F(13,91) = 1.60, p = .100, d = 0.19, indicating that individuals did not vary substantially in the inversion effect. This suggests that the significant interaction in Experiment 1 may have been caused by a subset of participants looking at the faces–accordingly, these individuals had low thresholds for both upright and inverted faces (i.e. little to no inversion effect), likely driven by their fixating the faces in both conditions.

To compare anisotropies more clearly, bar charts displaying the horizontal-vertical difference and upper-lower difference in both experiments are shown in Fig 5C. Only data for upright faces are included. The charts on the left-hand side show that the horizontal-vertical difference was consistent across both experiments, with significantly lower thresholds (better gender acuity) for faces at horizontal compared to vertical locations. The upper-lower difference was only significant in Experiment 1, where effects could have been driven by eye movements or variations in eye position within face stimuli. Although a trend in the same direction persisted when we controlled for these factors in Experiment 2 –showing that the upper-lower difference cannot be attributed to these factors alone–the difference was no longer significant.

To summarise, gender acuity was better along the horizontal vs. vertical meridian but did not differ significantly in the lower vs. upper field. However, even in Experiment 1 the upper-lower difference was smaller than the horizontal-vertical difference, suggesting that it may simply be harder to measure. We examined this possibility in Experiment 3.

Experiment 3

Given the trend towards better gender acuity in the lower vs. upper field in Experiment 2, we conducted further measurements at these locations to determine whether a significant difference would emerge with a greater number of trials.

Method

Participants.

14 participants (11 female, 2 male, 1 non-binary, Mage = 22.1 years) took part, including author AYM; the rest were naïve and newly recruited. All had normal or corrected-to-normal vision. 7 were right-eye dominant.

Stimuli.

Stimuli were as in Experiment 2, with faces shown at the upper (90°) and lower (270°) locations only.

Procedure.

Blocks were split according to whether faces were upright or inverted, with 128 trials in each block. Data were collected over 2 hour-long testing sessions, with each of the 2 conditions (upright/inverted) repeated 8 times per session (with an extra block completed at the start of the first session, which acted as a practice block and was not included in data analysis). This gave a total of 16 experimental blocks and 2048 trials over the experiment (doubling the number of trials per location compared to Experiment 2). Remaining parameters were as in Experiment 2.

Results

Mean gender acuity thresholds are plotted in Fig 6, with smaller values representing better gender acuity. The ANOVA revealed a main effect of location, F(1,13) = 22.97, p < .001, d = 0.91, indicating that gender acuity differed between the upper and lower fields. Gender acuity thresholds were significantly smaller in the lower field compared to the upper for upright faces, t(13) = 3.82, p = .002, and approached significance for inverted faces, t(13) = 2.07, p = .059. On an individual level, 12 of 14 participants showed better acuity in the lower vs. the upper field. These results highlight the presence of an upper-lower difference in face recognition, with better gender acuity in the lower half of the visual field.

thumbnail
Fig 6. Mean gender acuity thresholds for the upper (90°; lighter blue) versus lower (270°; darker blue) visual field, for both upright and inverted faces.

The asterisk represents a significant difference (p < .05). Individual data points represent thresholds for each participant, with lines connecting their performance between upper and lower locations.

https://doi.org/10.1371/journal.pone.0303400.g006

Like the previous two experiments, there was a main effect of orientation, F(1,13) = 8.38, p = .013, d = 0.39, indicating an overall inversion effect whereby gender acuity thresholds were larger for inverted compared to upright faces, t(27) = -2.92, p = .007. There was also a main effect of participant, F(1,13) = 31.12, p = .032, d = 1.00, again highlighting overall differences in gender acuity between individuals. There was no significant interaction between location and participant, F(1,13) = 0.58, p = .829, d = 0.37, suggesting that there was a common pattern of gender acuity across individuals. Interactions were similarly non-significant for location and orientation, F(1,13) = 0.91, p = .358, d = 0.07, and orientation and participant, F(1,13) = 1.02, p = .484, d = 0.51, indicating that gender acuity patterns were similar for upright and inverted faces, and that inversion effects did not vary significantly between individuals.

Discussion

We demonstrate that face perception varies across the visual field in a systematic pattern. Across 3 experiments, acuity for gender judgements showed a horizontal-vertical anisotropy for both upright and inverted faces, where recognition was possible with smaller faces on the horizontal vs. vertical meridian, and a small-but-reliable upper-lower difference, with better acuity in the lower vs. upper field. The presence of both these anisotropies, and the smaller magnitude of the upper-lower difference, matches the patterns of low-level vision [8, 11, 12, 42] and demonstrates that the resolution of face perception varies predictably across the visual field, rather than uniquely or idiosyncratically [35]. This suggests that spatial properties are preserved throughout the visual hierarchy, including in higher-level face-selective systems.

Our observation of systematic variations in face recognition differs from prior studies reporting purely idiosyncratic variations around the visual field. Prior studies demonstrating these idiosyncrasies have relied on judgements of appearance [3, 23]. Judgements of apparent size [43] and position [44] have similarly revealed perceptual idiosyncrasies in low-level vision, likely related to localised visual-field distortions. These individual-dependent distortions could in fact drive idiosyncrasies in facial-identity judgements (e.g. by altering facial features), given that the idiosyncrasies for one set of faces do not correlate with those for other faces [23]. The use of a single male/female pair to measure gender judgements in these prior studies [3] would be similarly susceptible to individual-based distortions. In contrast, our measurement of performance with multiple unambiguously gendered faces is less susceptible to smaller individual-dependent variations, allowing us to uncover larger systematic anisotropies.

Our results also differ from the systematic-but-unique variations found for faces previously. For instance, judgements of the identity of synthetic-contour faces show a left visual-field advantage [4], whereas we found no difference between the left and right fields. Our finding is nonetheless consistent with the broader observation that face perception is not reliably lateralised, but rather that these effects depend on the task [4548] and/or stimuli, with the clearest effects emerging for large, chimeric faces that span fixation [22, 49]. Additionally, whereas we observe better gender acuity in the lower vs. upper field, prior studies report faster gender recognition in the upper field [5, 20]. This difference likely reflects our measurement of performance rather than reaction times, given that an upper-field advantage is also evident in the temporal response to low-level stimuli [50, 51]. Altogether, our findings show that when the spatial resolution of face perception is measured in the same way as lower-level visual abilities, similar visual field anisotropies emerge.

In Experiments 2 and 3, the upper-lower difference was reduced (relative to Experiment 1) when the position of the eyes was matched across locations, confirming a particular importance of the eyes for gender perception [40, 41]. Across all experiments, acuity thresholds were nonetheless consistently lower for upright vs. inverted faces. This inversion effect confirms that our gender acuity task sufficiently engaged face-specific mechanisms, consistent with prior reports that the configural processing of upright faces occurs across both foveal and peripheral vision [46, 47, 52]. Interestingly, while inversion effects are typically measured using same-size faces [24], here we highlight a spatial component to face processing [30, 60]. This ability to identify upright faces at smaller sizes than inverted faces could reflect the added benefit of configural processing with upright faces [6]. Importantly, this inversion effect indicates that the thresholds for gender acuity that we measure cannot be solely attributed to lower-level limitations on spatial vision such as visual acuity and contrast sensitivity, nor to the recognition of individual facial features such as the eyes.

Nevertheless, the anisotropies that we observe for face perception could be partly driven by variations in featural selectivity, including properties like the radial bias, whereby peripheral contrast sensitivity and orientation discrimination are better for stimuli oriented towards the fovea [53, 54]. Because horizontal information is particularly informative for face recognition [5557], this could improve performance on the horizontal vs. vertical meridian–either passively through the pooling of variations in low-level information, or more actively by boosting the response to the optimal orientations for high-level processing, as argued recently [58]. However, though these factors could contribute to the horizontal-vertical anisotropy, these radial variations are matched across the upper and lower fields, making them unable to explain the upper-lower anisotropy.

The systematic anisotropies we observe for face recognition could also arise from variations in spatial selectivity within face-selective systems. In low-level vision, better acuity has been linked to smaller receptive fields and increased cortical magnification [1719, 59]. This could drive improved acuity for faces because higher levels passively inherit enhanced low-level input from locations on the horizontal (vs. vertical) meridian, and in the lower (vs. upper) field. Unlike low-level vision, however, better face perception has been linked with larger pRF sizes and the resulting increase in visual field coverage within face-selective regions [30, 60]. The observed anisotropies could thus derive instead from the way that face-selective neurons actively sample the visual field. Either way, the common pattern of anisotropies in low- and high-level vision suggests that retinotopic sensitivity within face-selective regions is not entirely distinct from that of earlier brain areas.

Given these possibilities for both the passive inheritance and active maintenance of featural and/or spatial selectivity, to what extent are face processing systems themselves contributing to the observed anisotropies? Note first that the presence of a consistent inversion effect indicates that performance on our gender acuity task derives from processing within face-selective systems (at least in part), rather than stemming purely from lower-level limitations on vision. However, the size of this inversion effect did not vary significantly around the visual field, as one might expect if face-selective processes were themselves to vary, and contrary to the recent finding of larger inversion effects in an identity-recognition task on the horizontal meridian than the vertical [58]. It could be that while the strength of face-selective processing varies around the visual field (leading to variable inversion effects with fixed-size stimuli), the resolution of these face-selective processes does not. We nonetheless observed similar anisotropies for both upright and inverted faces. These acuity thresholds are many times larger than those for the identification of simple elements–compared to data from Anstis [61], thresholds in Experiment 2 were approximately 8–13 times larger for upright faces, and 10–16 times larger for inverted faces. We suggest that face processing systems sample large regions of the visual field in order to support gender recognition, with greater efficiency for upright faces than inverted (and a requirement for broader sampling of information with inverted faces, given that inverted faces required larger stimulus sizes to reach threshold) and variations in sampling for both processes around the visual field. The spatial pooling required to extract holistic and/or configural information does not however appear to vary across the visual field.

In conclusion, face perception varies around the visual field with both horizontal-vertical and upper-lower anisotropies, matching patterns consistently found for low-level vision [8, 11, 12], and contrary to suggestions that face recognition varies in a unique or idiosyncratic manner [35]. Our results are consistent with a hierarchical model of face processing whereby the selectivity for faces is built on the selectivity of earlier levels. The variations in face perception that we observe are likely driven by variations in spatial selectivity, and perhaps in part by variations in featural selectivity, which may be inherited passively from earlier stages and/or actively maintained in face-selective regions. Ultimately, we demonstrate that common spatial variations are found throughout the visual system, causing location to systematically influence face perception.

Acknowledgments

We thank Alexia Roux-Sibilon at UCLouvain for supplying us with the face images.

References

  1. 1. Rosenholtz R. Capabilities and Limitations of Peripheral Vision. Annual Review of Vision Science. 2016;2:437–57. pmid:28532349
  2. 2. Himmelberg MM, Winawer J, Carrasco M. Polar angle asymmetries in visual perception and neural architecture. Trends Neurosci. 2023;46(6):445–58. pmid:37031051
  3. 3. Afraz A, Pashkam MV, Cavanagh P. Spatial heterogeneity in the perception of face and form attributes. Current Biology: Elsevier Ltd; 2010. p. 2112–6. pmid:21109440
  4. 4. Schmidtmann G, Logan AJ, Kennedy GJ, Gordon GE, Loffler G. Distinct lower visual field preference for object shape. Journal of Vision. 2015;15(5):18. pmid:26067536
  5. 5. Quek GL, Finkbeiner M. The upper-hemifield advantage for masked face processing: Not just an attentional bias. Attention, Perception, and Psychophysics2016. p. 52–68.
  6. 6. Robbins R, McKone E. No face-like processing for objects-of-expertise in three behavioural tasks. Cognition. 2007;103(1):34–79. pmid:16616910
  7. 7. Rossion B. Picture-plane inversion leads to qualitative changes of face perception. Acta Psychol (Amst). 2008;128(2):274–89. pmid:18396260
  8. 8. Carrasco M, Talgar CP, Cameron EL. Characterizing visual performance fields: effects of transient covert attention, spatial frequency, eccentricity, task and set size. Spat Vis. 2001;15(1):61–75. pmid:11893125
  9. 9. Greenwood JA, Szinte M, Sayim B, Cavanagh P. Variations in crowding, saccadic precision, and spatial localization reveal the shared topology of spatial vision. Proceedings of the National Academy of Sciences. 2017;114(17):E3573–E82. pmid:28396415
  10. 10. Westheimer G. Anisotropies in peripheral vernier acuity. Spat Vis. 2005;18(2):159–67. pmid:15856934
  11. 11. Abrams J, Nizam A, Carrasco M. Isoeccentric locations are not equivalent: the extent of the vertical meridian asymmetry. Vision Res. 2012;52(1):70–8. pmid:22086075
  12. 12. Barbot A, Xue S, Carrasco M. Asymmetries in visual acuity around the visual field. Journal of Vision. 2021;21(1):2. pmid:33393963
  13. 13. Himmelberg MM, Winawer J, Carrasco M. Stimulus-dependent contrast sensitivity asymmetries around the visual field. Journal of Vision. 2020;20(9):18. pmid:32986805
  14. 14. Wandell BA, Dumoulin SO, Brewer AA. Visual field maps in human cortex. Neuron. 2007;56(2):366–83. pmid:17964252
  15. 15. Curcio CA, Allen KA. Topography of ganglion cells in human retina. Journal of Comparative Neurology1990. p. 5–25. pmid:2229487
  16. 16. Perry VH, Cowey A. The ganglion cell and cone distributions in the monkey’s retina: Implications for central magnification factors. Vision Research1985. p. 1795–810. pmid:3832605
  17. 17. Silson EH, Reynolds RC, Kravitz DJ, Baker CI. Differential sampling of visual space in ventral and dorsal early visual cortex. Journal of Neuroscience2018. p. 2294–303. pmid:29382711
  18. 18. Silva MF, Brascamp JW, Ferreira S, Castelo-Branco M, Dumoulin SO, Harvey BM. Radial asymmetries in population receptive field size and cortical magnification factor in early visual cortex. NeuroImage2018. p. 41–52. pmid:29155078
  19. 19. Duncan RO, Boynton GM. Cortical magnification within human primary visual cortex correlates with acuity thresholds. Neuron2003. p. 659–71. pmid:12765616
  20. 20. Quek GL, Finkbeiner M. Face-sex categorization is better above fixation than below: Evidence from the reach-to-touch paradigm. Cognitive, Affective, & Behavioral Neuroscience. 2014;14:1407–19. pmid:24763922
  21. 21. Ellis HD, Shepherd JW. Recognition of upright and inverted faces presented in the left and right visual fields. Cortex. 1975;11(1):3–7. pmid:1149464
  22. 22. Harrison M, Strother L. Does face-selective cortex show a left visual field bias for centrally-viewed faces? Neuropsychologia. 2021;159(107956). pmid:34265343
  23. 23. Visconti di Oleggio Castello M, Taylor M, Cavanagh P, Gobbini MI. Idiosyncratic, Retinotopic Bias in Face Identification Modulated by Familiarity. eNeuro. 2018;5(5). pmid:30294669
  24. 24. Yin RK. Looking at upside-down faces. Journal of Experimental Psychology1969. p. 141–5.
  25. 25. Le Grand R, Mondloch CJ, Maurer D, Brent HP. Early visual experience and face processing. Nature. 2001;410(6831):890.
  26. 26. Maurer D, Grand RL, Mondloch CJ. The many faces of configural processing. Trends Cogn Sci. 2002;6(6):255–60. pmid:12039607
  27. 27. Yovel G, Kanwisher N. Face perception: Domain specific, not process specific. Neuron2004. p. 889–98.
  28. 28. Yovel G, Kanwisher N. The neural basis of the behavioral face-inversion effect. Current Biology2005. p. 2256–62. pmid:16360687
  29. 29. Kay KN, Weiner KS, Grill-Spector K. Attention reduces spatial uncertainty in human ventral temporal cortex. Curr Biol. 2015;25(5):595–600. pmid:25702580
  30. 30. Poltoratski S, Kay K, Finzi D, Grill-Spector K. Holistic face recognition is an emergent phenomenon of spatial processing in face-selective regions. Nature Communications. 2021;12(1):4745. pmid:34362883
  31. 31. Crider B. A Battery of Tests for the Dominant Eye. The Journal of General Psychology: Blake Crider; 1944. p. 179–90.
  32. 32. Kleiner M, Brainard D, Pelli D. What’s new in Psychtoolbox-3? Perception. 2007;36:14-.
  33. 33. Pelli DG. The VideoToolbox software for visual psychophysics: Transforming numbers into movies. Spatial vision: Brill; 1997. p. 437–42. pmid:9176953
  34. 34. Brainard DH. The Psychophysics Toolbox. Spat Vis. 1997;10(4):433–6. pmid:9176952
  35. 35. Laguesse R, Dormal G, Biervoye A, Keufner D, Rossion B. Extensive visual training in adulthood significantly reduces the face inversion effect. Journal of Vision. 2012;12(4). pmid:23019119
  36. 36. Párraga CA, Troscianko T, Tolhurst DJ. The human visual system is optimised for processing the spatial information in natural visual images. Current Biology: Cell Press; 2000. p. 35–8. pmid:10660301
  37. 37. Watson AB, Pelli DG. QUEST: a Bayesian adaptive psychometric method. Percept Psychophys. 1983;33(2):113–20. pmid:6844102
  38. 38. Tailor VK, Theodorou M, Dahlmann-Noor AH, Dekker TM, Greenwood JA. Eye movements elevate crowding in idiopathic infantile nystagmus syndrome. Journal of Vision. 2021;21(13):9-. pmid:34935877
  39. 39. Eccentric Vision Lab Github. [Internet]. 2023. https://www.github.com/eccentricvision.
  40. 40. Schyns PG, Bonnar L, Gosselin F. Show me the features! Understanding recognition from the use of visual information. Psychol Sci. 2002;13(5):402–9. pmid:12219805
  41. 41. Brown E, Perrett DI. What gives a face its gender? Perception. 1993;22(7):829–40. pmid:8115240
  42. 42. Kurzawski JW, Burchell A, Thapa D, Majaj NJ, Winawer JA, Pelli DG. An enhanced Bouma model fits a hundred people’s visual crowding. bioRxiv. 2021.
  43. 43. Moutsiana C, De Haas B, Papageorgiou A, Van Dijk JA, Balraj A, Greenwood JA, et al. Cortical idiosyncrasies predict the perception of object size. Nature Communications: Nature Publishing Group; 2016. pmid:27357864
  44. 44. Kosovicheva A, Whitney D. Stable individual signatures in object localization. Curr Biol. 2017;27(14):R700–R1. pmid:28743014
  45. 45. Bourne VJ, Vladeanu M, Hole GJ. Lateralised repetition priming for featurally and configurally manipulated familiar faces: evidence for differentially lateralised processing mechanisms. Laterality. 2009;14(3):287–99. pmid:18949655
  46. 46. Kovacs P, Knakker B, Hermann P, Kovacs G, Vidnyanszky Z. Face inversion reveals holistic processing of peripheral faces. Cortex. 2017;97:81–95. pmid:29096198
  47. 47. McKone E. Isolating the special component of face recognition: peripheral identification and a Mooney face. J Exp Psychol Learn Mem Cogn. 2004;30(1):181–97. pmid:14736306
  48. 48. Schweinberger SR, Baird LM, Blümler M, Kaufmann JM, Mohr B. Interhemispheric cooperation for face recognition but not for affective facial expressions. Neuropsychologia. 2003;41(4):407–14. pmid:12559158
  49. 49. Burt DM, Perrett DI. Perceptual asymmetries in judgements of facial attractiveness, age, gender, speech and expression. Neuropsychologia. 1997;35(5):685–93. pmid:9153031
  50. 50. Honda H, Findlay J. Saccades to targets in three-dimensional space: Dependence of saccadic latency on target location. Percept Psychophys. 1992;52:167–74. pmid:1508624
  51. 51. Abegg M, Pianezzi D, Barton JJ. A vertical asymmetry in saccades. Journal of Eye Movement Research. 2015;8(5).
  52. 52. Kalpadakis-Smith AV, Goffaux V, Greenwood JA. Crowding for faces is determined by visual (not holistic) similarity: Evidence from judgements of eye position. Sci Rep. 2018;8(1):12556. pmid:30135454
  53. 53. Westheimer G. The distribution of preferred orientations in the peripheral visual field. Vision Res. 2003;43(1):53–7. pmid:12505604
  54. 54. Rovamo J, Virsu V, Laurinen P, Hyvarinen L. Resolution of gratings oriented along and across meridians in peripheral vision. Invest Ophthalmol Vis Sci. 1982;23(5):666–70. pmid:7129811
  55. 55. Dakin SC, Watt . Biological “bar codes” in human faces. Journal of Vision. 2009;9(4):2-.
  56. 56. Goffaux V, Greenwood JA. The orientation selectivity of face identification. Sci Rep. 2016;6:34204. pmid:27677359
  57. 57. Goffaux V, Duecker F, Hausfeld L, Schiltz C, Goebel R. Horizontal tuning for faces originates in high-level Fusiform Face Area. Neuropsychologia. 2016;81:1–11. pmid:26683383
  58. 58. Roux-Sibilon A, Peyrin C, Greenwood JA, Goffaux V. Radial bias in face identification. Proceedings of the Royal Society B: Biological Sciences. 2023;290(2001):20231118.
  59. 59. Benson NC, Kupers ER, Barbot A, Carrasco M, Winawer J. Cortical magnification in human visual cortex parallels task performance around the visual field. eLife. 2021;10. pmid:34342581
  60. 60. Witthoft N, Poltoratski S, Nguyen M, Golarai G, Liberman A, LaRocque KF, et al. Reduced spatial integration in the ventral visual cortex underlies face recognition deficits in developmental prosopagnosia. BioRxiv. 2016;051102.
  61. 61. Anstis SM. Chart demonstrating variations in acuity with retinal position. Vision Res. 1974;14(7):589–92.