Handedness and Graspability Modify Shifts of Visuospatial Attention to Near-Hand Objects

We examined how factors related to the internal representation of the hands (handedness and grasping affordances) influence the distribution of visuospatial attention near the body. Left and right handed participants completed a covert visual cueing task, discriminating between two target shapes. In Experiment 1, participants responded with either their dominant or non-dominant hand. In Experiment 2, the non-responding hand was positioned below one of two target placeholders, aligned with the shoulder. In Experiment 3 the near-monitor hand was positioned under the placeholder in the opposite region of hemispace, crossed over the body midline. For Experiments 2 & 3, in blocked trials the palmar and back-of hand surfaces were directed towards the target placeholder such that targets appeared towards either the graspable or non-graspable space of the hand respectively. In Experiment 2, both left and right handers displayed larger accuracy cueing effects for targets near versus distant from the graspable space of the right hand. Right handers also displayed larger response time cueing effects for objects near the graspable versus non-graspable region of their dominant hand but not for their non-dominant hands. These effects were not evident for left-handers. In Experiment 3, for right handers, accuracy biases for near hand targets were still evident when the hand was crossed over the body midline, and reflected hand proximity but not functional orientation biases. These findings suggest that biased visuospatial attention enhances object identity discrimination near hands and that these effects are particularly enhanced for right-handers.


Introduction
A substantial body of research has shown changes to the distribution of visuospatial attention towards objects when they are near hands. Specifically, the location of our hands, their posture, and individual differences such as handedness all have demonstrable impacts on how visual attention is distributed to objects near the body [1,2,3,4,5,6,7,8]. In addition, how we plan to use our hands to act upon objects in our environment (action goals) impacts near-body visual processing [9,10]. Typically, measuring near-hand attention confounds the goal of the action with the proximity of the hand. For example, one hand often is adjacent to the display with the distant hand responding to the target [5,6,7,8] responding to targets [2,3,4]. By contrast, studies with non-manual response measures (i.e. saccade / foot-pedal) disambiguate the action goal from manual responses, but still do not assess directly the impact of the action goal on the effects found. As a result, it is difficult to determine which elements of the action system contribute to near-hand visuospatial attention in each case. The aim of the present research was to investigate systematically how the relationship between the goal of the action and two internal states of the action system: handedness and graspability, influence the distribution of attention to objects near the body.
Handedness provides an index of the internal representation of our primary effectors, the hands. It plays an important role in determining how we use our hands to interact with objects in our environment. Right-handers as a group tend to be strongly right lateralised and use their right hand for most unimanual tasks, whereas left-handers are heterogeneous in their degree of laterality overall and may use either hand depending on circumstance [11,12,13]. Importantly, left and right-handers differ in the neural representation of their dominant and non-dominant hands. The degree of dominance is a reflection of this representation [14,15]. Imaging studies have shown that the volume of motor cortex dedicated to the dominant hand is directly correlated with degree of handedness. In vivo recordings have shown that there is also greater synaptic connectivity in regions of motor cortex that represent the dominant hand compared with those that represent the non-dominant hand [14,15]. Thus, handedness provides a behavioural indicator of structural and functional variations in the brain that manifest in differences in a variety of other cognitive domains and it is for these reasons that left handers are often excluded from cognitive research [14,16,17].
Importantly, evidence suggests that the relationship between handedness and experience is bidirectional. Frequency and /or familiarity of hand use modulates the perceptual processing of near body stimuli. For example, handedness influences which hand we use to complete an action [16,18,19,20]. When we use one hand more often than the other, as we do with our dominant hand, we reinforce its representation over the non-dominant hand [14,15]. Because left-handers tend to be less lateralised than right-handers, sensorimotor representation of their dominant hand is not as substantial as in right-handers, but is also subject to changes relative to experience, such that if they use their non-dominant hand often, they strengthen its representation [15]. Moreover, behavioural evidence has shown that left-and right-handers have different kinematic profiles when using their dominant and non-dominant hand to grasp objects [20,21,22,23,24]. Further, evidence from neuro-imaging [25] confirms that such differences are subserved by different patterns of neural activation in left-and right-handers (see [26] for review).
There is recent evidence that handedness modulates basic elements of visual perception [27,28]. Namely, target detection accuracy is increased for objects in proximity to the hands. In one study, left-and right-handed participants completed a non-speeded discrimination task (responding to the identity of a left, right or centrally aligned stimulus) while their dominant hand, non-dominant hand, both hands or no hand was adjacent to the monitor [27]. Compared with the no-hand condition, right handers displayed greater visual sensitivity for stimuli near their dominant but not their non-dominant hand. By comparison, left handers displayed similar sensitivity for their dominant and non-dominant hand rather than a mirror pattern of the right handed participants [27].
Other intrinsic factors such as the functional properties of hands themselves (grasping capabilities) have also been shown to influence visuospatial attention [3,6,7,28]. For example, Abrams and colleagues [2] found that when hands were held to either side of a display (and responses made using display mounted response buttons) both spatial and temporal shifts in visual attention slowed compared with when responses were made by the hands distant from the display, even when hands were obscured from view. Similarly, Lloyd, et al., [6] used a covert exogenous cueing paradigm ( [29] see Fig 1) to investigate the influence of hand location on attention shifting. Right handed participants held either their dominant or non-dominant hand beneath one of two target placeholders and responded via foot-pedal to target identity. The cueing effect (difference between valid and invalid RTs) was larger for targets in the handadjacent versus hand-distant placeholder. When the participants' hands were crossed over the body midline and adjacent to the opposite hemispace placeholder, there were attentional biases only for targets near the right (dominant) hand and not the left. Also using an exogenous cueing task, Reed and colleagues [8] demonstrated visuospatial biases that were specific to hands and not just a result of an additive visual anchor provided by the hand. The participant's own hand, a non-hand visual anchor or a fake hand were placed adjacent to one of two potential target locations whilst the task was performed. Targets that appeared in the handadjacent location were detected faster than those equidistant from fixation but distant from the hand irrespective of cue validity. Moreover, this bias was not present for the visual anchor which indicates it was specific to representation of the hand [7].
Further evidence suggests that visuospatial attention is distributed near the body, relative to both the location and grasping affordances of hands. Specifically, visual objects are detected more rapidly when near the palmar 'grasping' surface of the hand compared with the back-ofhand [8,30]. To investigate this, Reed et al. [7] had participants hold their hand with either their palm or back-of-hand directed towards one of two potential target locations in a visual cueing task. Irrespective of cueing, participants were faster to detect targets appearing to their palm compared with the back-of-hand. This suggests that visuospatial attention was engaged more rapidly to the palmar (versus back) surface of the hand, because targets appearing in that location are more actable, thus reflecting an affordance bias in attention distribution [7]. Additionally, Thomas [31] found that when hands were adjacent to a visual display, precision grip postures enhanced observer performance on a form detection task (spatial sensitivity) whereas power grasps postures enhanced motion detection (temporal sensitivity).
The above findings qualify the results of earlier studies because they suggest that the orientation of the hand in relation to the target has a differential effect on the allocation of attentional resources [2,6]. In the studies conducted by Reed, et al. [7] and Abrams, et al., [2] participants directed the palmar surface of their hand towards the screen for all conditions in which hands were held near the display. By comparison, Lloyd, et al., [6] had participants direct the back-of the hand towards the target location. Thus we do not have a clear understanding of which attentional effects are attributable to mere hand proximity and which are related to the graspable properties of the objects. This is important because each has presented evidence that the location of the hand influences different mechanisms of visuospatial attention (engagement, shifting and spatial coding). For example, Abrams et al. [2] showed that both spatial and temporal shifts in visual attention were slowed amongst visual objects near the hands. Reed et al., [7,8] demonstrated speeded engagement and Lloyd et al., [6] found slower disengagement of attention for targets appearing near versus distant from the hand. Due to the different postures and response measures used in each, the functional orientation of the hand may contribute differently to each of these. It is also not possible to disambiguate whether the effects found by each reflect consistent visuospatial hand biases depending on task demands or the differential influence of the graspability of the targets. Moreover all of the described studies evaluated right-handed participants so it is unclear how intrinsic representation of the hands themselves may contribute to these effects.
This highlights another important consideration for paradigms which investigate hand proximity. All of the outlined intrinsic factors influence the distribution of visuospatial attention in one manner or another. Yet in examining how attention contributes to near-body visual processing, we also need to take into account the response demands of the tasks used to evaluate these (e.g., to press a key in response to the onset of a target). Extrinsic factors such as the goal of the action (e.g., to depress a key following the appearance of a target or to discriminate the identity of an object) also determine how visual attention is distributed within nearbody space. Neurophysiological and behavioural research has found that visual stimuli near the body are coded relative to action centred reference frames [9,10,32]. For example, Cosman and Vecera [33] found that proximity of one or both hands to a visual object resulted observers more often identifying objects as foreground figures irrespective of the presence of contextual cues suggesting the object was concave (i.e. not in the foreground). These and similar findings suggest that rather than just being delineated by mere proximity to the body, visual objects are prioritised perceptually based on how we may use our hands to act on them or within the space that they are situated.
Neurons in ventral premotor cortex selectively activate to visual stimuli relative to their proximity to the hands. These neurons display maximal activation when stimuli are on or near the hands, or critically, when they move towards them [9]. In line with this, behavioural evidence has shown that action goals modulate attentional selection. When the goal of an action is to reach from one point to another, visual inputs which fall within the zone between the start and goal location of the reach, are prioritised for attention [10]. The action itself defines the area of attentional prioritisation. For example, Tipper and colleagues [10,32] found that response times (RTs) were slower when distractors were presented between the hand and target compared with visual objects that were near the body but not within the frame of the action [10,32].
The aforementioned research employed either the non-adjacent hand or both hands to respond to visual objects. Thus we cannot dissociate effects attributable to the manual action goal from those attributable to internal representation of the hand, because they share a bidirectional relationship. We can examine the spatial relationship between the action goal and the proximity of the hands to the target stimulus, and how that relationship modulates visuospatial attention is distributed near the body. In the present study we did so by systematically varying the relationship between the goal of the action and the proximity of the hand to examine the added contributions of handedness, hand posture and proximity. In Experiment 1 we examined how handedness influences the distribution of visuospatial attention relative to the laterality of the response hand. Experiment 2 evaluated how the combination of hand proximity, posture and the laterality of the response hand influences the distribution of visuospatial attention relative to handedness. Experiment 3 spatially dissociated the response hand from the hand adjacent to the monitor to evaluate whether lateralised biases remap with the location of the hand in the opposite region of hemispace.

Experiment 1
The aim of Experiment 1 was to provide a measure of baseline visuospatial biases relative to handedness. We examined how the relationship between the laterality of the response hand (left versus right) and handedness (left-handed versus right-handed) influenced the distribution of attention within a visual display, to ascertain whether any response-biases were present. Left and right-handed participants completed a Posner [34] cueing task with predictable lateral cues responding to targets with either their dominant or non-dominant hand with both hands distant from the display. Because both hands were distant from the display, hand proximity and posture should not impact the pattern of results. Thus any resultant biases in visuospatial attention may be attributable to changes in visual attention based on handedness. If handedness influences the overall distribution of visuospatial attention, we expect responses to be faster to targets when responding with their dominant versus non-dominant hand irrespective of cue validity (main effect of response hand). Alternatively, if handedness influences shifts in visuospatial attention, we expect a greater cueing effect for targets aligned with the dominant hand, when participants responding with their dominant hand. Due to the greater degree of laterality exhibited by right handers it is also possible that right-handers will show more of an effect of response hand (dominant versus non-dominant) either via faster responses overall compared with left-handers when responding to targets with the dominant hand, or interacting with cueing to alter cueing effects.

Methods
Subjects. Forty-two undergraduate students (25 females; 21 left-handed and 21 righthanded by self-report) from the University of Queensland (mean age 20.48 years) completed the experiment in return for course credit. The School of Psychology Ethics Review Committee, University of Queensland, Australia approved the running of this study. Verbal consent was obtained from all participants, no written consent was required by the aforementioned committee because at no stage was any uniquely identifying information obtained from the participants. Instead after they gave consent to participate they were assigned a random numerical code to identify their data set which was not able to be linked to their identity.
Stimuli and procedure. Stimuli were created in Microsoft PowerPoint and presented using E-Prime 2.0 at a 50-cm viewing distance on 32-cm × 48-cm LCD colour monitors (resolution of 1,680 × 1050 pixels). Two blue placeholder rectangles (7.44˚× 5.73˚) with the far edge 24.27˚from fixation were presented in the bottom corners or the monitor on a black background either side of a white fixation cross (0.5˚). A white peripheral cue box (6.87˚× 5.15˚) appeared in one of the placeholders. Targets were solid two-dimensional yellow shapes: either a triangle or circle (1.14˚), that appeared in centre of one of the placeholders, 20.96f rom central fixation with spatial location (left or right) randomised across trials.
The fixation cross and placeholders appeared for 700-1000ms and remained on display when a 250ms duration cue appeared in either the left or the right placeholder. At cue offset, in replication of Lloyd et al., [6] there was a 250ms stimulus onset asynchrony (SOA). The target (yellow circle or triangle) appeared in either the same location as the cue (valid cue) or the opposite location (invalid cue). Target location was invalidly cued on 30% of trials (Fig 1a) and validly cued on 70% of trials (Fig 1b). Participants were instructed to respond as quickly and accurately as possible to the target identity (circle or triangle; equally probable) by clicking either the left or right button on the computer mouse with their index / middle finger respectively, counterbalanced across participants.
Participants completed the task in a quiet, dimly lit room with their eyes approximately 50cms from the monitor. A subset of left and right handed participants were randomly assigned to complete the task with either their left hand (11 left handed participants; 10 righthanded participants) or right hand (10 left handed participants; 11 right-handed participants) resting on the table, 40cms from the centre of the monitor. The hand not engaged in completing the task was rested in the participants lap.
Of the right handed participants who responded with their left hand; five responded to triangles with their index finger and circles with their middle finger and five with the reverse. Of the right handed participants who responded with their right hand; five responded to triangles with their index finger and circles with their middle finger and six with the reverse. Of the left handed participants who responded with their left hand; five responded to triangles with their index finger and circles with their middle finger and six completed the task with the reverse arrangement. Of the left handed participants who responded with their right hand; five responded to triangles with their index finger and circles with their middle finger and five with the reverse.

Results and Discussion
Anticipatory RTs < 150ms (0.29%) and non-stimulus driven RTs > 1000ms, (1.29%) were also excluded (see Table 1 for summary of mean RTs and accuracy for each level of each condition).
RT analysis. A mixed repeated measures ANOVA was conducted with the within subjects factor target location (left target; right target) and the between subjects factors of handedness  Fig 1A) represents the graspable posture, Fig 1B) represents the non-graspable posture. Each participant completed a block each with their monitoradjacent hand in the A and B posture, with the posture-order counterbalanced between participants. Error analysis. We conducted a mixed repeated measures ANOVA with the same factors as the RT analysis. Accuracy cueing effect served as the dependant variable. This was computed as the mean accuracy for validly cued targets minus the mean accuracy invalidly cued targets. The analysis revealed a main effect of target location F(1,38) = 7.79, p = .008, Z 2 p ¼ : 17. There was a larger cueing effect for left versus right targets, consistent with the changes in RT. We interpret these results with regards to both accuracy and as a likely reflection of stimulusresponse compatibility effects related to the use of the computer mouse for which the more frequent usage is a left click [35,36]. However, such stimulus-response compatibility effects are unlikely to impact systematically the pattern of findings in the experiments we subsequently conducted as both left and right handers would be equally impacted. Importantly, the above findings suggest that there was no demonstrable impact of handedness, or the laterality of the response-hand on covert exogenous attention. Thus, intrinsic differences in body representation related to handedness and hand laterality did not likely influence how attention was distributed/ shifted when performing a distal task.

Experiment 2
Experiment 1 rules out any effect of handedness on the distribution of visuospatial attention in a covert orienting paradigm. Following from this, the aim of the second experiment was to systematically the combined influence of handedness, hand proximity and functional orientation of the hand (i.e. whether the palm or back-of-hand was oriented towards the display) on the distribution of visuospatial attention. We adopted posture elements from the methodologies of Lloyd et al., [6] and Reed et al., [7] presenting the palmar (grasping) and back-of-hand (nongrasping) surfaces towards one of two target placeholders (in blocked trials) whilst keeping the hand location constant across postures. The participant's hand was positioned below the monitor directly under one of the two target placeholders with either the palmar or back-of-hand surface oriented upwards (towards the target placeholder). Thus when targets appeared in the hand-adjacent location they were either in graspable or non-graspable space of the hand. Participants held either their dominant or non-dominant hand directly under the placeholder that corresponded with the hand side (e.g., right hand place under the right placeholder) and we compared performance between left and right-handed participants.
In this experiment, we aimed to establish which is perceptually important as a cue for attention: the relationship between the hand near the display and the target, or the relationship between response hand and the target or any combination of these. We predicted that the proximal hand would be weighted as the stronger cue for visuospatial attention. If this is the case there should be evidence of attentional prioritisation (faster RTs and improved accuracy) for targets near versus distant from the hand, more so when the grasping versus non-grasping region of the hand is oriented towards the display, as a reflection of attentional prioritisation relative to grasping affordances. Moreover, these effects should be more pronounced (in the form of greater cueing effects) when the proximal hand is dominant compared with non-dominant, as a result of greater structural and functional representation [14]. Alternatively, if the goal of the action is the primary driver for visual attention, changes to visuospatial attention should be evident when the dominant hand is responding.

Methods
Subjects. Forty-six undergraduate students (35 females; 22 left-handed and 24 righthanded by self-report) from the University of Queensland (mean age 19.79 years) completed the experiment in return for course credit, all gave informed consent.
Stimuli and procedure. These were identical to those of Experiment 1 with the exception of the following. Of the right handed participants, 11 responded to triangles with their index finger and circles with their middle finger and 13 with the reverse. Of the left handed participants; 11 responded to triangles with their index finger and circles with their middle finger and 11 completed the task with the reverse arrangement. Participants completed two blocks of 128 trials (256 total) one with the hand graspable (Fig 1a) and one with the hand non-graspable (Fig 1b) with block order counterbalanced across participants.
Half of the participants positioned their left hand under the left bottom corner of the screen and responded to targets with their right hand (left hand proximal) and the other half positioned their right hand under the right bottom corner of the screen and responded to targets with their left hand (right hand proximal). Postures and time-courses are shown in Fig 1. In the grasping condition, the hand was positioned directly under one of the target placeholders with the palm oriented towards the placeholder (Fig 1a). In the non-grasping condition, the hand was positioned directly under one of the target placeholders with the back of the hand oriented towards the placeholder (Fig 1b).

Results and Discussion
Participants with overall accuracy < 70% (2 participants, both right-handed) were excluded from further analyses. This resulted in 11 left-handed and 11 right handed participants completing the task with their left hand adjacent to the display and 11 left-handed and 10 righthanded participants completing the task with their right hand adjacent to the display. Anticipatory RTs < 150ms (0.13%) and non-stimulus driven RTs > 1000ms, (5.98%) were also excluded (see Table 2 for summary of mean RTs and accuracy for each level of each condition). All follow-up comparisons to interactions were evaluated against a Bonferroni corrected pvalue based on the number of comparisons conducted.
RT analysis. A mixed repeated-measures ANOVA was conducted on the mean RTs the within subjects factors of target location (hand adjacent; hand distant), hand posture (nongraspable; graspable) and the between subjects factor of handedness (left handed; right handed) and proximal hand laterality (left; right). As computed in Experiment 1, RT cueing effect was the dependant variable.
There was an interaction between hand posture, target location, and proximal hand laterality F(1,42) = 5.81, p = .020, Z 2 p ¼ :12. We conducted pairwise t-tests comparing between proximal hand laterality groups for each level of hand posture and target location and found larger cueing effects for the non-graspable hand distant condition for the right proximal hand laterality versus left proximal hand laterality t(42) = 2.85, p = .007.
To examine handedness effects directly, we conducted separate ANOVAs for each handedness group with the factors, hand posture, target location and proximal hand laterality. We found no significant differences for left handers. For right handers, there was a trending interaction between hand posture, target location and proximal hand laterality F(1,42) = 4.06, p = .057, Z 2 p ¼ :17. Follow-up planned comparisons were conducted comparing the graspable and nongraspable postures for each target location (two comparisons per handedness group). For right handed participants with the right proximal hand laterality, there was a larger cueing effect Table 2 when targets appeared near versus distant from the hand when the hand was in the graspable posture t(20) = 2.53, p = .032. No other effects reached significance. Error analysis. Mean error was evaluated by comparing cueing effects (valid-invalid accuracy) in a mixed ANOVA with the within subjects factors of target location (hand adjacent; hand distant) and hand posture (non-graspable; graspable) and the between subjects factor of handedness (left handed; right handed) and proximal hand laterality (left; right). There was a main effect of target location F(1,42) = 8.52, p = .006, Z 2 p ¼ :17. Cueing effects were larger for hand adjacent versus hand distant targets. There was also a main effect of hand posture F(1,42) = 4.91, p = .032, Z 2 p ¼ :11, with larger cueing effects for the graspable versus nongraspable posture, consistent with a visuospatial bias relative to grasping affordances.

. Mean RTs (in ms), with standard errors (in parentheses), percentages of accuracy, for left and right handed participants, separated by hand side (left, right), target location (hand distant, hand adjacent), hand posture (non-graspable, graspable), cue validity (valid, invalid) and mean cueing effect.
An interaction between target location and hand posture clarified the above main effects F (1,42) = 8.00, p = .007, Z 2 p ¼ :16. Follow-up t-tests compared target locations between hand postures. These revealed larger cueing effects for objects adjacent to versus distant from the grasping region of the hand t(42) = 2.66, p = .001. In addition, there was also an interaction between hand posture and proximal hand laterality F(1,42) = 5.37, p = .025, Z 2 p ¼ :11. We followed this up by comparing between proximal hand laterality groups (left; right) for each hand posture. There was a greater cueing effect for the graspable posture when the right versus left hand was proximal t(25) = 2.49, p = .042.
There was also a three-way interaction between hand posture, target location and hand under the monitor F(1,42) = 5.93, p = .019, Z 2 p ¼ :12. To follow this up, we compared proximal hand laterality groups for each level of hand posture and target location (4 comparisons). Cueing effects were larger for targets adjacent to the graspable region of the hand when the right versus left hand was proximal t(23) = 2.49, p = .020 (see Fig 2 for graphical representation). Taken together, these findings suggest that there was enhanced ability to discriminate target identity, when targets onset near the graspable region of the right hand, compared with both the non-graspable region. Moreover, they indicate that there was impairment of processing for targets which onset distant from this region. To examine effects related to handedness, we conducted separate ANOVAs with the within-subjects factors hand posture, target location and proximal hand laterality for each handedness group. As with the RT findings, we found no significant differences for left handers. For right handers there was a main effect of target location F(1,20) = 6.25, p = .021, Z The findings of Experiment 2 indicate that there are biases in near-hand attention relative to the grasping affordances of the hand which influence the observer's object identification accuracy. When considered in combination with the RT results, these findings suggest that visuospatial attention is biased to the grasping space of the right hand. As a result, object identification is enhanced in this location and associated impairments in identifying targets in opposite the site of biased attention. For right handers specifically, the results also suggest that there may be speeded engagement and delayed disengagement of attention to the graspable region of the dominant hand. In line with this, the results provide further evidence that lefthanders may not mirror attention biases for their dominant hand in the same manner as right handers. Instead the current findings suggest that both left-and right-handers have greater accuracy in detecting objects near the graspable space of the right hand. This may reflect usespecific changes in representation, because left handers must often employ their non-dominant hand to complete tasks, due to the fact that many every-day objects (e.g., door handles) afford action from the right hand. It may alternatively reflect visuomotor biases that are specific to the left hemisphere of the brain, consistent with earlier findings [20,21,22]. However, it is also important to note that when broken down by handedness group, these effects dissipate which suggests that strong dominant hand grasping-space biases in right-handers may contribute to this attentional bias.

Experiment 3
In Experiment 2, we found that the functional representation of the right hand, when proximal to the display, influenced the distribution of visuospatial attention. Accuracy of object identification was enhanced near the palm of the right hand and right handers also displayed faster engagement and delayed disengagements of attention to the grasping space of the dominant hand. The aim of Experiment 3 was to examine whether such visuospatial biases remain when the hand is crossed over the body midline, and thus are specific to functional representation of the limb.
In many instances in everyday life, we use our hands in an ipsilateral location or towards the body midline. Yet we are also capable of using hands to complete actions in the opposite region of visual hemispace; crossed over the body midline [32]. Hand-specific visuospatial biases have been interpreted by earlier research to be a reflection of the response properties of bimodal neurons [7,8,9,33]. Attentional biases result from the overlapping visual and tactile representation of the space near the hand. If it is the case then the biases found in Experiment 2 should occur irrespective of the hands location in visual space. Moreover, it is important to establish whether the laterality of targets (relative to the hand) and observer handedness impact attentional distribution in space.
We used a similar methodology to Experiment 2 but the dominant or non-dominant hand under the contralateral rather than ipsilateral target placeholder. We predicted that if the visuospatial and response biases found in the second experiment were due specifically to internal representation of the right hand and not to an overall lateralisation bias, we would expect that the findings to will be replicated when the hands cross over the body midline. That is, we should find the same cueing effects for targets appearing near the right hand now in the contralateral region of visual space.

Methods
Subjects. Thirty-five undergraduates (21 females; 19 right-handed; 16 left-handed by selfreport) from the University of Queensland (mean age 22.14years) participated in return for course credit or for AUD10 paid remuneration, all gave informed consent.
Stimuli and procedure. Stimuli and procedure were identical to Experiment 2 with the following exceptions. Participants crossed either their right or left hand (randomised between participants) over the midline and held it under the opposite corner of the monitor (see Fig 3  for diagrammatic representation). Of the right handed participants in the left hand adjacent condition; five responded to triangles with their index finger and circles with their middle finger and five with the reverse. Of the left handed participants in the left hand adjacent condition, four responded to triangles with their index finger and circles with their middle finger and four completed the task with the reverse arrangement. Of the right handed participants in the right hand adjacent condition; five responded to triangles with their index finger and circles with their middle finger and four with the reverse. Of the left handed participants in the right hand adjacent condition, four responded to triangles with their index finger and circles with their middle finger and four completed the task with the reverse arrangement.

Results and Discussion
Participants with overall accuracy < 70% (3 participants; all right-handed) were removed from further analyses. Anticipatory (< 150ms, < .001%) and non-stimulus driven (> 1000ms, 0.05%) RT trials were excluded from latency and accuracy analyses. Of the remaining sample, right handed and 9 left handed participants completed the task with the left hand adjacent to the monitor and 8 right handed and 8 left handed participants completing the task with their right hand adjacent to the monitor (see Table 2 for summary of mean RT and accuracy for each level of each factor). All follow-up comparisons to interactions were evaluated against a Bonferroni corrected p-value based on the number of comparisons conducted.
RT analysis. We conducted a mixed ANOVA with the same factors as the RT analysis for Experiment 2. There were no significant findings. To examine how handedness influenced the distribution of attention we conducted separate ANOVAs by handedness group (left; right) using cueing effect as the variable of interest. No main effects or comparisons reached statistical significance.
Error analysis. We conducted a mixed ANOVA with the same factors as the error analysis for Experiment 2. These revealed an interaction between target location and handedness F (1,31) = 4.79, p = .036, Z 2 p ¼ :13. Follow up t-tests compared mean accuracy for each target location between handedness groups. Consistent with the results of Experiment 2, this revealed that right handers had a greater cueing effect for targets appearing adjacent to the hand, compared with those appearing distant from the hand t(33) = 2.50, p = .017. To examine how handedness influenced accuracy we conducted separate ANOVAs by handedness group (left; right) using accuracy cueing effect as the dependant variable. For left handers there was a main effect of target location F(1,15) = 4.81, p = .045, Z 2 p ¼ :24, such that there was a larger cueing effect for targets distant from versus adjacent to the hand (see Fig 4). No other effects reached significance. For right-handers there were no significant differences. The above findings suggest that hand-centred bias in accuracy remained when hands were crossed over the body midline. This was such that object discrimination was more accurate for objects which onset near the right hand compared with those distant from it, irrespective of handedness. However, there was no longer any evidence for graspable biases for the right hand. This may be due to the reduction in the actionable value of the hand when crossed over the body midline, or to the increased difficulty of completing the task with the hand in a crossed posture. That is, the posture undertaken with the hand crossed over the body midline and the grasping side of the hand (palm) was oriented towards the display, reduces the capability for grasping. The crossed position makes grasping awkward and therefore may reduce the relevance of the hand orientation as a cue for visuospatial attention. While it is still possible to grasp an object in this posture, it is a less posturally comfortable action than the uncrossed position. Thus the affordances typically conveyed to the object near the hand by the orientation of the palm may be reduced. In turn, this may mitigate the significance of the palm orientation as a cue for attention. In addition, the results of Experiment 3 suggest that left handers no longer exhibit grasping biases for the right hand or generalised biases in the accuracy of object discrimination relative to the proximity of either the dominant or non-dominant hand, when hands are crossed over the body midline. Instead we found an exaggerated cueing effect in accuracy when targets are distant from the hand. There are a number of potential explanations for this, for example that left handers are more accurate when responding to objects in their dominant (left) side of hemispace, irrespective of hand location. However, further investigation is required to ascertain whether this is the case, as the same effect was not found in Experiment 1 where both hands were distant from the monitor.

General Discussion
The aim of the present research was to investigate how grasping affordances and handedness work in combination with action goals to influence the distribution of visuospatial attention near hands. Experiment 1 found that when completing a distal task, exogenous cues modulated attention. Handedness or the laterality of the response hand, conversely, do not influence attention systematically. Experiment 2 found that the grasping affordances of the right hand, when proximal to a visual display, biased shifts in visuospatial attention. Both left and right handers displayed greater accuracy costs when detecting objects near versus distant from the palm of the right hand. In addition, right handers displayed more rapid engagement slower disengagement of visuospatial attention to hand-adjacent targets near the graspable region of their dominant hand. In Experiment 3, when hands were crossed over the body midline, only right-handers retained dominant hand biases in the accuracy of target identification, and grasping space biases were no longer apparent. Taken together, these findings provide evidence that visuospatial attention is distributed to objects near the hands, relative to grasping affordances, and the strength of the underlying representation.
The current research presents a number of novel findings relating to the distribution of visuospatial attention near the body. Foremost, we extended upon the findings of Reed and colleagues [8] by showing that when the task requires a higher-order aspect of visual processing (shape discrimination versus onset detection), hand location and posture modulated shifts of attention. Importantly we also illustrated that such attentional biases enhanced object identification at the site of the attention shift, and impaired object identification in the unattended zone.
The aforementioned changes in visuospatial attention run somewhat counter to the overall biases in the distribution of visuospatial attention found by Reed and colleagues [8]. Those authors found that observers were faster to detect objects near the palm of their hands, irrespective of cue validity and hand laterality [8]. We propose that this is likely the result of the discrimination (as opposed to detection) task employed in the present research. With regards to task demands, the difference between target detection and discrimination is that the latter requires focal attention whereas the former does not [37]. Specifically, object shape discrimination requires a higher order judgement of object properties than detection does, and as a result great attentional resources to discern target identity. Thus, the present findings likely reflect that internal representation of hands relative to grasping capability enhances focal attention which in turn facilitates object processing. This is a critical finding because it speaks to the level of attentional processing that is affected by the functional properties of the hands. When considered in combination with the findings of Reed et al. [8], detection of visual objects may be speeded when they are adjacent to the hands, but there are additive attentional benefits to processing object properties when they are graspable.
Our results support the proposal by Lloyd and colleagues [6] that right-lateralised biases in attention shifting evident in their research, were potentially the result of hand dominance in their sample. We extended on these by investigating a left-handed sample and found that left handers do not display the mirror attentional biases for their dominant hand that right-handers do. Instead we showed that left-handers also display biases in visuospatial attention relative to the location and functional orientation of the right hand. Also, we found that visuospatial biases for right-handed subjects are restricted to the dominant (right) hand; they did not affect performance when the non-dominant hand was proximal to the display. The present findings further clarify those of Lloyd and colleagues [6] because they show that shifts of attention occur relative to the grasping affordances of the right hand, when ipsilateral versus contralaterally aligned. Moreover, these findings fit with earlier research [20,21,22,23], which has shown that visuomotor mechanisms specific to the left hemisphere of the brain have a crucial role in visually guided actions, and that individuals display similar patterns of space use, irrespective of handedness. Thus, the robust right-hand specific biases found in the present research across handedness groups may reflect an impact in attention biases that stem from or are impacted by such visuomotor biases. Stemming from this, it is important for future studies to disambiguate the left at which handedness begins to play a role in shaping or biasing the distribution of visuospatial attention, as such findings suggest that it is not as clear-cut as just a generalised bias towards faster/ more accurate detection of objects near the dominant hand.
The functional account of near-body attention posits that the allocation of additional attentional resources near the hands serves to enhance the cognitive processing of action-relevant stimulus properties [1,2,7,8]. These in turn are thought to guide sensorimotor transformations required to act on hand-adjacent objects [38,39]. We extended upon this to show that biases in near-hand attention were associated with enhancements in object identification, to the detriment of locations distant from the hand which supports the improved perception for action hypothesis. Thus, the current findings provide a precise picture of the visuospatial attention biases by illustrating that they are not only relative to handedness but more to grasping affordances. Furthermore, visuospatial biases are likely the result of expertise-dependant changes to visual perception rather than solely a reflection of strengthened right hand representation in right-dominant people. When considered in combination, the findings of Experiment 2 and 3 suggests that graspable biases in visuospatial attention do not occur in contexts where the orientation of the hand in relation to the stimulus makes grasping awkward. Yet one element of affordances not addressed in the present research, is how object affordances themselves may impact such grasping space biases. In the present study we utilised two-dimensional objects in our experimental paradigm which do not optimally elicit shape specific grasping affordances compared with three dimensional objects. As we most often interact with three dimensional objects in everyday life, it is an important direction for further study to establish how the shapes of objects themselves modulate hand-centred biases in visuospatial attention.
Following from the above, we have shown that the underlying limb representation is not the sole factor driving near-hand attention. The majority of near-hand attention research has focused primarily on right handers. The current findings contribute to understanding of body representation by demonstrating that mirrored biases are not displayed by left handers. Instead we found evidence for right-hand grasping space biases and no evidence for dominant hand associated biases in visuospatial attention for left handers. One possible explanation for these effects is that left handers are more heterogeneous in terms of the degree of hand laterality compared with right handers [14,16]. Thus it may be the case that the lack of significant differences found in the present study between the two groups is a reflection of this heterogeneity. Moreover, the current findings, particularly with regards to Experiment 2, may reflect use-specific changes in visuospatial attention. Because left handers exist in a world designed to afford actions for the right hand, this may as a result have enhanced the underlying representation of right-grasping affordances, as was evident in the present study.
Moreover, it is also likely that left handers exhibit changes to visuospatial attention based on the functional properties and proximity of their hands but that these do not impact the mechanisms of attention probed in the present task paradigm. Earlier evidence provides support for lower level perceptual differences in the representation of the dominant and non-dominant hand between left and right handers. For example, Le Bigot and Grosjean [27] demonstrated that left handers display greater visual sensitivity (d ') for objects near both their dominant and non-dominant hand. When considered in combination with the present findings, this indicates that differences in visuospatial attention may stem from or be contributed to by underlying changes in visual sensitivity. Thus it is important for future research to establish which mechanisms of visual perception and visuospatial attention are impacted by lefthand dominance, because they appear to be less lateralised than those evident for right handers. In addition to this, the present study only investigated handedness as a discrete construct: left or right handed. Of course, a large body of previous research has shown that people exist upon a continuum when it comes to degree of overall laterality. In line with this, future research should evaluate the extent to which degree of laterality (i.e. the degree to which one is either left or right handed) impacts the attentional mechanisms evaluated in the present study.
To sum, the current study has shown that intrinsic representation of hands, both with regards to the strength of underlying representation and actable properties, modifies the distribution of visuospatial attention near the body. Our results suggest that there are biases in visuospatial attention relative to the grasping properties of hands which serve to enhance the identification of objects in this space. We have also shown that this occurs to the detriment of perceptual processing of objects which appear near to the body, but distant from the hand. These findings extend current understanding of near body attention because they demonstrate that both proximity and functional orientation of the hands play critical roles in directing visuospatial attention. Thus future research regarding near-hand attention and perception must account for the impact of both on effects of interest.