Abstract
Shadows in physical space are copious, yet the impact of specific shadow placement and their abundance is yet to be determined in virtual environments. This experiment aimed to identify whether a target’s shadow was used as a distance indicator in the presence of binocular distance cues. Six lighting conditions were created and presented in virtual reality for participants to perform a perceptual matching task. The task was repeated in a cluttered and sparse environment, where the number of cast shadows (and their placement) varied. Performance in this task was measured by the directional bias of distance estimates and variability of responses. No significant difference was found between the sparse and cluttered environments, however due to the large amount of variance, one explanation is that some participants utilised the clutter objects as anchors to aid them, while others found them distracting. Under-setting of distances was found in all conditions and environments, as predicted. Having an ambient light source produced the most variable and inaccurate estimates of distance, whereas lighting positioned above the target reduced the mis-estimation of distances perceived.
Similar content being viewed by others
Introduction
Many visual cues to distance are available in natural environments1, and the precision and accuracy with which judgements can be made will depend on the particular combination of information available in a given context2. For example, Surdick et al.3 found that in a stereoscopic display, perspective cues were more effective at producing an accurate perception of distance than other depth cues, while the effectiveness of some lighting cues, such as relative brightness, contributed very little. The effectiveness of many cues, especially parallax distance cues, also tends to reduce rapidly with distance1,2.
Perception of distance in sparse environments tends to be poor, as predicted from geometrical considerations of the limited visual cues that are available. For example, distance tends to be underestimated, with relatively high uncertainty, when few visual cues are available4,5,6,7,8,9. We have previously tested whether adding specific environmental cues (binocular cues, linear perspective, surface texture, and scene clutter) could enhance the accuracy and precision of distance-dependent perceptual tasks using virtual reality10. Performance in this instance was measured via the degree to which individual cues contributed to a reduction in the amount of bias and variability in participant responses. It was found that adding visual information did indeed improve performance, such that a full-cue context allowed for the highest levels of accuracy and precision. In this case, both binocular and pictorial cues were important for obtaining accurate judgements. To further to these findings, the current experiment assessed the specific contributions made by clutter and cast shadows to the perception of relative distance, to explore the importance of lighting cues in complex environments for the accurate perception of 3D space.
Shadows occur when an object or surface disrupts the visibility of another surface or object to a light source, thus reducing the degree to which it is illuminated. The surface in shadow may belong to an object that is the same as, or different from, the one that is responsible for the shadowing. Mamassian et al.11 provide definitions for a number of distinct aspects of shadows and shading, including shading, attached shadows, and cast shadows. Shading refers to the variation in the amount of light reflected by surfaces, as a result of their orientation relative to the light source. Shading may be distinguished from the shadowing effect under consideration in the presented experiment, which is a result of the occluding of regions of surfaces from the light source. Attached shadows are those that are formed on the same surface that is occluding the light source. In contrast, cast shadows are those which are formed on a different surface from the one occluding the light source. It is cast shadows that are the focus of the current experiment, and in particular shadows that are cast on a horizontal surface (such as a ground plane or table top) by objects that are not attached to the surface.
When an object is placed on a surface, its cast shadow will be directly adjacent to the object in the image, as shown in Fig. 1. In this case, the size of the shadow, relative to the height of the object, can be used to provide information about the direction of the light source. For an object with a height (h\(_1\)), and a shadow with a length (f), the direction (\(\theta _1\)) of the light source is given by:
When an observer views an object above a table-top at eye-height, and the object is illuminated from a single direction, the shadow will tend to be detached from the object, examples of this are presented in Figs. 1 and 2. When the light source is directly above the object, its shadow will be cast on the horizontal surface at the same distance as the object. If the light source if further away than that the object, the shadow will be cast at a distance that is nearer than the object itself. Conversely, if the light source is nearer than the object, or behind the observer, the shadow will be cast at a distance that is further away than the object. The location of the shadow on the surface provides information about the distance of the object relative to locations on the surface, albeit in a way that is ambiguous due to its dependence on the direction of the light source. Geometrically, it can be seen that, when the shadow is further away than the object, its distance from the object in the horizontal direction is given by:
where D is the distance, h\(_2\) is the height of the object and \(\theta _2\) is the direction of the light source. Thus, if the direction of the light source can be estimated, the distance of the object along the surface, relative to the location of its shadow, can be inferred.
Allen12 demonstrated that shadows can be used a source of information about the distance to objects, by disambiguating the contributions of distance and vertical location to the height in the visual field of objects. In addition, another study by Allen13 showed that the direction of cast shadows contributes to distance judgements, and that shadows have also been found to improve distance judgements regardless of the direction of lighting14. Cavanagh et al.15 showed that the lateral displacement of a shadow from a target object influenced the perceived depth between the two; therefore, for a given depth separation, the size and direction of this offset will be determined by the direction of lighting. The results obtained were consistent with both a model in which the visual system assumes a single light source16, and a Bayesian model that weighted different lighting directions by their relative likelihood in the natural environment.
Wanger et al.17 used two tasks to assess the contribution of shadows to the perception of 3D surface layout. They found that shadows improved observers’ ability to accurately align objects in three-dimensional space, and to accurately match their size. Additionally, te Pas et al.18 conducted an odd-one-out experiment, in which participants were presented with three stimuli with varying lighting and shadows. By collecting data of participant’s eye fixations, reaction time, and percentage of correct answers they showed that participants relied primarily on shadows in identifying differences in the direction and intensity of lighting.
Cast shadows can also have a strong influence on the apparent 3D layout and motion in moving scenes19,20,21,22. Kersten et al.19 showed that when presented with a ball moving linearly away from the observer in depth, and a shadow which would correspond with a bouncing ball, an observer was more likely to use the shadow information to perceive the ball as having non-linear movement. One of the conclusions of this study was that when a shadow moves, observers assume that this is caused by the movement of the objects, rather than the light source. That is, the locations of shadows are used to inform the estimation of the locations of objects under the assumption of a fixed light source.
Dee et al.23 reviewed the literature on the visual effects of shadows. For simple cases with an object on a flat surface, when a light source does not move, then physical properties of the object such as its size, motion, and shape can be inferred from the shadow.
Shadows have great potential for improving the perception of distance in augmented reality, by helping users to accurately locate virtual objects within the physical environment. In virtual reality, Hu et al.24 showed that cast shadows can be used to determine when an object is in contact with a tabletop, or to judge the distance between the two. The presence of cast shadows can increase the accuracy of distance judgement in augmented reality25,26, and this is affected by the degree of mismatch in the direction of lighting in the real and virtual environments25,27.
The current study had three goals. The first was to determine whether the presence of cast shadows could improve the accuracy of distance judgements for an object that was located above a horizontal surface. The second was to determine how these distance judgements were affected by the position of the light source, which governs the spatial location of the shadow relative to the target object. The third goal was to determine whether accuracy could be further improved by the presence of visual clutter on the tabletop, which could be used to determine the direction of the light source, and thus allow the shadow to serve as an unambiguous cue to distance.
The experiment consisted of two virtual environments which were tested separately, both using a perceptual matching task in a VR headset. Participants were tasked with matching the distance of a floating target object to that of a horizontal reference line on a tabletop. The location of the light source was varied between blocks.
In one lighting condition, only ambient lighting was present, so that no shadows were cast. In all other conditions, one light source was present, and its location relative to the object was varied. This light source was either in a fixed location in the environment so that it did not move with the object, or it was locked to the object, so that it moved and maintained a constant position relative to the object. In the fixed light condition, the light source was either at the near (observer’s) end of the table or the far end. For the conditions in which it was locked to the object, the light was either nearer, directly above, or further away than the object. These lighting conditions are depicted in Fig. 2. The experiment was repeated in a Sparse and a Cluttered environment.
The first hypothesis relates to the overall performance in the Sparse and Cluttered environments: (1) regardless of the influence of shadows, the presence of additional visual cues in the Cluttered environment compared to the Sparse environment is predicted to improve the accuracy of distance judgements10,28.
In the Sparse environment, shadows are predicted to have a number of effects. As shadows are expected to contribute to distance estimation, the first hypothesis here is that (2) performance will be worse overall in the Ambient condition. In addition, (3) performance is expected to be more precise for the condition in which the light source is directly above the target (Locked on top), when it is perfectly aligned with the reference at the correct distance, than when the light source is in front or behind. This is because in this condition, the observer is required to align two features (the shadow and reference line) that overlap, rather than alignment with a spatial offset. Furthermore, (4) performance is hypothesised to be more accurate when the light source is locked to the moving object than when it is static in the environment, since this creates a constant shape and offset of the shadow. While the shadow provides a potentially useful cue to the location of the object, when the light is not directly above the object there will be an offset between the reference line and the shadow when the object and reference are perfectly aligned. Additionally, it is hypothesised (5) that observers may have a tendency to align the shadow with the reference line. This would lead to an the object being positioned nearer when the light is in front of the object than when it is behind it.
Finally, (6) it is hypothesised that conditions which improve the accuracy of settings will be more evident in the Cluttered environment than in the Sparse, since this provides more reliable information about the direction of the light source, from the sizes of the shadows cast by the surrounding objects on the table-top (Eq. 1). This should also reduce any systematic biases that are associated with the lighting direction. Note that this predicted increase in the effects of shadows is in addition to the overall improvement in performance in the cluttered environment relative the sparse environment.
Methods
Participants
Total of 21 naive participants were recruited using the University’s online system as well as through word of mouth. Participants received a £5 voucher for approximately 45 minutes of their time to complete the study.
The methods and protocols were approved by the Psychology Ethics Officer and carried out in accordance with the ethical guidelines at the University of Essex. All participants gave informed consent.
Materials and apparatus
An Oculus Rift headset and associated controllers were used, with two environments created in Unreal Engine 4.2. Figure 3 shows the virtual Sparse environment presented to participants: a room containing just a long table (1 m width by 4 m length) and the target object (an 10 by 10 by 5 cm cuboid presented at eye-height). The architecture of the room, target, and reference line were created using the Engine’s Starter Content. The Cluttered environment incorporated objects added on to the tabletop, shown in Fig. 4. These were scans obtained from Unreal Engine Marketplace29,30 or real objects scanned by the experimenters. A total of 21 objects were positioned randomly across the full length of the table surface.
The target object had a point light locked to the participant-facing side, with an intensity of one lumen and attenuation radius of 10 cm. This light was programmed to not cast any shadows during the experiment, and ensured that the luminance of the target object itself did not differ across the lighting conditions of the experiment, which did however affect the presence, location and shape of the cast shadows.
Shadow conditions
Both experiments contained the following six shadow-lighting conditions, depicted in Fig. 2:
-
Ambient: no point light source, no cast shadows.
-
Static back: a light source locked to the edge of the table furthest from the participant (position in environment: X = 36.7, Y = 0, Z = 175 cm; rotation: X = 360 °, Y = \(-65\) °, Z = 180 °). The size and shape of the shadow changed as a function of the distance between the target and the light source, which was positioned between the target and the far edge of the table.
-
Static in front: a stationary light source, positioned at the edge of the table closest to the participant (position in environment: X = 37, Y = 0, Z = 175 cm; rotation: X = 0 °, Y = \(-65\) °, Z = 0 °). The size and shape of the shadow varied as a function of the distance between the target and light source, which was positioned between the target and participant.
-
Locked on top: a light source attached to the target object (position from target: X = 0, Y = 0, Z = 120; rotation pointing directly down onto target). The cast shadow was directly below the target.
-
Locked back: a light source attached to the target object (relative transformation position from Locked on top: X = \(-60\), Y = 0, Z = 0 cm; rotation: X = 0 °, Y = 50 °, Z = 0 degrees °). The cast shadow moved with the target object and was positioned between the target and the far edge of the table.
-
Locked in front: a light source attached to the target object (relative transformation position from Locked on top: X = 60, Y = 0, Z = 0 cm; rotation: X = 0 °, Y = \(-50\) °, Z = 0 °). The cast shadow moved with the target object and was positioned between the participant and target
In Fig. 2b,c the light is static in the environment, whereas in (d), (e), and (f) the light source is attached to the target object and moves along with the distance adjustments to the target. There was no point light present in (a). The light sources were obtained from Unreal Engine’s Starter Content.
The conditions were presented in separate blocks, each having 20 trials. In the Cluttered environment, the objects on the virtual table cast their own shadows from the light sources; these were not manipulated in any way. The light sources were not visible objects themselves, rather the emitted light was visible only on the objects within the environment.
Task and procedure
The goal was for each participant to match the distance of the target object (a floating box) to that of a reference line on the table surface. Participants were told to read instructions and these were also described to them by the experimenter; there was an opportunity to ask questions before starting. The order of the environments was counterbalanced, alternate participants were allocated to start in either the Cluttered or Sparse environment. In all conditions the task was to move the target object so that it aligned in distance with the reference line on table.
Once inside the VR headset participants adjusted the height of a chair to ensure targets appeared at eye-height, and sat in an indicated area for the duration of the experiment. They were instructed to not move from this position during the experiment. A horizontal reference line appeared on the surface of the table in front of the observer at a random distance between 40 and 365 cm. The target object was positioned at eye-height and, using the controller’s A and B face-buttons, participants could move the target along the X axis (forwards/backwards in egocentric distance) to a position where it looked as though it was directly above the reference line. Once the distance was decided upon, the participant pressed the controller’s trigger which saved the trial and moved onto the next one. The next trial automatically began by randomising the distance of the reference line and target box. No indication was given to participants about when the condition was to change, and no feedback on the accuracy of their settings was provided.
The lighting condition was altered across blocks of trials. The metrics recorded were the final position participants set the target object to be at in each trial and the position of the reference line, so that the offset for each trial could be calculated.
Results
Formatting data
The raw data output from the experiments were in a comma separated value format; each session had an independent .csv file named by the experiment name, time, and date of the session. Each row represented an individual trial with columns for the condition number, the distance of the reference line (in cm), and the set distance of the target (in cm). A numerical key was used to identify the conditions.
The signed and unsigned error for each trial was calculated by subtracting the distance of the reference stimulus from the set distance of the target stimulus. For the signed errors, a positive value indicated that the object was positioned further away than the references, and a negative value indicated that it was positioned closer than the reference. Unsigned errors reflect the overall error in settings, incorporating both systematic biases and random variability. Signed errors reflect any systematic biases in the settings, since random variability will tend to cancel out across repeated trials.
The following exclusion criteria were decided upon before the experiment: any participants who did not finish the entire experiment, and any participant who withdrew their data after a two-week window from completion. Additionally, removal of trials where the set distances were further or closer than the ends of the table were to be excluded, but no data were set beyond these limits (see Section “Evaluation of the methodology”). No data were excluded based on these criteria.
Analysis strategy
A Pearson’s correlation was conducted, separately for the two environments, on the distance of the reference line and set target positions in each trial to identify the general relationship between participant estimates and correct settings.
A linear mixed effects model was used to analyse the data with the following equation, where the distance set, by participants (P) in each trial, is predicted by the distance of the reference line (D) and the lighting condition (L, numbered from 1 to 6):
Two-way (environment-by-condition) repeated measures ANOVAs were used to identify how these factors influenced the unsigned and signed distance-setting errors. These were followed up with one-way repeated measures ANOVAs to assess how settings were affected by the shadow condition in each environment, Tukey’s HSD to which pairs of conditions differed.
Analyses
There was a consistent under-setting of distances, in that regardless of the lighting or sparsity of the context, participants set the target to be not as far away as it should have been, which can be seen in Figs. 5 and 6. A correlation was conducted on the distance of the reference line and set target positions, to determine the overall relationship between the estimates and correct settings. In the Sparse environment, there was a strong positive relationship between the two (r = 0.935, p<0.001); the same positive trend was found in the Cluttered environment (r = 0.933, p<0.001).
The data for each environment were analysed separately: Table 1 presents the slope and intercept information for the lighting conditions from both environments. Accurate performance would produce a slope value of one and an intercept of zero. In both environments the intercept was largest in the Static back lighting condition, and closest to zero in the Locked in front condition. All slopes were below one, indicating under-setting of distances. The closest to accurate slope estimate was produced by the Locked on top condition in the Sparse environment, and Locked back in the Cluttered environment.
Unsigned error
Unsigned errors, which reflect the overall accuracy of the settings, are plotted as a function of environment and condition in Fig. 7. A two-way repeated measures ANOVA showed a significant effect of condition (F (5, 100) = 12.187, p< 0.001) but no effect of environment nor interaction . Overall performance, and the effects of shadows, were thus no more different in the Cluttered environment (hypotheses 1 and 6).
The effects of shadows were analysed separately for the sparse and cluttered environments. Unsigned errors were significantly influenced by the lighting conditions in both the sparse (F (5, 100) = 9.11, p< 0.001) and cluttered (F (5, 100) = 7.47, p< 0.001). The only conditions with consistently smaller unsigned errors was the lit from above condition, which were significantly smaller than in all other conditions (all p< 0.02).
These results show that performance was most accurate when the light was directly above the target, and the shadow directly aligned with the reference when the object is at the correct location (hypothesis 3). However, there was no evidence that unsigned errors were reduced in any of the other shadow conditions (hypothesis 2).
Signed errors
Signed errors, which reflect systematic biases in the settings, are plotted as a function of environment and condition in Fig. 7. A two-way repeated measures ANOVA showed a significant effect of condition (F (5, 100) = 19.815, p< 0.001) but no effect of environment nor interaction . Overall bias, and the effects of shadows, were thus no different in the Cluttered environment (hypotheses 1 and 6).
The effects of shadows were analysed separately for the two environments. Signed errors were significantly influenced by the lighting conditions in both the Sparse (F (5, 100) = 10.6, p< 0.001) and Cluttered (F (5, 100) = 12.7, p< 0.001).
It was predicted that, if observers had a tendency to align the shadow with the reference, then settings should be closer when objects were lit from in front than behind (hypothesis 5). This was found for both environments, and for both the static lights and those that were locked to the target objects (all p< 0.05).
Residual errors
Using the data from Table 1, the residual errors for each participant in each environment and condition were calculated. These were calculated by performing a regression of set distance on reference distance for each participant and each condition, then calculating the average absolute difference between each setting and the values predicted by the regression. This separates out the between-trial variation in settings from the systematic error captured by the regression. The mean unsigned residuals were then used in two way repeated measures ANOVA. A significant effect of condition was found (F (5, 100) = 3.654, p = 0.004), but no effect for environment nor interaction. The distribution of these data are shown in Fig. 7. Residual errors were lowest when the light source was locked to the top of the object.
Discussion
This experiment aimed to identify which cast-shadow conditions produced the most accurate responses in a perceptual matching task, and whether the addition of scene anchors enhanced the influence of cast shadows. Six lighting conditions were created and presented in virtual reality. The task was repeated in Cluttered and Sparse environments. A thin box was used as a target so that the most visible face was flat and therefore did not have shading as an added visual cue. The presence of scene clutter did not improve the accuracy of settings, and did not affect the influence of shadows on these settings. A light source directly above the target, projecting a shadow at the same distance as the object onto the table-top, produced the most accurate settings. When the light source was in front of or behind the target, there was a tendency for observers to align the shadow with the reference line.
Interpreting the findings
There was a strong correlation between the reference stimulus location and set target location, showing that participants understood the task and completed it appropriately. There was an overall tendency however, for participants to set the target stimulus at a shorter distance than the reference line Fig. 7. This indicates an over-estimation of target distance relative to the reference line, which may reflect a partial misinterpretation of its greater height-in-the-field as a cue to distance1,2. This effect has been found for similar contexts in augmented reality31.
No differences in accuracy or precision were identified between the environments. However, a difference was found between the lighting conditions. The overall error in responses for each condition can be seen in Fig. 6, where the Locked on top lighting clearly produces the most precise distance estimates. This is supported by the unsigned errors made, Fig. 7, which were lowest in this condition.
Evaluation of the methodology
Initially, an exclusion criterion was to remove any trials with set distances outside the range of the table length, as this would not be possible should the experiment have been completed correctly. However this exclusion criterion was not used, because in the hypothesised event that the shadow was visible and being aligned with the reference at either end on the table, but the target was subsequently located beyond, this would have led to the removal of trials in which the participant was following a coherent strategy.
Lighting conditions experienced in daily life tend to include ambient light from overcast weather diffusing the light source, and directional lighting, for example from the sun or from ceiling lights32. The lighting conditions used here were created to replicate these, with overtly visible cast shadow effects (with the notable exception of the Ambient lighting condition). Additionally, some less common light instances were created, although all were physically plausible.
We hypothesised that performance would be more most accurate when the position of the light source was locked to the target objects, so as to produce a constant offset in the location of the shadow. While light sources that are a fixed location produce more complex shadow effects, these are also more likely to be encountered in everyday life, and that participants may be more familiar with this dynamic shadow cues. We might also predict that the additional shadows present in the cluttered condition might be most useful when the light source was moving with the target object, by providing information about the movement of the light source.
Retinal size is a reliable cue when either the dimensions of a stimulus or the distance is known by the observer33,34. Here, the participants were not familiarised with either the dimensions of the target, or the testing area within the virtual environment. The participants were not primed as to whether the size of the target would be consistent or changing throughout the experiment. However, if they assumed the target was of a constant size, then its retinal size provides a reliable cue to changes in distance, which are likely to contribute to the reliable distance estimation shown in Fig. 5.
The reference stimulus was presented at a randomly chosen distance on each trial, unlike many other distance-matching studies where the target appears at a limited few distances35,36,37. This avoided the possibility of participants remembering the relative size of the target at each set distance, and then try to replicate the size, rather than complete the trial as required38.
The two environments were created so that a comparison of whether scene clutter aids in the perception of distances. The specific objects used in the Cluttered environment were chosen as they are typical everyday objects that varied in size and shape so that the shadows produced would also be varied. The positioning for these objects was random, a structured approach to the positioning of the clutter may influence the perception of the scene.
Clutter objects and an abundance of shadows on the tabletop may have been distracting for some participants, which may have contributed to the non-significant difference between the two environments. After discussions with some of the participants it became apparent that a subset of the participants valued the objects in the Cluttered environment differently than other participants. In particular, individuals with attention deficit hyperactivity disorder may find the additional objects in the Cluttered environment too distracting. Thus, increasing the availability of extra visual cues may not always lead to the anticipated improvements in performance39.
Practical implications
Sugano et al.27 showed that shadows are an important depth cue in AR, but that lighting needs to be consistent with physical lighting due to the see-through nature of the device. Gao et al.25 used an AR device to test the impact of shadows on distance perception. It was predicted that distance would be under-estimated, that virtual shadows would enhance accuracy in the perceptual matching task, and that misalignment between physical and virtual lights would however impair these estimates. Our results also showed a misestimation of distances and that lighting conditions also had a significant impact on these judgements. Our results are thus consistent with those of Gao et al.25, in that we found an improvement in performance, shown in the reduction in signed, unsigned and residual errors when objects were lit from above. This is consistent with other studies that have that the presence of these ‘drop’ shadows support the most accurate perception of distance in AR40. In our VR study we did not consider the possibility of an inconsistency in the shadows created for different objects. This is an important practical implication for AR, in which some shadows will be created by natural light sources in the real world, and others by virtual light sources. This misalignment can negatively affect the accuracy of distance judgements25.
The lit from above condition in this experiment, where the light was attached to and moving with the object, is a special case in that it creates a situation in which the shadow is cast on the horizontal plane at the same distance as the object. Moreover, the fact that it moves with the object may mark it out as an additional light source, thus potentially reducing any effects of conflict between this and other light sources in the virtual scene. As such, this condition is likely to also improve accuracy in augmented reality applications.
Conclusions
Adding clutter to an environment, where there are more shadows cast and more points of reference, does not enhance the accuracy of distance estimates within the scene compared to an identical environment minus the clutter. Clutter objects are therefore not used as points of reference, or at least there are some subsets of participants who do not find them useful for improving performance.
Consistent with previous literature, there was an under-setting of distances throughout the experiment. From the lighting techniques tested, both the precision and accuracy were highest in the Locked on top lighting condition. This condition provides the observer with a simple strategy of aligning the shadow with the reference line in order to complete the task accurately. Differences in biases between the cases where the light was in front of or behind the object suggest that observers did in fact have tendency towards aligning the shadow and reference line, even in cases where lighting was not from directly above.
Some of the lighting conditions presented in this experiment can be considered to be ‘unnatural’ in typical viewing of physical space, as there are few instances where a light source is attached to an object with some sort of offset. The two static lighting conditions are more representative of typical viewing. However, implementations of unnatural lighting can be seen in specific applications of VR and cinematography. It is in these instances where the perception of distance can be enhanced through the use of the conditions presented in this experiment.
Data availability
The datasets generated and analysed during the current study are available in the University of Essex Research Data Repository, http://researchdata.essex.ac.uk/143/
References
Cutting, J. E. & Vishton, P. M. Perceiving layout and knowing distances: The integration, relative potency, and contextual use of different information about depth. In Perception of Space and Motion, 69–117 (Elsevier, 1995).
Hibbard, P. B. Estimating the contributions of pictorial, motion and binocular cues to the perception of distance. Perception 50, 152 (2021).
Surdick, R. T., Davis, E. T., King, R. A. & Hodges, L. F. The perception of distance in simulated visual displays: A comparison of the effectiveness and accuracy of multiple depth cues across viewing distances. Presence Teleoperator Virtual Environ. 6, 513–531 (1997).
Gogel, W. C. & Tietz, J. D. Absolute motion parallax and the specific distance tendency. Percept. Psychophys. 13, 284–292 (1973).
Johnston, E. B. Systematic distortions of shape from stereopsis. Vis. Res. 31, 1351–1360 (1991).
Viguier, A., Clement, G. & Trotter, Y. Distance perception within near visual space. Perception 30, 115–124 (2001).
Yang, Z. & Purves, D. A statistical explanation of visual space. Nat. Neurosci. 6, 632–640 (2003).
Chen, J., McManus, M., Valsecchi, M., Harris, L. R. & Gegenfurtner, K. R. Steady-state visually evoked potentials reveal partial size constancy in early visual cortex. J. Vis. 19, 8–8 (2019).
Hornsey, R. L., Hibbard, P. B. & Scarfe, P. Size and shape constancy in consumer virtual reality. Behav. Res. Methods 52, 1587–1598 (2020).
Hornsey, R. L. & Hibbard, P. B. Contributions of pictorial and binocular cues to the perception of distance in virtual reality. Virtual Real. 25, 1087–1103 (2021).
Mamassian, P., Knill, D. C. & Kersten, D. The perception of cast shadows. Trends Cognit. Sci. 2, 288–295 (1998).
Allen, B. P. Shadows as sources of cues for distance of shadow-casting objects. Percept. Motor Skills 89, 571–584 (1999).
Allen, B. P. Angles of shadows as cues for judging the distance of shadow casting objects. Percept. Motor Skills 90, 864–866 (2000).
Allen, B. P. Lighting position and judgments of distance of shadow-casting objects. Percept. Motor Skills 93, 127–130 (2001).
Cavanagh, P., Casati, R. & Elder, J. H. Scaling depth from shadow offset. J. Vis. 21, 15–15 (2021).
Mamassian, P. & Goutcher, R. Prior knowledge on the illumination position. Cognition 81, B1–B9 (2001).
Wanger, L. R., Ferwerda, J. A. & Greenberg, D. P. Perceiving spatial relationships in computer-generated images. IEEE Comput. Graph. Appl. 12, 44–58 (1992).
te Pas, S. F., Pont, S. C., Dalmaijer, E. S. & Hooge, I. T. Perception of object illumination depends on highlights and shadows, not shading. J. Vis. 17, 2–2 (2017).
Kersten, D., Mamassian, P. & Knill, D. C. Moving cast shadows induce apparent motion in depth. Perception 26, 171–192 (1997).
Taya, S. & Miura, K. Cast shadow can modulate the judged final position of a moving target. Atten. Percept. Psychophys. 72, 1930–1937 (2010).
Ouhnana, M. & Kingdom, F. A. Objects versus shadows as influences on perceived object motion. i-Perception 7, 2041669516677843 (2016).
Katsuyama, N., Usui, N., Nose, I. & Taira, M. Perception of object motion in three-dimensional space induced by cast shadows. Neuroimage 54, 485–494 (2011).
Dee, H. M. & Santos, P. E. The perception and content of cast shadows: An interdisciplinary review. Spat. Cognit. Comput. 11, 226–253 (2011).
Hu, H. H., Gooch, A. A., Creem-Regehr, S. H. & Thompson, W. B. Visual cues for perceiving distances from objects to surfaces. Presence Teleoperators Virtual Environ. 11, 652–664 (2002).
Gao, Y. et al. Influence of virtual objects’ shadows and lighting coherence on distance perception in optical see-through augmented reality. J. Soc. Inf. Disp. 28, 117–135 (2020).
Adams, H., Stefanucci, J., Creem-Regehr, S. & Bodenheimer, B. Depth perception in augmented reality: The effects of display, shadow, and position. In 2022 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), 792–801 (IEEE, 2022).
Sugano, N., Kato, H. & Tachibana, K. The effects of shadow representation of virtual objects in augmented reality. In The Second IEEE and ACM International Symposium on Mixed and Augmented Reality, 2003. Proceedings., 76–83 (IEEE, 2003).
Wu, B., Ooi, T. L. & He, Z. J. Perceiving distance accurately by a directional process of integrating ground information. Nature 428, 73–77 (2004).
Patchs. Food pack 01 (2016). Unreal Engine Marketplace.
Dekogon Studios. Construction site vol. 2 - tools, parts, and machine props (2018). Unreal Engine Marketplace.
Rosales, C. S. et al. Distance judgments to on-and off-ground objects in augmented reality. In 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), 237–243 (IEEE, 2019).
Hibbard, P. B., Goutcher, R., Hornsey, R. L., Hunter, D. W. & Scarfe, P. Luminance contrast provides metric depth information. R. Soc. Open Sci. 10, 220567 (2023).
Baird, J. C. Retinal and assumed size cues as determinants of size and distance perception. J. Exp. Psychol. 66, 155 (1963).
Holway, A. H. & Boring, E. G. Determinants of apparent visual size with distance variant. Am. J. Psychol. 54, 21–37 (1941).
Singh, G., Swan II, J. E., Jones, J. A. & Ellis, S. R. Depth judgement measures and occluding surfaces in near-field augmented reality. In Proceedings of the 7th Symposium on Applied Perception in Graphics and Visualization, 149–156 (ACM, 2010).
Dey, A., Jarvis, G., Sandor, C. & Reitmayr, G. Tablet versus phone: Depth perception in handheld augmented reality. In 2012 IEEE international symposium on mixed and augmented reality (ISMAR), 187–196 (IEEE, 2012).
Swan, J. E. et al. A perceptual matching technique for depth judgments in optical, see-through augmented reality. In IEEE Virtual Reality Conference (VR 2006), 19–26 (IEEE, 2006).
Keefe, B. D. & Watt, S. J. The role of binocular vision in grasping: A small stimulus-set distorts results. Exp. Brain Res. 194, 435–444 (2009).
Smallman, H. S. & John, M. S. Naive realism: Misplaced faith in realistic displays. Ergon. Des. 13, 6–13 (2005).
Diaz, C., Walker, M., Szafir, D. A. & Szafir, D. Designing for depth perceptions in augmented reality. In 2017 IEEE international symposium on mixed and augmented reality (ISMAR), 111–122 (IEEE, 2017).
Acknowledgements
This was funded by the Economic and Social Research Council.
Author information
Authors and Affiliations
Contributions
Conceptualization (RH, PH); Data Curation (RH); Formal Analysis (RH, PH); Investigation (RH); Methodology (RH, PH); Software (RH); Supervision (PH); Visualisation (RH); Writing (RH, PH); Review & Editing (RH, PH)
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Hornsey, R.L., Hibbard, P.B. Distance mis-estimations can be reduced with specific shadow locations. Sci Rep 14, 9566 (2024). https://doi.org/10.1038/s41598-024-58786-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-58786-1
- Springer Nature Limited