In virtual reality (VR) applications such as games, virtual training, and interactive neurorehabilitation, one can employ either the first-person user perspective or the third-person perspective to perceive the virtual environment; however, applications rarely offer both perspectives for the same task. We used a targeted-reaching task in a large-scale virtual reality environment (N=30 healthy volunteers) to evaluate the effects of user perspective on the head and upper extremity movements, and on user performance. We further evaluated how different cognitive challenges would modulate these effects. Finally, we obtained the user-reported engagement level under the different perspectives.
We found that first-person perspective resulted in larger head movements (3.52±1.3m) than the third-person perspective (2.41±0.7m). First-person perspective also resulted in more upper-extremity movement (30.08±7.28m compared to 26.66±4.86m) and longer completion times (61.3±16.4s compared to 53±10.4s) for more challenging tasks such as the “flipped mode”, in which moving one arm causes the opposite virtual arm to move. We observed no significant effect of user perspective alone on the success rate. Subjects reported experiencing roughly the same level of engagement in both first-person and third-person perspectives (F(1.58)=0.9,P=.445).
User perspective and its interaction with higher-cognitive load tasks influences the extent of movement and user performance in a virtual theater environment, and may influence the choice of the interface type (first or third person) in immersive training depending on the user conditions and exercise requirements.
Interactive virtual reality (VR) applications that engage human motor performance span a wide variety of domains, including piloting remote vehicles, movement performance training, and neurorehabilitation therapy following brain injury. These applications often seek to stimulate brain activity, facilitate practice-based learning, and to leverage the brain’s capability for reorganizing its structure . In this sense, using a display to provide visual feedback to the trainer or trainee can reinforce learning and performance skills. As technology evolves, new opportunities for trainee feedback become possible. For example, out-of-body perspectives in immersive environments allow for motion feedback that may otherwise not be visible or apparent to the trainee. However, whether such interactive approaches impact the user performance and mental load is an open research question. For instance, it is not currently known which characteristics of virtual reality are most beneficial for rehabilitation training , although one avenue might be to increase the patient’s engagement with the therapy . Engagement, as defined in the videogame literature, is the experience involving a lack of awareness of time and the real world, and a sense of being in the task environment .
From the multiple VR characteristics to consider, the user perspective is of particular interest, since it can dramatically influence the user’s perception of their body within the environment. A first-person perspective presents the world directly through the character’s eyes, though sometimes without a body representation. In contrast, a third-person perspective, while not natural, allows a full view of body motions. In our study, we displayed the virtual character in front of the user’s viewpoint; the avatar did not act as a mirror. Three studies have compared these user perspectives in general VR applications. Salamin’s experiments using a head-mounted display endorsed the use of the first-person perspective for precision tasks and the use of third-person perspective for movement tasks . Later, Salamin found that participants attempting to catch a ball estimated distances better and performed more similarly to real world-conditions in third-person perspective . Covaci et al.  compared the perspectives in a cave automatic virtual environment (CAVE) with a ball throwing task. In this task, participants underestimated distances in both perspectives; however, similarly to Salamin results, they performed slightly better in third-person perspective. It remains unclear what other effects if any, the viewer perspective may have on aspects like motion or perceived engagement and whether VR applications in some domains justify their choice (e.g., limb-studies ).
Not only the engagement can vary depending on the user perspective but also based on the chosen immersive environment. Immersive environments use stereoscopic imagery to enhance depth perception and position tracking to react to user’s movement, being the most common head-mounted displays and small rooms with projectors . Another environment is a virtual-reality theater where instead of using projectors, the setup uses tiled-displays. Compared to head-mounted displays, a virtual-reality theater provides a nonintrusive and collaborative experience (although not in this study), only requiring the users to wear lightweight polarized glasses. Besides, virtual-reality theaters provide higher-resolution than most of the commercial head-mounted displays and allow the user to see the real world. Despite the potential benefits of these virtual-reality theaters to a variety of interactive training, few studies have yet explored such benefits.
In this work, we evaluate the relationships between user perspective and user performance, motion, and perceived level of engagement. From a motor rehabilitation perspective, such an evaluation would be enriched by further considering the context of cognitive load levels. Therefore, we test under varying cognitive loads the existing null hypothesis that user perspective does not influence user performance, body head, and upper-extremity movements, or self-perceived level of engagement. To test this hypothesis, we use a virtual target-reaching game to evaluate the effects of both first- and third-person perspectives in a virtual reality theater environment, a CAVE2 . The game makes possible the use of different challenge levels, leading to exercises with varying cognitive loads.
We found that first person user perspective caused larger head movements (3.52±1.3SD m) than third-person perspective (2.41±0.7SD m; main effect F(1,29)=159.2, P<.0001; see Fig. 1). The flipped exercise head movement was also significantly higher than both normal and trail exercises in first-person (F(2,58)=19.4,P<.0001) and similarly in third-person. Data distribution passed the normality test (P=0.064) and sphericity assumption after the application a Box-Cox transformation. Results statistics are available in Table 1.
Exercise type caused the largest effect on the upper-extremity movement (F(2319) = 168.4, P<.0001). The flipped exercise mode lead to longer trajectories: flipped exercises (Fig. 2) in first-person perspective (30.08±7.28SD m) had longer trajectories than flipped exercises in third-person perspective (26.66±4.86SD m). The second effect was caused by the interaction between the perspective and exercise type (F(2319)=25.6,P<.0001).
We noted that the displayed trail paths did not influence the movement chosen by a subject; none of the subjects followed a straight line, but rather the expected geodesic curve. We further noticed that in the flipped exercise mode, different subjects adopted different strategies when negotiating the higher cognitive load: some subjects aimed for the targets directly, while others explored the space from the frontal plane towards the sagittal plane, to first scan the area. Results statistics are available in Table 2.
The results show complex variation in the completion time, related to both user perspective (first or third) and exercise type (normal, trail, or flipped) (Fig. 3). Surprisingly, perspective effects varied interactively with the exercise type (ANOVA main effect F(2,58)=39.7,P<.0001). In particular, for the normal exercise type, the third person completion times were the same as the first person times. For the trail type, the third person times were higher than the first person times (M_trail_third=42.2s, SD_trail=6.8s; M_trail_first = 38.2s, SD_trail_first=11.2s). For the flipped type, the third person times were lower than the first person times (M_flipped_third=53s, SD_flipped_third=10.4s; M_flipped_first=61.3s; SD_flipped_first=16.4s).
The largest effect on completion time was caused by the flipped exercise type, which lead to the largest completion time (F(2,58)=200.4,P<.0001, followed by pairwise post hoc comparisons; Fig. 3). User perspective also influenced completion time (F(1,29)=7.5,P=.01, followed by pairwise post hoc comparisons; Fig. 3). Target size apparently also influenced completion time (F(1,29)=10.9,P=.003); although the ANOVA results suggested an effect of the target sphere size, we did not find significant differences in the post hoc analysis. We further found interactions between perspective and target size (F(1,29)=6.8,P=.015). The data passed the Lilliefors test (P=.094) after performing a Box-Cox transformation, and did not violate the sphericity assumption. Results statistics are available in Table 3.
Hitting the target - score
Subjects obtained high scores consistently in all three exercise types (M=19.66,MED=20,MIN=9,MAX=20), in both first and third-person perspectives. Again, we found differences due to multiple factors. User perspective caused the largest effect on the score (F(1319)=86.4, P<.0001); the first-person perspective correlated with lower scores in flipped mode (M=18.9,SD=2.1) in comparison to the third-person perspective (M=19.9, SD=0.7). The exercise type caused the second largest effect on the score (F(2319)=37.1,P<.0001); scores for exercises in flipped mode were significantly lower than in the normal and trail modes. Results also suggest the effect of other factors on the score, such as the target size (F(1319)=30.7,P<.0001), the interaction between perspective and exercise type (F(2319)=31.2,P<.0001), the interaction between perspective and target size (F(1319)=14.5,P=.0002), and the interaction of all the factors (F(2319)=3.25,P=.04). However, the post hoc analysis indicated that flipped exercises were the most common factor associated with subjects missing targets (Fig. 4). Additionally, the skewed distribution to the right (higher scores) suggests the reaching exercise game had overall low difficulty. Results statistics are available in Table 4.
Level of engagement
We found no significant differences in the engagement questionnaire score between first and third-person perspectives (F(1,58)=0.59,P=.445; Fig. 5). Non-parametric analysis also revealed no significant differences (Kruskal-Wallis test, X2=2.72,P=.0989). The data distribution transformed with the Box-Cox function satisfied the normality test (P=.1192) and the Levene’s test (P=.766). Results statistics are available in Table 5.
As shown by our results, we found differences in head and hand movements depending on the user perspective and the three exercise types. These differences are of particular importance for VR-based rehabilitation, where the patient’s condition constrains the range of motion that patient can perform.
The user perspective had the main effect on the head movements; there was significantly longer total distance traveled by the head over 20 reaching attempts in the first-person (M=3.52, SD = 1.30) perspective than third-person perspective (M=2.41, SD = 0.70). Compared to the third-person perspective, in which the avatar and targets appear in front of the user, the first-person perspective required the subjects to explore a wider area of screen space, which implicitly required the use of peripheral vision. From the perspective of VR-based rehabilitation, the use of the first-person perspective may be beneficial when the patient requires broader spatial exploration as in rehabilitation of spatial neglect .
For upper-extremity movements, the first-person perspective with higher cognitive load exercises (i.e., flipped) yielded longer trajectories. In contrast, upper-extremity movements for the normal and trail exercise settings did not reveal significant differences between the two user perspectives; thus, revealing a zero-effect from the visual cues shown as trails and the shortest path between the hand and target on the trajectory distance.
Our results indicate that the user perspective alone is not the main factor influencing user performance, but a combination of user perspective and cognitive load. As expected, we found that flipped exercises (higher cognitive load) required more time to be completed (Fig. 3) in both user perspectives and had lower success rates (Fig. 6), compared to the normal and trail exercises. However, for this higher cognitive load exercise, subjects took more time to complete the task and had lower task success rates in the first-person perspective than in the third-person perspective. For normal exercises, there were no differences between user perspectives. For trail exercises, subjects took longer to complete the task in third-person perspective compared to first-person perspective (opposite case than trail exercises) and had similar success rates between the two perspectives. Similar to Salamin , user perspective alone did not influence completion time, and similar to Covaci et al. , user perspective alone did not influence task success.
We believe that the opposite results in completion time between the higher cognitive load task and the trail task relies on two factors: movements feedback and environmental feedback. In third-person perspective, the subjects were able to instantly observe wrong actions in the avatar and make corrections accordingly while in the first-person perspective, the restricted field of view did not ease this immediate feedback. This movement feedback was particularly useful for the higher cognitive task, the most unfamiliar for the subjects. As for environmental feedback, in the trail exercise, the displayed segment connecting the closest hand to the target was more visible in the first-person perspective; thus, it caught the subject’s attention faster. These visual cues may also be helpful to get people’s attention back, especially for patients with spatial neglect. Opposed to Salamin’s preference for using the first-person perspective for precision tasks , we believe that the third-person perspective feedback is preferred for precision tasks with a higher cognitive load that require unfamiliar movements and performed on an environment with field of view constraints. The recommendation may not apply to other cognitive tasks such as memory attention unless mixed in a dual attention setup with unfamiliar movements.
Level of engagement
In this study, we found that the user perspective had no significant effect on the degree of user-perceived engagement. Our findings are supported by the conflicting results reported by Denisova and Cairns  (first-person perspective more immersive) and respectively by Salamin  (third-person perspective more engaging). We believe that the low degree of task difficulty (for a healthy subject) and the static foot positioning required by the rehabilitation application may have influenced the subjects’ sense of engagement. As suggested by Faria et al. , exercises requiring more cognitive effort and attention may gather more attention. On the other hand, the study of Schuurink and Toet  suggests that the level of stimulation induced by a virtual reality environment is independent of the question of first- and third-person perspectives; our results support Schuurink and Toet’s conclusion in the context of tasks with relatively reduced cognitive effort.
We looked into the self-perceived level of engagement due to its potential to improve a patient’s experience with rehabilitation tasks. To measure this we used the definition of engagement from the gaming literature, while implementing our experiments in an immersive VR environment. We note that the term ‘immersion’ is overloaded between gaming research and VR research. Gaming research operates under the definition given by Jennet et al. , who describe the different levels of engagement as immersion, an experience felt by gamers. In the gaming literature, immersion is therefore used interchangeably with engagement and involvement . Jennet et al. further created the IEQ questionnaire to assess immersion (defined as engagement), based on components that influence the gamer experience. In contrast, VR research uses the term immersion to describe a property of the technology  that influences the level of presence — the “illusion of being there, notwithstanding that you know for sure that you are not” . Each of the two definitions of immersion is equally influential in its field of origin. Because we seek to measure game engagement, in our questionnaire we followed the Jennet et al.  definition. This definition acknowledges the influence of presence on engagement, and suggests that presence only appears at the deeper level of engagement.
Depth perception errors
Based on the questionnaire feedback, in the third-person perspective a few subjects experienced misperceptions, with no further impact, for objects placed to the sides and close to the frontal plane. Two of the subjects reported that in the third-person view the displayed scene seemed to be rendered in 2D; another subject thought mistakenly that the objects were behind the avatar. It is possible the lack of geometrical visual cues in spherical targets contributed to these isolated depth perception issues, although we provided consistently shadows as a cue. For example, Powell et al.  suggest using complex geometries (more complex than our simple sphere geometry) to further provide visual cues for reaching tasks. Nevertheless, none of these reported ln appear to have resulted in lower user performance than in the first-person view.
Conversely, no subjects reported depth-perception errors in first-person perspective, despite lower user performance results under specific settings. A recent study  in CAVE-like environments suggested that targets near the screen—that hence appear large to the user (as in first-person perspective)—are easier to understand than objects far into the scene (as in third-person perspective). Results from the Bruder et al. study  further report up to 50% misinterpretations of distance for objects that are further away. We believe their findings help explain the difference in our user reports of depth-perception errors.
Assumptions and limitations
Concerning assumptions and limitations, our study used a specific infrastructure that is currently not readily accessible to the public due to cost, technical support, and space requirements. We note, however, that the CAVE2 merely serves as a vehicle for testing our hypothesis, since it allows both first person and third person large screen feedback display, aside from the CAVE2’s many other, vast capabilities, which were not used in this study. We believe these findings would transfer to any another immersive system that allows both first person and third person large screen display. Given a strong need in domains like rehabilitation, such systems are not inaccessible to rehabilitation clinics. Our findings may also not be directly transferable to other platforms, for example, head-mounted displays. However, immersive large environments have specific advantages over head-mounted displays (e.g., non-intrusive equipment) that make such environments relevant to rehabilitation clinics. Last but not least, as suggested by Levin et al. , in general, further evaluation of the performance, quality and surrogate aspects of motor behavior are needed to analyze the fidelity of VR environment tasks to physical environment tasks. We note, nevertheless, on the deliberate simplicity of the exercise set we designed and we tested in this work, in collaboration with a domain expert.
For our study design, even though we randomized the starting user perspective per subject, the order of the exercises with different target sizes was not randomized which could introduce confounding effects. Moreover, our study considers engagement from the perspective of a self-reported “experience that produces a lack of awareness of time and the real world” . We do not consider other aspects of the virtual reality experience such as embodiment (the sensation of being inside a body) .
The success score median of 20 (out of 20) across all exercises and both perspectives indicates that our game had an easy level of difficulty for healthy subjects. We corroborated this last statement with the results of the user survey; on a 1 to 5 Likert scale from easy to difficult, subjects scored the first-person perspective as 2.04 ± 0.92 and the third-person perspective as 2.21 ± 0.92. These results are not surprising as the arm-reaching exercises were designed for physical therapy patients, but tested on healthy subjects; rehabilitation patients would find the exercises more difficult. Furthermore, we allowed the healthy subjects eight seconds to target each object, when in fact they only needed an average of 2.3 seconds ± 1.1. These numerical results may not transfer to a population of stroke subjects. Besides, even though we developed our application in collaboration with rehabilitation experts, further studies with actual actions is required to evaluate its application on the target population and the transferability of our results. Finally, while using smaller target sizes did not increase the difficulty, it may be possible that smaller objects than the ones we used could make the game more challenging.
We presented the results from a user study comparing the effect of first-person and third-person user perspectives, under varying cognitive load, on user performance and upper-extremity movements for a set of reaching tasks in a virtual reality theater environment. Also, we analyzed the effect of user perspectives on the degree of engagement using a questionnaire. Considering only user perspective, the first-person perspective required the participants to make more head movements compared to the third-person perspective. For upper-extremity motion, the interaction between user perspective and higher-cognitive load exercises influenced the movements. These head and upper-extremity movements are essential when assessing the range of motion suitable for VR-based rehabilitation therapy. For user performance, the user perspective and the type of exercise influenced the completion time and the scores. Finally, as hypothesized initially, the user perspective did not influence the self-perceived engagement for this set of relatively easy tasks.
Thirty students (8 women and 22 men), aged 18–32 years (M=24.6 years, SD = 3.6 years) from the University of Illinois at Chicago were recruited for the user study (UIC IRB #2016-0332). Inclusion criteria included being over 18 years old, being an enrolled student, and not having significant physical impairments. We asked every participant about their physical conditions (e.g., movement problems in their trunk, upper-body or head) and their ability to perceive objects in 3D. We explained that any issue would limit their experience with the environment and would affect the expected results. The participants self-reported not having any physical condition that could limit their experience with the environment. Also, none of them reported having stereoscopic vision problems (i.e., problems perceiving depth cues).
Apparatus and materials
The CAVE2 environment  is a large-scale VR environment composed of 72 high-resolution displays arranged in eighteen columns in a circular fashion (Fig 7). The arrangement provides an area of 320 degrees of screens with a resolution of 36 Megapixels per eye. Fourteen Vicon Bonita infrared cameras track the position of retroreflective markers in tracking glasses and controllers in an area of 34 square meters. Twenty speakers and two subwoofers mounted at the top and bottom of the columns provide stereo 3D sound.
A Kinect v2 Ⓡ sensor is located in front of the central display (73 cm from the ground) to track the subject upper-extremity joints; the Vicon cameras only track the head position. In our study, to provide quality tracking points for the Kinect sensor, each subject was required to stand in the center of the environment approximately 3.3 meters away from the screens.
We designed and implemented a prototype of a VR-based rehabilitation game focused on upper-extremity tasks using Unity3D 5.2.3. The main exercise consists of reaching out with an upper-extremity in order to touch virtual spheres that appear at positions drawn from a uniform distribution, 70 cm away from the player’s chest. The virtual spheres appear one at the time. Experimenters can test different gaming conditions by choosing between first-person and third-person perspectives (Fig. 8). Experimenters can also select the number of targets to reach to. In order to increase or decrease the difficulty of the exercise, experimenters can also vary the target size. To further investigate the role of cognitive function on the user performance under the two perspectives, the prototype implements three exercise types (Fig. 9): 1) easy: normal view; 2) difficult: flipped (the left limb actions moved the right limb, and vice versa); and 3) very easy: trail—a mode that displays a trajectory path and a trace of the hand’s movements to help guide and document the upper-extremity motion.
The CAVE2 environment has, like most second-generation, modern immersive theaters, no ceiling or floor projection. This is due partly to display-technology constraints (while providing remarkably higher resolution and brighter environments than older systems with projected light, displays cannot provide passive/active stereo for multiple users at ceiling or floor locations), and partly due to collaboration requirements (for collaborative use, analysts wish to roll in desks and chairs to create a comfortable-enough environment; in this way, in addition to the environment immersive capabilities, they can continue to use their laptops and web-based technologies as they do in the office). Because of this design, the field of view in the first-person perspective is restricted; subjects only see their hands for fully extended arm elevations between 90 to 120 degrees approximately. Additionally, in the first person perspective the associated trail in trail mode is naturally only visible within the field of view. Object placement is constrained to be always inside the screen-visible space. We use the same space constraint for the third-person perspective, in order to compare both views under the same conditions. Moreover, we used a transparent material for the avatar body, to avoid potential occlusion issues for objects appearing in front of the avatar in the third-person perspective. Before running the experiment, we verified that the distance of an object was perceived similarly on both user perspectives by placing a styrofoam sphere at different locations and a virtual sphere at the same location in the virtual scene.
Movement logs from each exercise were recorded in JSON format by the application. The data was preprocessed using Python scripts and further analyzed using R for the statistical tests and G*Power for the power analysis. In addition to the movement logs, we recorded each user study session using a wide lens camera.
To determine whether the user perspective and the training settings have an effect on the user performance and body movements in the reaching-objects rehabilitation task, we designed twelve exercises. These activities combined the two user perspectives (first-person and third-person), the three exercise types (normal, flipped [i.e., right physical upper-extremity activates left virtual upper-extremity, and vice-versa], and trail), and two target sizes (spheres of 10 and 15 cm in diameter). Each exercise presented 20 targets, one at a time, and gave up to eight seconds to attempt to touch the target before it disappears. At the beginning of each user study session, each subject performed two rounds of exercises that combined the different settings. We updated the CAVE2 parameters to adjust the virtual scene to the subject’s interpupillary distance. Next, we randomly assigned the starting perspective. However, the six exercises within each user perspective had the same fixed order. The subject first worked with bigger targets in normal mode, then changed to the flipped mode, and finished with the trail mode. Finally, each subject repeated the exercises, this time using the smaller targets. The smaller targets were used to increase the difficulty of the game. Arm reaching was selected by our collaborator at the Shirley Ryan AbilityLab in Chicago, IL, as “the simplest full body task appropriate for a stroke survivor”, and thus appropriate for this study. The number and size of targets, the exercise timing, and the exercise frequency were furthermore selected empirically in order to pose a sufficient challenge to a healthy player. Nevertheless, the number and size of targets, the timing, and the frequency could be adjusted per subject.
We derived two user performance metrics: the total time required to complete the 20 attempts (completion time) and the score per exercise represented as the number of objects caught. In addition, to obtain body movement metrics, we recorded the subject’s joint positions. Body movement metrics included the total number of meters traveled by the hands and head displacements.
Finally, to determine whether the user perspective had an effect on the perceived level of engagement of the CAVE2 environment for the given task, we used the IEQ questionnaire . As recommended by Jennett et al. , we grouped the IEQ questions in sub-domains: basic attention, temporal dissociation, challenge, emotional involvement, and enjoyment. The sum of thirty questions on a Likert scale, from one to five, determines the engagement score. At the end of each study session, subjects filled out one survey per user perspective, and were asked to provide any additional feedback to the experimenter.
Analysis of data
To analyze the effects on user performance and body movements, we performed a multi-factor within-subjects analysis of variance. We checked the normality assumption for the within-subjects test using the Lilliefors (Kolmogorov-Smirnov) test . Data distributions that passed the tests were analyzed with a parametric ANOVA test. As required, we used the Aligned-Rank test  for data distributions that failed the tests. As required for repeated-measures ANOVA tests, in the parametric approach we performed an additional sphericity verification using the Mauchly’s test , which guarantees that the variances between all within-subject condition pairs are equal. Results for distributions that violate the sphericity assumption were adjusted using the Greenhouse-Geisser correction . We used P<0.05 to determine significant differences. Furthermore, we performed post hoc analysis using pairwise comparisons with the Holm-Bonferroni adjustment  for parametric approaches. For nonparametric methods, we used the differences of differences cross-factor contrast test  to analyze interactions and pairwise comparisons for independent variables.
For the analysis of the level of engagement, we performed a one-way ANOVA test for data that passed the normality and homogeneity of variance assumptions, and a Kruskal-Wallis test  for data that violated the assumptions. As in the previous design, we also used P<0.05 to determine significant differences.
Availability of data and materials
Data are available on request.
Cave automatic virtual environment
Cheung KL, Tunik E, Adamovich SV, Boyd LA. Neuroplasticity and virtual reality. In: Virtual Reality for Physical and Motor Rehabilitation. New York: Springer: 2014. p. 5–24.
Salamin P, Thalmann D, Vexo F. The benefits of third-person perspective in virtual and augmented reality? In: Proceedings of the ACM Symposium on Virtual Reality Software and Technology. New York: ACM: 2006. p. 27–30.
Salamin P, Tadi T, Blanke O, Vexo F, Thalmann D. Quantifying effects of exposure to the third and first-person perspectives in virtual-reality-based training. IEEE Trans Learn Technol. 2010; 3(3):272–6.
Covaci A, Olivier A-H, Multon F. Third person view and guidance for more natural motor behaviour in immersive basketball playing. In: Proceedings of the 20th ACM Symposium on Virtual Reality Software and Technology. New York: ACM: 2014. p. 55–64.
Febretti A, Nishimoto A, Thigpen T, Talandis J, Long L, Pirtle J, Peterka T, Verlo A, Brown M, Plepys D, et al.Cave2: a hybrid reality environment for immersive simulation and information analysis. In: The Engineering Reality of Virtual Reality 2013. Burlingame: International Society for Optics and Photonics: 2013. p. 864903.
Denisova A, Cairns P. First person vs. third person perspective in digital games: do player preferences affect immersion? In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. New York: ACM: 2015. p. 145–8.
Faria A, Couras J, Cameirão M, Paulino T, Costa G, i Badia SB. Impact of combined cognitive and motor rehabilitation in a virtual reality task: an on-going longitudinal study in the chronic phase of stroke. Los Angeles: 11th ICDVRAT; 2016.
Powell V, Powell W. Locating objects in virtual reality-the effect of visual properties on target acquisition in unrestrained reaching. In: Proc 10th Intl Conf on Disability, Virtual Reality and Assoc. Technologies, Gothenburg, Sweden. Reading: 2014.
Bruder G, Argelaguet F, Olivier A-H, Lécuyer A. Cave size matters: Effects of screen distance and parallax on distance estimation in large immersive display setups. Presence Teleoperators Virtual Environ. 2016; 25(1):1–16.
Levin MF, Deutsch JE, Kafri M, Liebermann DG. Validity of virtual reality environments for sensorimotor rehabilitation. In: Virtual Reality for Physical and Motor Rehabilitation. New York: Springer: 2014. p. 95–118.
Wobbrock JO, Findlater L, Gergle D, Higgins JJ. The aligned rank transform for nonparametric factorial analyses using only anova procedures. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York: ACM: 2011. p. 143–6.
We thank the faculty, staff and students at the Electronic Visualization Laboratory.
This work was supported in part by NSF awards CNS-1456638 and IIS-0952720. The funding body has no role in the design of the study, and collection, analysis, and interpretation of data and in writing the manuscript.
Authors and Affiliations
Department of Computer Science, University of Illinois at Chicago, 851 S Morgan Street, Chicago, 60607, USA
Juan Trelles Trabucco, Andrea Rottigni, Marco Cavallo, Daniel Bailey & G. Elisabeta Marai
Department of Bioengineering, University of Illinois at Chicago, 851 S Morgan Street, Chicago, 60607, USA
GEM and JP conceived the idea and supervised the project. JT, AR and MC designed and implemented the CAVE2 application. GEM and JT designed and implemented the user study. JT analyzed the data. JT, DB, JP and GEM wrote the manuscript. All authors gave authorization for the manuscript publication. All authors read and approved the final manuscript.
The user study protocol (#2016-0332) was reviewed and approved by the Institutional Review Board at the University of Illinois at Chicago. All the participants provided written informed consent before participating in a user study session.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
Trabucco, J.T., Rottigni, A., Cavallo, M. et al. User perspective and higher cognitive task-loads influence movement and performance in immersive training environments.
BMC biomed eng1, 21 (2019). https://doi.org/10.1186/s42490-019-0021-0