698 lines
135 KiB
Plaintext
698 lines
135 KiB
Plaintext
University of Texas Rio Grande Valley
|
||
ScholarWorks @ UTRGV
|
||
|
||
Marketing Faculty Publications and Presentations
|
||
|
||
Robert C. Vackar College of Business & Entrepreneurship
|
||
|
||
8-2022
|
||
What is XR? Towards a Framework for Augmented and Virtual Reality
|
||
Philipp A. Rauschnabel Reto Felix The University of Texas Rio Grande Valley Chris Hinsch Hamza Shahab Florian Alt
|
||
|
||
Follow this and additional works at: https://scholarworks.utrgv.edu/marketing_fac Part of the Marketing Commons
|
||
Recommended Citation Rauschnabel, P.A., Felix, R., Hinsch, C., Shahab, H. and Alt, F., 2022. What is XR? Towards a framework for augmented and virtual reality. Computers in Human Behavior, 133, p.107289. https://doi.org/10.1016/ j.chb.2022.107289
|
||
This Article is brought to you for free and open access by the Robert C. Vackar College of Business & Entrepreneurship at ScholarWorks @ UTRGV. It has been accepted for inclusion in Marketing Faculty Publications and Presentations by an authorized administrator of ScholarWorks @ UTRGV. For more information, please contact justin.white@utrgv.edu, william.flores01@utrgv.edu.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289 Contents lists available at ScienceDirect
|
||
Computers in Human Behavior
|
||
journal homepage: www.elsevier.com/locate/comphumbeh
|
||
|
||
What is XR? Towards a Framework for Augmented and Virtual Reality
|
||
Philipp A. Rauschnabel a,b,c,*, Reto Felix d, Chris Hinsch e, Hamza Shahab f,g, Florian Alt h
|
||
a Digital Marketing and Media Innovation, College of Business, Universita¨t der Bundeswehr München, Werner-Heisenberg-Weg 39 85577 Neubiberg, Germany b Management Center Innsbruck (MCI), Universita¨tsstraße 15, 6020 Innsbruck, Austria c Social Sciences, Economics, and Business Administration, Otto-Friedrich-Universita¨t Bamberg, Kapuzinerstraße 16, 96047 Bamberg, Germany d Robert C. Vackar College of Business & Entrepreneurship, The University of Texas Rio Grande Valley, 1201 W University Dr, Edinburg, TX 78539, USA e Seidman College of Business, Grand Valley State University, 1 Campus Dr, Allendale, MI 49401, USA f Faculty of Business and Economics, University of Malaya, 50603 Kuala Lumpur, Malaysia g Faculty of Management Sciences, University of Wah, 47040 Wah Cantt, Pakistan h Usable Security and Privacy Group, CODE Research Institute for Cyber Defense and Smart Data, Universita¨t der Bundeswehr München, Carl-Wery-Straße 18-22, 81739 München, Germany
|
||
|
||
ARTICLE INFO
|
||
Keywords: Definition Augmented reality Virtual reality Mixed reality Assisted reality Metaverse
|
||
|
||
ABSTRACT
|
||
Augmented Reality (AR), Virtual Reality (VR), Mixed Reality, and Extended Reality (often – misleadingly – abbreviated as XR) are commonly used terms to describe how technologies generate or modify reality. However, academics and professionals have been inconsistent in their use of these terms. This has led to conceptual confusion and unclear demarcations. Inspired by prior research and qualitative insights from XR professionals, we discuss the meaning and definitions of various terms and organize them in our proposed framework. As a result, we conclude that (1) XR should not be used to connote extended reality, but as a more open approach where the X implies the unknown variable: xReality; (2) AR and VR have fundamental differences and thus should be treated as different experiences; (3) AR experiences can be described on a continuum ranging from assisted reality to mixed reality (based on the level of local presence); and (4), VR experiences can be concep tualized on a telepresence-continuum ranging from atomistic to holistic VR.
|
||
|
||
“If we don’t all agree on the same definitions, then we have imme diate ambiguity and confusion when we are talking about Mixed Reality, Augmented Reality, or Virtual Reality, which then requires further explanations and alignment which then leads to wasted time and energy and potentially misalignment of expectations and so on”. (informant “MIKE”, Head of XR for a leading consulting firm)
|
||
1. Introduction
|
||
Recent advances in information technology, such as high-speed mobile Internet, artificial intelligence, increased computing power, and high-resolution displays, create new ways for users to experience reality (Dwivedi et al., 2020; Hoyer et al., 2020). Important industry players have developed a plethora of devices, brands, and labels to po sition themselves in this market. For example, Microsoft is promoting their Hololens as a “Mixed Reality” device (Rauschnabel, 2018). Meta Platforms, Inc. (formerly Facebook, Inc.) purchased Oculus – a “Virtual
|
||
|
||
Reality” (VR) company (Hoffman et al., 2014), to complement their primary social media products as a “metaverse” company. PTC discusses “Assisted Reality” as a new reality format for warehousing companies (Coon, 2018). Apple touted “Augmented Reality” (AR) as a technology that will disrupt the world (Raymundo, 2016). Furthermore, Deloitte (2018) uses the term “Digital Reality” and Accenture embraces “Extended Reality” (Raghavan & Rao, 2018).
|
||
This ambiguity and confusion of terms and concepts is also notable in the academic literature. For instance, Wedel et al. (2020, p. 443) state that “mixed reality (MR) merges both VR and AR”, indicating that “we refer to all these technologies [AR, VR, and MR] as VR and use the term AR only when the distinction is needed in a specific context.” At the same time, Milgram and Kishino’s (1994) influential real – virtual environment continuum conceptualizes mixed reality as an umbrella term, combining virtual and real elements. However, other scholars contest the Milgram and Kishino view by suggesting that mixed reality is a very specific type of reality, situated between AR and “augmented virtuality (AV)” (Farshid et al., 2018; Flavia´n et al., 2019). Hoyer et al.
|
||
|
||
* Corresponding author. College of Business, Universit¨at der Bundeswehr München, Germany. E-mail address: philipp.rauschnabel@gmail.com (P.A. Rauschnabel).
|
||
https://doi.org/10.1016/j.chb.2022.107289 Received 2 May 2021; Received in revised form 8 January 2022; Accepted 27 March 2022 Available online 31 March 2022 0747-5632/© 2022 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
(2020) propose that mixed reality is an extension of AR and argue that “while AR is mainly available through smartphone apps, MR requires a headset or an equivalent wearable device” (p. 59). In addition, some authors point out that AR is fundamentally different from VR (Tan et al., 2021). The discussion is further complicated by Milgram et al.’s (1995) observation that “Perhaps surprisingly, we do in fact agree that AR and VR are related and that it is quite valid to consider the two concepts together” (p. 283). Finally, the meaning of the term or abbreviation “XR” remains ambiguous.
|
||
Importantly, the ongoing ambiguity regarding AR, AV, mixed reality, and related concepts may be detrimental to the user experience1 for a number of reasons. First, this ambiguity “holds back those eager to explore the different opportunities these new technologies represent” (Farshid et al., 2018, p. 658), which, in turn, constricts both consumer value realization and cash flow for producers. Second, ambiguity and user confusion impact managerially-relevant outcome variables such as customer intention to use a product (Deng et al., 2010). By definition, customer perceptions that do not align with customer expectations will result in issues with satisfaction. Since satisfaction is linked to equity and other important managerial variables (Poushneh & Vasquez-Parraga, 2017; Szymanski & Henard, 2001), user experience is important for managers. In short, we concur with Flavia´n et al. (2019) who observe that the boundaries of AR, VR, and mixed reality have not been defined adequately, and we posit that the extant literature is ripe for a reorganization and reconceptualization of existing approaches to reality.
|
||
We address this gap by detailing a coherent framework to consoli date the existing and often contradictory perspectives currently found in both industry and academic literature. Specifically, this research has three objectives. First, we identify and organize extant terms, views, and definitions in the academic and practitioner-oriented literature. Second, we synthesize these concepts and terms into an ordering framework that is externally informed and validated through insights from focus groups and in-depth interviews with industry experts. Third, we delineate the core differences between new reality formats and guide future scholarly work by proposing avenues for future research in various disciplines.
|
||
Our work provides contributions to several streams of literature. Specifically, we advance the current literature on AR (e.g., Chylinski et al., 2020; Hilken et al., 2020; Rauschnabel et al., 2019) by concep tualizing local presence, defined as the degree to which a user perceives AR content as being actually here (Verhagen et al., 2014), as a key cri terion for AR. Our proposed AR continuum ranges from assisted reality (with low levels of local presence) to mixed reality (with high levels of local presence). Furthermore, we advance the extant literature on VR (e. g., Cowan & Ketron, 2019; Hudson et al., 2019; Mütterlein, 2018) by conceiving telepresence, defined as the degree to which a user feels present in the virtual rather than the physical environment (Steuer, 1992), as the focal construct. Our VR continuum ranges from atomistic VR (with low levels of telepresence) to holistic VR (with high levels of telepresence). Importantly, our conceptual framework clearly separates AR from VR as opposed to previous streams of research (Milgram et al., 1995; Milgram & Kishino, 1994) which proposed a fluent AR-VR con tinuum. In other words, we suggest that users can either be immersed2 in
|
||
1 Congruent with Hassenzahl and Tractinsky (2006), we conceptualize user experience as a multidimensional, complex construct that is the consequence of the user’s internal state (e.g., needs, moods, and expectations), the character istics of the technology (e.g., complexity, functionality, and usability), and the environment or context within which the interaction occurs (e.g., organiza tional vs. recreational setting).
|
||
2 Although we acknowledge that presence and immersion share communal ities (Mütterlein, 2018), we follow Slater and Sanchez-Vives’ (2016) proposi tion that immersion refers to the technical capabilities of a system, whereas presence (with its subforms of local and telepresence) describes the user’s subjective experience with the system.
|
||
|
||
an AR environment or in a VR environment, but not both simulta neously. Finally, our work addresses the ongoing confusion regarding the term XR, which we do not define as “extended reality” but with X as a placeholder for any form or new reality.
|
||
2. New realities
|
||
2.1. Augmented reality
|
||
Augmenting the view of the world has a long history. In his novel “The Master Key”, L. Frank Baum’s (1901) protagonist receives the su pernatural power of “character marker” – a special set of spectacles that superimpose a letter indicating an individual’s underlying personality on their forehead. Before this, the concept of Pepper’s ghost symbolized an “AR-like” illusion technique – although not digitally – in stage pro ductions from the 1860s.
|
||
While the concept of AR dates back to the 1950s (Carmigniani et al., 2011), the phrase is generally considered to have been coined by Tom Caudell and David Mizell in 1990 (Berryman, 2012). AR has been defined in a number of ways, but it typically refers to a combination of digital information with the real world that is presented in real-time (Azuma, 1997; Feiner et al., 1993; Milgram et al., 1995). A Google Scholar search of AR literature reveals hundreds of articles exploring this topic focusing on diverse topics like the development of the un derlying technology (Zhou et al., 2008), the impact on social interaction (Miller et al., 2019), and leveraging this technology in differentiated application areas, including medical training (Berryman, 2012), tourism (Yin et al., 2021; tom Dieck & Han, 2019), manufacturing (Schein & Rauschnabel, 2021), marketing (Chylinski et al., 2020; Hilken et al., 2017, 2020; Heller et al., 2019a, 2019b; Tan et al., 2021; Hoffman et al., 2022; Scholz & Duffy, 2018; Rauschnabel et al., 2022), service man agement (Heller et al., 2021), and architecture (Lin & Hsu, 2017). Overall, technological advances and the ubiquity of mobile devices have made AR available to a substantial number of users (Billinghurst et al., 2015).
|
||
It is important to note that multiple AR classifications coexist in the extant literature, and many are either incompatible or contradictory. Fig. 1 presents a classification of AR characteristics based on the most prominent devices, enablers, and display types for visual AR (non-visual AR will be discussed later). As a general rule of thumb, newer, dedicated AR devices typically include more specialized hardware (e.g., depth sensors, eye tracking, see-through/retinal displays, etc.), which also allow new forms of human-computer interfaces (e.g., controllers, hand and finger tracking, voice commands, retinal control, and brain com puter interfaces). Moreover, newer AR devices typically provide a higher level of embodiment by moving the technology closer to the human body, whereas more established approaches leverage ubiquitous tech nologies characterized by a wide market penetration (e.g., smartphones or WebAR on a laptop computer).
|
||
2.2. Virtual reality
|
||
The idea of providing users with an immersive, artificiallyconstructed reality predates the concept of AR. While the concepts of engaging in a technology-driven, fabricated reality go back to early science fiction and fantasy novels, the first approaches to “VR” were in fact panoramic paintings that sought to fill an individual’s field of view and make the viewer feel as if he/she were actually embedded in the scene (Bown et al., 2017). Whereas panoramic paintings utilize fore shortening to create a feeling of presence in a scene, stereoscopic photo viewers more effectively used this concept to create a realistic percep tion of “being there”. The first widespread use of technology that resembled contemporary VR was the Link Trainer application used to train pilots before and during World War II (Jeon, 2015). When we think of VR today, most picture a head mounted system that occludes infor mation from the environment while presenting information depicting a
|
||
|
||
2
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
Fig. 1. A classification of AR use cases (visual AR focus).
|
||
|
||
virtual environment to the user. These ‘head-mounted displays’ (HMDs) were initially designed for gaming and entertainment, but usage has gradually broadened to include areas like job training, prototyping, marketing, and tourism (Shahab et al., 2021). Researchers have also explored the usage of VR in several commercial applications such as retail outlets and supermarkets (Krasonikolakis et al., 2018), the fashion industry (Yaoyuneyong et al., 2018), manufacturing (Berg & Vance, 2017), tourism (Lee et al., 2020; Wei et al., 2019), healthcare (Fertleman et al., 2018), and as a research tool (Holla¨nder et al., 2019; Stadler et al., 2019).
|
||
With few exceptions (e.g., VR caves; compare Lu & Smith, 2009), VR has traditionally been limited to headset-based applications. The main distinction between different VR devices is the number of degrees-of-freedom (DoF), i.e., the number of parameters in a system that can vary independently of each other. For example, 3 DoF only supports rotational tracking, whereas 6 DoF supports both rotational and translational tracking (Pan & Hamilton, 2018).
|
||
2.3. Four views of new realities
|
||
When screening the literature for existing definitions and frame works, we searched for publications that introduce new reality formats to the literature in various disciplines, such as marketing, tourism, human-computer interaction (HCI), management information system (MIS), and computer science. As a first step, we conducted a literature search in common academic databases, including Google Scholar, Web of Science, IEEE Xplore, and the ACM Digital Library. By iteratively comparing similarities and differences expressed in extant classifications of these concepts, we grouped existing definitions and perspectives into four prototypical views. Once we determined that the four views adequately encapsulated the perspectives of the existing literature, we requested feedback from our informants during the interviews and found general support for our classification. Fig. 2 outlines these views which will be discussed in detail below.
|
||
2.3.1. The “MR-dominant view” In what is probably the most cited definition in this area, Milgram
|
||
et al. (1995) presented the Reality-Virtuality Continuum as a way to understand the relationship between real and virtual elements of the user experience. To the left side of the continuum lies the real envi ronment without the addition of virtual objects. To the right of the continuum lies the virtual environment, which refers to a fully virtual
|
||
|
||
user experience without the inclusion of elements from the real world. Milgram et al. (1995) define any user experience combining real and virtual objects as mixed reality, with mixed reality’s two sub forms being AR and AV.
|
||
The “MR-dominant view” is compelling and well-known due to its simplicity and flexibility (Skarbez et al., 2021). However, several issues are notable with this perspective. First, the MR-dominant view states that other realities (AR, AV) are a subclass of mixed reality (Milgram & Kishino, 1994) and that “AR and VR are related and that it is quite valid to consider the two concepts together” (Milgram et al., 1995, p. 283). However, given the differences in both designer goals and user experi ences associated with AR and VR, considering them together might be problematic. Second, some authors argue that the MR-dominant view distinguishes between AR and AV based on the proportion of real vs. virtual content (Leclet-Groux et al., 2013; Looser et al., 2004). This proportion-based interpretation, however, is not without limitations. For instance, consider a user wearing a pair of functional AR glasses integrating textual content into the person’s field of view. This use case would, in the MR-dominant view, be considered a “mixed reality envi ronment”, and, more specifically, represent either AR or AV, depending on the proportion of the user’s field of view that is covered by text. As we will show later, such examples contradict current industry practices. Third, some authors also interpret this view by the dichotomous distinction of whether virtual content is overlaid on the real-world (AR) or whether real objects are overlaid on virtual content (AV). While a dichotomous distinction contradicts the idea of a continuum, this distinction might become especially challenging in video see-through systems (where everything is presented on a digital screen), and it also may not matter to consumers. Fourth, real-world occurrences of AV are difficult to find. As we will show later, several industry informants with numerous years of experience were unfamiliar with this term, and a Google trend analysis3 supports this conclusion.
|
||
2.3.2. The “VR-dominant view” The “VR-dominant view” argues that VR is the main medium standing
|
||
above all other formats. For instance, Azuma (1997) suggests that AR “is a variation of Virtual Environments (VE), or Virtual Reality as it is more commonly called” (p. 2). In a similar vein, Guttentag (2010, p. 638)
|
||
3 https://trends.google.com/trends/explore?date=2004-01-01% 202021-08-06&q=%22augmented%20virtuality%22,%22augme nted%20reality%22,%22virtual%20reality%22,%22mixed%20reality%22.
|
||
|
||
3
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
Fig. 2. Schematic representation of prior “views” on new reality formats.
|
||
|
||
states that “this paper accepts augmented reality (AR) – the projection of computer-generated images onto a real-world view [ …] as a type of VR.” Wedel et al. (2020) recently argued that the term VR is sufficient and the use of the term AR is only necessary when a distinction is spe cifically needed in a given context. Hence, the VR-dominant view tends to classify AR as a sub form of VR and discusses mixed reality as something merging both VR and AR (Wedel et al., 2020), without further specifying the relationship between mixed reality on the one hand and VR/AR on the other. As with the MR-dominant view, given the substantially different goals and experiences of AR and VR, it seems dissatisfying to declare AR to be a subset of VR.
|
||
2.3.3. The “MR-centered view” Other scholars (e.g., Farshid et al., 2018) propose a continuum
|
||
including mixed reality in the center between AR and AV, which is surrounded by the real world and VR. In addition, and contrary to Milgram et al. (1995), mixed reality is not conceptualized as an umbrella term for these realities that include both real and virtual elements, but rather as a very specific type of reality that “combines what’s real with what’s possible” (Farshid et al., 2018, p. 660). Flavi´an et al. (2019) similarly designate “pure mixed reality” as a specific technology that fits between AR (defined as when virtuality overlaps reality) and AV (when reality overlaps virtuality). Overall, the MR-centered view makes an important distinction by segregating the real from the possible and follows Deleuze (1966), who first proposed this split between the real (e. g., the telephone on your desk) and the virtual (e.g., using Siri as a virtual assistant) to virtuality (e.g., playing a virtual game that has no connection to reality) (Farshid et al., 2018). However, Farshid’s distinction between VR (left side of the VR continuum) and virtuality (right side of the VR continuum) seems to remain ambiguous, especially given that mixed reality and AV are positioned between these two poles. For example, it remains unclear why VR is described as being “real”, whereas virtuality is described as being “possible” (cf. Fig. 1, p. 658). Furthermore, as with the “MR-dominant view” and the “VR-dominant view,” the “actual reality/virtual reality continuum” view does not define a role for XR.
|
||
2.3.4. The “extended reality view” A myriad of firms and consultants have developed their own ap
|
||
proaches and terminologies. XR (often used as an abbreviation for extended reality) is frequently employed as an umbrella term for a va riety of distinct concepts – most prominently AR and VR. The term mixed reality is often loosely and vaguely incorporated, typically as “a combination of AR and VR” and without specifying further what this means, while other authors specify mixed reality in more detail. For instance, Kunkel and Soechtig (2017, pp. 48–63), in a recent report from Deloitte, propose that in mixed reality, “the virtual and real worlds come together to create new environments in which both digital and physical objects—and their data—can coexist and interact with one another” (p. 49). Dalton (2021, p. 5) states that AR is “sometimes subcategorized as mixed reality” which represents “another form of AR”. He explains the
|
||
|
||
unique characteristics of mixed reality in technical terms, such as that “digital elements can be anchored to points in the physical environment” (p. 5), and discusses it in “contrast … [to being] … simply overlaid” (p. 5). This specification is basically the translation of technical mixed re ality characteristics (e.g., spatial anchors, stereoscopic 3D, etc.) on a higher level of abstraction in the user’s voice (“very realistic”), as dis cussed in Dwivedi et al. (2020). However, it is important to note that we observed a shift in the use of the term mixed reality toward “realistic AR” with the launch of the HoloLens device in 2016, which the experts in our study also confirmed. The question mark and the unclear boundaries (blurred circle) in Fig. 2 indicate that many industry pro fessionals have not yet scrutinized the organization of these terms under a clearly defined umbrella. Furthermore, this view has not yet received sufficient academic attention.
|
||
3. The need for an updated framework
|
||
In the previous section, we observed an inconsistent and incomplete use of new reality terminology. We also noted that three of the four different “views” are academically driven whereas the “extended reality view” is primarily informed by industry. Since the new reality field is shaped by various stakeholders, such as academics, commercial players (e.g., hardware and software providers, and their marketing de partments), industry associations, and so forth, we conclude that both academics and industry will benefit from reconceptualizing and orga nizing the field through expert informant opinion. Against this back ground, the current research aims at identifying, (re-)defining, distinguishing, and organizing relevant terms in a managerially focused framework by consolidating published research and input from a variety of expert informants. More formally, the proposed framework is designed to:
|
||
• include all relevant terms • provide an informant-driven definition • identify and explain core differences between relevant terms • organize these concepts into a coherent framework
|
||
4. Focus group and expert interviews as an iterative validation
|
||
We began our research with a detailed inspection of the academic literature, industry publications, and other publicly available materials (e.g., self-descriptions from companies). Based on the review of these materials, we developed a first tentative draft of the xReality framework which included definitions and delineations between and across different terminologies.
|
||
Next, we conducted a focus group with seven experienced industry practitioners, all with extensive but varied AR and VR expertise. We conducted the focus group via a video conferencing tool, which allowed participants to engage in active discussions and conversations (Bosco & Herman, 2010). The focus group lasted 1.5 h and was recorded. After a general introduction, participants were presented with the four views of
|
||
|
||
4
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
Fig. 3. XR (xReality) as an umbrella term for AR and VR.
|
||
|
||
new realities (described in section 2) as well as the draft of the xReality framework, and then asked for impressions and feedback on these ma terials. This procedure produced intense discussions regarding the existing draft. In addition, concepts and terminologies that were not part of the initial framework were discussed. Four participants later con tacted the research team with additional input. Following qualitative analysis of the focus group, the authors iteratively referred back to the literature and compared the results with documented views and findings.
|
||
Additionally, the authors conducted 15 qualitative interviews with AR and VR professionals of different backgrounds, foci, and tenure. The interviews lasted between 40 and 105 min and were guided by pre senting the framework as it evolved with each interview. Specifically, after each interview, we adjusted the model, revised terms, added new concepts whenever necessary, and documented additional feedback and suggestions. Four experts shared further documents, videos, or thoughts with us after the interviews. We ceased data collection when saturation was reached and additional interviews failed to generate novel insights (Saunders et al., 2018).
|
||
We identified both focus group and interview informants based on public presentations, recommendations, publications, and personal contacts. The sample included individuals with differing perspectives, terminologies, and backgrounds (e.g., academics, managers, developers, consultants, etc.), and Appendix 1 displays their demographics. Our empirical approach to integrating the voice of both industry and aca demic experts iterated “through data collection and analysis in such a way that preceding operations shape subsequent ones” (Spiggle, 1994, p. 495). The qualitative approach was appropriate as we were exploring a new and emerging area that is rapidly changing, and understanding the processes involved in designing and categorizing these interrelated concepts requires the flexibility availed by the qualitative process.
|
||
5. Development of the integrative xReality framework
|
||
Based on our review of related literature and the insights provided by our informants, we developed a contemporary classification of new media formats. We name this framework the “xReality framework” (see Figs. 3 and 4). In the subsequent sections, we explain each element in more detail.
|
||
5.1. XR: extended reality vs. xReality
|
||
|
||
this view. However, we also found that many experts felt that this conceptualization could be misleading, as the term “extended”, per definition, excludes VR since reality in VR is not extended but rather replaced. Some of our informants, for example JOE, indicated this immediately (“I do not say extended realities [to XR] because VR, to me, is not an extended reality but rather an alternative reality”), whereas others altered their stance after a discussion about the appropriateness of the term “extended” as including VR since reality is replaced. Some informants suggested alternative terms, such as “reality x” (DORIAN), “digital reality” (PAT), “new realities” (JOE), or sticking with “XR” (CARL). MARTIN specified this as follows: “XR, in my mind, always was where X is replaced with whatever; it’s ‘something R’“. We decided to follow the general consensus that X represents a placeholder for any digital reality format, embracing the notion of using XR as an abbrevi ation for X Reality (as, for instance, suggested by BILL), conceptualized as an established umbrella term for a variety of digital reality formats.
|
||
Proposition 1. We posit that X – in XR – represents a placeholder (similar to an X variable in algebra) for any form of new reality.
|
||
5.2. Strict separation of AR from VR and of experience from hardware
|
||
Section 2.3 above outlined four views of the conceptual space described in the XR literature, where most of them conceptualize AR and VR on the same continuum (Flavia´n et al., 2019; Milgram et al., 1995; Milgram & Kishino, 1994) or AR as a specific sub form of VR (e.g., Guttentag, 2010; Wedel et al., 2020). Contrasting these views, we found a general agreement4 among our informants that AR and VR represent “fundamentally different concepts” (CARLA), where typically different types of content are relevant (MAX) and need to be “separated” (ANNA), especially “from the user’s viewpoint” (JAKE). In short, as stated by GARY, AR and VR “are not the same thing at all […] I don’t think they should be considered on the same scale [ …] and it is better to have a split between those two”. RICK reported his observation that “many firms completely separate AR from VR’’. ANNA discussed “different purposes” of these two formats which are driven by different success factors; for instance, “how well the content is integrated in the reality” matters in AR, but not in VR (SAM).
|
||
Following suggestions from the human computer interaction litera ture (e.g., Hassenzahl & Tractinsky, 2006), we conclude that a concep tualization of new realities should be based on user experience, and that
|
||
|
||
The term XR is often used as a generic expression covering both AR and VR (Ço¨ltekin et al., 2020). Our informants supported this notion of an “umbrella term” where reality formats are “put in” (JAKE). Extant literature frequently establishes XR as an abbreviation for “extended reality” (Alcan˜iz et al., 2019), and some of our informants first echoed
|
||
|
||
4 We acknowledge that some informants (e.g., TRISTAN, MIKE) provided some less common examples where the differentiation might not be immedi ately clear (we discuss some of them in the discussion section). However, we found a general agreement that common use cases can be well separated into AR or VR.
|
||
|
||
5
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
Fig. 4. XReality (XR) framework: Augmented and Virtual Reality.
|
||
|
||
AR should be clearly differentiated from VR. This contradicts some streams of research (e.g., Milgram et al., 1995; Milgram & Kishino, 1994), yet echoes the views of nearly all of our informants. Specifically, we observed a general agreement that the distinction between AR and VR should be made based on whether the physical environment is, at least visually, part of the user’s experience or not. That is, any distinc tion based on the underlying hardware alone is not appropriate.
|
||
For example, one could use a “VR-branded device” with front cam eras and present video-see-through AR to a user. Although the device itself might be classified (or marketed) as “VR”, the user would in fact experience AR (similar to AR on a smartphone). Our informants corroborated this perspective. For instance, SVEN observed:
|
||
“You are either in an AR environment, so you see the real-world, or you are in a VR environment where […] you are basically in a digital environment. […]. You can move from one to the other – that is not happening very often now, the technology is not really there. But in the future, I see that as a thing that can happen. But you are only either in one, you can’t be in both at the same time”.
|
||
We also identified general differences between AR and VR (see Table 1). As discussed earlier, there are a variety of AR devices, whereas consumer VR is typically limited to HMDs (and in rare instances to caves or similar formats). Many of the informants emphasized AR technology’s potential to develop into something that is used always and everywhere (e.g., BEN), whereas VR devices – without substantial innovation – remain a device for temporary use. In addition, a precise understanding of the physical environment through tracking technology is necessary to realistically integrate virtual objects into the real-world in AR. On the other hand, in VR this is usually less crucial and often limited to collision avoidance, i.e., the identification of potentially dangerous real objects close to a user. Furthermore, from a practical perspective, specific content might be effective (e.g., accepted by consumers) in AR, but not in VR, and vice versa. Therefore, developing an application in one or the other requires an understanding of both the specific use case and the availability of devices within the user group.
|
||
The physical environment is, at least visually, replaced in VR and represents an experience where users “go in” (CARL). The notion “at least visually” is important since other external sensual stimuli (e.g., smell) are challenging to suppress. However, similar to AR, typical VR
|
||
|
||
systems are “primarily centered around vision” (Slater & Sanchez-Vives, 2016, p. 4). In contrast, the physical environment is extended and enriched (also by diminishing real objects) in AR, and thus, AR experi ences are driven by the experience of digital content within a physical space (BEN, JOE, RICK). Furthermore, barriers to the experience tend to
|
||
|
||
Table 1 Main differences between common AR and VR use cases.
|
||
|
||
Augmented Reality
|
||
|
||
Virtual Reality
|
||
|
||
Role of the local physical environment
|
||
Usage time frame (potential)
|
||
Typical Usage context
|
||
Technology
|
||
Physical risks
|
||
Privacy concerns Motion sickness Specific
|
||
Mechanism Typical Use Cases
|
||
|
||
Is extended/diminished
|
||
Enduring
|
||
Everywhere
|
||
Devices: Stationary, mobile, wearable, on-/in-body, projectors Display techniques: Video see-through displays Optical see-through displays Projection Collisions or accidents through distraction
|
||
The user and surrounding people Rarely applicable Local Presence
|
||
Situations where combined experiences of real and virtual content is beneficial (e.g., to compare sizes, e.g., of furniture) and possible (e. g., the home for the furniture already exists)
|
||
|
||
Is replaced
|
||
Temporarily
|
||
In a “secure” area (e.g., at home) or in specific contexts (e.g., therapy, amusement parks, shop etc.) Devices: Wearables (HMDs), caves (declining practical relevance) Display techniques: Video displays, projection
|
||
Collisions or accidents through disconnection with the real world The user
|
||
Significant Telepresence
|
||
Situations where the physical or story context does not exist (e.g., a fictitious game), is not accessible to a user (e.g., the moon, time travel) or where the actual physical context is not desirable (e.g., in training situations that would be dangerous in the real-world).
|
||
|
||
Note: We refer here to “generic” experiences and standard devices. There may be situations where these differences do not (fully) apply.
|
||
|
||
6
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
|
||
Fig. 5. Four examples on the Assisted-Mixed-Reality Continuum.
|
||
|
||
matter. In VR, users may feel “lost”, struggle with motion sickness, or fear collisions with physical objects (e.g., due to abstraction, complete mental immersion, and an inability to perceive the real world). In AR, physical threats may result from distraction or misinterpretation (e.g., by perceiving real objects as virtual), which can eventually also result in collisions with physical objects. Furthermore, users can compromise not only their own, but also other people’s privacy (Cowan et al., 2021; Lammerding et al., 2021; Rauschnabel et al., 2018). These aforemen tioned differences are also echoed in typical use cases. For example, both AR and VR can be effective for training and education. However, SVEN, for instance, argues that VR is effective in training before starting a new job, whereas AR allows training on the job. In more general terms, typical AR use cases usually emerge in situations where combined experiences of real and virtual content are beneficial (e.g., to compare sizes of furniture or clothing) and possible (e.g., when the space for a specific piece of furniture already exists). VR, in contrast, is preferred in situa tions where the physical context does not exist (e.g., a fictitious game), is not accessible to a user (e.g., the moon), or where the actual physical context is not desirable (e.g., in training situations that would be dangerous in the real world).
|
||
Furthermore, as with other media, both AR and VR have the po tential to cause psychological or physical harm – either intentionally, due to carelessness, or out of malicious intent. Examples include expe riencing a war scene in which people are killed in VR, the inclusion of scary or disturbing virtual objects in AR, or the design of visual content in a way causing nausea, headaches or even seizures, as was demon strated by the so-called Pokemon Shock in 1997 where strobe effects in a TV episode caused health issues among 600 viewers. Fig. 4 presents the final model related to Proposition 2.
|
||
Proposition 2. There is a need to separate AR from VR based on whether the physical environment is, at least visually, part of the user experience (=AR) or not (=VR).
|
||
5.3. Refining AR: the assisted versus mixed reality continuum
|
||
Following the general consensus of our expert informants, we sug gest that AR represents a combination of real and virtual content that is displayed in real-time. Furthermore, our empirical findings suggest that it is meaningful to distinguish between different types of AR (cf. Fig. 5). For example, workers can use AR glasses to obtain text-based work in structions overlaid on the physical environment (Mura et al., 2016), and tourists can gain access to overlaid information for places of interest when on a sightseeing tour (Han et al., 2013). Many of our informants (e. g., MEL, PAT, SVEN) used the term “assisted reality” to describe this form of AR because the purpose of the virtual objects is to assist the user in obtaining a better understanding of the physical environment rather than to merge virtual objects with the real world. On the other hand, our informants described a highly sophisticated form or AR that tracks and
|
||
|
||
maps the environment in three dimensions, and which integrates digital objects realistically and seamlessly into the user’s perception of the real world. This seamless integration of virtual and real objects is termed “mixed reality” because the two realities (real and virtual) merge and, in its extreme form, become indistinguishable to the user (MEL, PAT, SVEN, MIKE).
|
||
Although we acknowledge that the term “mixed reality” is often used differently in the literature [compare, e.g., also Speicher et al. (2019) who observe that different authors treat mixed reality either as a syno nym for AR, as a combination of AR and VR, or as a “stronger” version of AR], we argue that our conceptualization of the assisted reality – mixed reality continuum is meaningful and beneficial for academia and prac titioners. First, we found substantial support from our informants for this conceptualization, especially from those informants who felt that the traditional mixed reality continuum suggested by Milgram et al. (1995) has become conceptually problematic, given the development of AR over the last 25 years. For example, one of our informants emphasized that “there is a big step from assisted into mixed reality” (SVEN) within the range of the AR continuum. Others associated mixed reality with “hybrid experiences” (BILL), where virtual content is “interacting with physical objects and logically matching” (DORIAN), or user experiences where “you cannot really tell anymore what is real from what not” (LENA). Second, this view is echoed by recent industry publications from reputable players in the XR market (e.g., Dalton, 2021) who use the term mixed reality similarly to our conceptualization.5 Third, the term “mixed” is etymologically closely related to the idea of a realistic inte gration of real and virtual content. For instance, the Oxford English Dictionary (2021a) defines the term mix as “combine and put together to form one substance of mass” (here: one experience) and uses “oil and water do not mix” as a negation example. This example metaphorically represents the opposite end of the continuum, assisted reality, where virtual content (oil) is just overlaid (“floating”) on top of the real world (“water”) – or, as stated by MIKE, “floating in front” of the user.
|
||
The discussions with our informants about the differences between assisted and mixed reality centered around the term realism (e.g., LENA), and identified different views of what this term means to them. Some argued that fictitious characters (e.g., monsters) can, per defini tion, never be realistic, whereas others discussed this issue through the lens of “suspension of disbelief”, which describes users’ willingness to suppress information that contradicts real-world knowledge (Weibel et al., 2015). Linking these observations to the literature, we identified
|
||
5 The experts in the study highlighted their associations of mixed reality with the Microsoft Hololens device from 2016 (e.g., PAT, ANNA; RICK: “a super marketing term”), which was marketed as the first technology that realistically integrates, rather than overlays, virtual content. Others stated the term is, in their view, poorly defined and used without a clear understanding of what it means (e.g., JAKE).
|
||
|
||
7
|
||
|
||
P.A. Rauschnabel et al.
|
||
the term “local presence6” to best describe this distinguishing factor between assisted and mixed reality. Drawing on prior research (e.g., Lombard & Ditton, 1997; Smink et al., 2020; Spagnolli et al., 2009; Verhagen et al., 2014; Vonkeman et al., 2017), we define local presence as the degree to which a user experiences AR objects as being actually present in his or her own physical environment. In assisted reality, content is perceived as clearly artificial and overlaid, and thus, not perceived as being actually there. In contrast, when it comes to mixed reality, users experience virtual content as being actually in their physical environment (e.g., a decorative vase on a table or a flying monster in a game).
|
||
Proposition 3. Assisted reality and mixed reality are the opposite poles of the AR continuum. This categorization depends on the level of local presence perceived by the user.
|
||
In addition, two more terms were subject to the discussion: First, several informants mentioned diminished reality where physical objects are omitted from user perception. In other words, the technology “era ses” objects from the real-world by overlaying them with virtual objects (Mann & Fung, 2002). The xReality framework appropriately handles this emerging concept. At the assisted reality endpoint, diminished re ality would utilize an unrealistic overlay (e.g., a censor bar blurring content). Mixed reality would seamlessly remove perception of a real object in a way that is difficult or impossible to detect by users. An example use case might be an ad blocker in AR glasses that realistically overlays virtual content over environmental ads, allowing users to experience an “ad free” environment (Rauschnabel, 2021).
|
||
Second, when discussing different reality views (e.g., the MRdominant view or the MR-centered view), we asked informants about their view of the concept of AV. Surprisingly, many informants were not aware of the term, and others consciously choose not to use it (e.g., BILL, DORIAN, GARRY, MIKE). Still others see it as “maybe a niche” (JOE) with limited use cases, or associated it with green screen technology (RICK) or other non-XR formats.
|
||
Throughout our research, we identified numerous characteristics of AR that determine the position of an AR experience on the assisted re ality – mixed reality continuum. Before discussing them in more detail, we need to acknowledge two premises of our framework. First, AR/VR conceptualizations based on Milgram et al.’s (1995) reality–virtuality continuum determine the specific type of reality via the proportion of visual “content that is real versus how much is computer-generated” (Looser et al., 2004, p. 2; see also; Leclet-Groux et al., 2013). For example, in the Milgram et al. (1995) view, environments with a larger proportion of real objects are termed AR, and environments with a larger proportion of virtual elements are termed AV. However, we argue that such a “view of proportions” remains limited and does not acknowledge substantial changes and developments in recent AR technology. Hence, our AR continuum considers aspects related to the type of content (e.g., the quality, transition, and integration of real vs. virtual objects) rather than merely the proportion between these elements. Second, we suggest that mixed reality is not per se “better” than assisted reality – rather, this depends on the general context. Hence, user goals determine whether users perceive one specific AR application as being better than another one. For example, even though mixed reality provides a substantially
|
||
6 Note that in the extant literature, similar terms such as “local presence”, “object presence”, and “spatial presence” are used, among others. We deliber ately opted for using local presence to avoid misunderstandings among different scientific communities. For example, “spatial presence” could be un derstood as a term to describe the spatial relationship between the physical location of the user and the virtual location in which he or she is. In this case, spatial presence would be ambiguous. Likewise, “object presence” refers to the object-focused appearance or position of virtual objects (in AR or VR), rather than a user’s perception that virtual content is perceived as actually being in his or her local physical environment.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
higher integration of virtual and real objects than assisted reality, assisted reality may very well be the superior environment if the goal of the user is to enrich the real environment with factual information. In this example, a high integration of virtual and real elements might distract or confuse the user rather than being able to generate benefits. On the other hand, in highly hedonic contexts (e.g., games), a high integration between virtual and real objects (leaning towards mixed reality) is likely to improve user experience. Finally, given our concep tualization of AR as an experience, we argue that personal user char acteristics (e.g., expectations, prior experiences with AR, etc.) determine how realistically they perceive specific uses.
|
||
The following paragraphs outline elements that our informants identified as the most important drivers of the assisted/mixed reality distinction. It is important to note that the factors listed below are not presented as either exhaustive or perpetual, as other factors may play a role and some factors may lose importance as both user expectations and technology change.
|
||
5.3.1. Content stability and persistence There are two approaches to placing virtual content in the real world
|
||
(Jaekl et al., 2002), and these approaches determine content stability. Head-stable content moves according to the orientation of the user’s head and is appropriate in cases where no relationship between virtual and real-world objects exist or where it is essential that users can quickly process information (e.g., text or notifications). This approach aligns well with the notion of assisted reality. On the other hand, world-stable content is anchored in a fixed position within the 3D space. This approach is commonly used in cases where a relationship exists between a virtual and a physical object and, hence, is prominent in mixed reality applications. From a technical point of view, world-stable content re quires extensive tracking technology so that virtual content can be rendered in 3D space realistically and in registration with both real and virtual objects (Keil et al., 2019; Tamura et al., 2001).
|
||
Persistence is a specific characteristic of world-stable content that refers to how augmented content is spatially attached to specific phys ical objects (e.g., a digital vase on a user’s physical desk), or attached to a specific geographic location through geo-coordinates (Bachras et al., 2019). Augmented content is required to respond to movements within the physical world. Informant MEL stated that mixed reality experiences allow users “to place an object in the space, to turn around and back, and it still appears at the same place” and added that this does not always work well with all technologies, such as those that suffer from calibra tion drift (i.e., a loss of calibration quality). RICK highlighted the importance of world-wide and device-independent platforms (“mirror worlds’’) for applications that enable AR content to persist in a specific location forever and for multiple users to access (AR metaverses).
|
||
5.3.2. Dimensionality of content The type of display influences how a perception of depth can be
|
||
created in AR applications (Greene et al., 2021). In general, the pro cessing of depth information based on binocular disparity requires the use of stereoscopic AR displays (Heinrich et al., 2019). Alternatively, if no stereoscopic display is available (for example, when using a smart phone), other depth cues such as perspective, occlusion or shading could be used. The visual content displayed in assisted reality is typically 2D (e.g., text), whereas content toward the mixed reality endpoint is usually 3D, and thus, typically increases local presence.
|
||
5.3.3. Contextual embedding In general terms, contextual embedding refers to how cues in the
|
||
environment are situated and interpreted within a specific context (cf. Hornecker, 2010). Assisted reality provides a minimum level of contextual embedding as the technology is not fully aware of the context. Here, the technology requires the user to update when a specific step in a process has been completed so that it can advance to display the next step (processual context). Mixed reality, on the other hand,
|
||
|
||
8
|
||
|
||
P.A. Rauschnabel et al.
|
||
recognizes and identifies objects in the surroundings and can also track the user’s progress in a task, thereby advancing the display as progress is made (physical context). As AR moves from assisted to mixed reality, the requirements for tracking and understanding the environment increase (c.f., Fig. 4). For instance, LUKE commented on the need to understand and incorporate real world lighting characteristics to model realistic shadows. RICK discussed varifocal technology that allows users to perceive virtual objects displayed a centimeter from the eye as if they were far away. Advancements in tracking technology (e.g., markerless tracking, LIDAR scanners, and recent versions of Apple’s AR Kit and Google’s AR Core) substantially improve the perception of embedded content. Hence, embedding in the context is reflected by perceived augmentation quality (Rauschnabel et al., 2019) which in turn leads to higher levels of local presence (cf. Daassi & Debbabi, 2021) and thus moves the user experience more towards the mixed reality end of the AR continuum.
|
||
5.3.4. Technological embodiment The concept of technological embodiment describes how AR tech
|
||
nology can become an extension of a user’s body (Tussyadiah et al., 2017). Flavia´n et al. (2019) suggest that wearable devices increase and stationary devices decrease embodiment. Furthermore, our informants reported industry developments on AR brain interfaces that could directly integrate virtual information into the optic nerve.
|
||
An integral aspect of technological embodiment is how interaction is designed. For interaction with (2D) content on the assisted reality end of the continuum, techniques that have been designed for other contexts in which interaction with 2D content is common (e.g., desktop and smartphone applications) typically work. Yet novel approaches may be required for interaction with content in 3D space, with the objective of enabling intuitive interaction and ultimately supporting a stronger sense of technological embodiment. Examples include, but are not limited to, interfaces based on speech, gaze, EEG, and EMG. Overall, assisted reality is usually characterized by lower levels of technological embodiment, as opposed to higher levels of technological embodiment for mixed reality.
|
||
5.3.5. Interactivity In the context of technology-mediated communication, interactivity
|
||
has been defined as “the extent to which users can participate in modifying the form and content of a mediated environment in real time” (Steuer, 1992, p. 84). Previous research suggests that interactivity is important for an effective AR experience (Park & Yoo, 2020; Yim et al., 2017). Park and Yoo (2020) find that perceived interactivity increases mental imagery, which in turn improves consumers’ attitudes towards a product and increases purchase intentions in an AR-mediated shopping context. In a similar vein, Yim et al. (2017) show that higher levels of interactivity in AR applications increase consumers’ perceived useful ness and enjoyment. The extant literature acknowledges that inter activity and realism may interact in AR experiences (Montero et al., 2019), yet higher levels of interactivity may not necessarily result in higher perceived local presence. For instance, assisted reality applica tions can provide a high level of interactivity by allowing users to manipulate and control superimposed objects such as text or virtual icons. However, such applications may not provide high levels of local presence as virtual objects and the physical environment are (inten tionally) not seamlessly integrated.
|
||
5.3.6. Shared and social experiences Content in traditional AR has typically been restricted to a single user
|
||
at a time. However, advances in technology create the opportunity for multiple users to experience the same (virtual) content together (Chy linski et al., 2020; Hilken et al., 2020; Lebeck et al., 2018), which is often discussed in terms of co-presence (Nowak, 2001) and co-experience (Battarbee & Koskinen, 2008). Shared experiences build upon persis tence (as outlined above) since the content needs to remain in the same location for multiple users to perceive it (e.g., in an AR metaverse).
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Shared experiences can not only enable interactions between one user and the (real and virtual) content, but also between users. This enables a form of interaction that is more similar (i.e., realistic) to real-world in teractions. For example, Carrozzi et al. (2019) found that shared AR experiences enhance feelings of psychological ownership for virtual objects, and Hilken et al. (2020) showed that shared AR experiences can enhance users’ social empowerment and improve joint decision-making. Hence, we suggest that with increasing levels of shared experiences, the user experience shifts towards mixed reality in the AR continuum. In addition, yet not fully explored, shared experiences might also occur with anthropomorphic or animalistic virtual creatures (e.g., a user-pet relationship with an AR animal).
|
||
5.3.7. Augmentational/environmental control Environmental control refers to the level of control users have of
|
||
content in AR applications and how this content interacts with real objects (Brooks, 1990). Previous research suggests that increased levels of perceived control raise the need for user identification and psycho logical ownership of the AR application (Carrozzi et al., 2019). Aug mentational control is very low or potentially non-existent in a typical assisted reality application. Control increases in common mobile AR apps (e.g., a makeup app where users can augment their eyes, lips, etc.) and tends to be very high in mixed reality where users can manipulate virtual objects and enhance real objects. For example, in a mixed reality environment, a user might be able to click on a virtual light switch that is connected to a real lamp.
|
||
5.4. Refining VR: the atomistic vs. Holistic Virtual reality continuum
|
||
Compared to their categorizations of AR and mixed reality, Milgram and Kishino (1994) as well as many authors building on Milgram et al.‘s work, remain surprisingly silent when it comes to specific forms or types of VR (Farshid et al., 2018; Flavia´n et al., 2019; Milgram et al., 1995; Milgram & Kishino, 1994). However, given the drastic advances in VR technology and applications (Hollebeek et al., 2020), a distinction be tween different types of VR will further advance our understanding of virtual environments.
|
||
Informed by the experts and inspired by prior research, we argue that telepresence allows us to distinguish between different forms of VR. Hence, telepresence with its notion of “being there” (Rodríguez-Ardura & Martínez-Lo´pez, 2014) is clearly delineated from local presence, as discussed in the section on AR. More formally, we draw on existing research (Lim & Ayyagari, 2018; Mantovani & Riva, 1999; Mollen & Wilson, 2010; Steuer, 1992; Tussyadiah et al., 2018) and define tele presence as the degree to which a user feels present in the virtual rather than the physical environment. We acknowledge that existing research uses several variations of this term (e.g., simply “presence”, Mantovani & Riva, 1999), virtual presence (Sheridan, 1992), or mediated presence (Bourdon, 2020) synonymously for our definition of telepresence. However, by adding the prefix “tele,” we highlight the distinction from local presence. More specifically, according to the Oxford English Dic tionary (2021b), the prefix “tele” is borrowed from Greek and denotes or relates to “action, observation, or communication at, over, or across a distance, or denoting devices used for this.” In this sense, telepresence refers to presence mediated through a fully virtual environment (Man tovani & Riva, 1999).
|
||
Based on the notion of telepresence, we propose VR applications to be positioned on a continuum between atomistic and holistic VR expe riences. On the one hand, atomistic VR refers to applications of VR for which the quality of the user experience is often secondary to some other goal. For example, VR can be used for training or modeling physical spaces (such as virtual blueprints in construction applications) where the completion of a task is a primary concern. In these cases, the user’s perception of telepresence is less important than accomplishing a spe cific goal or outcome. On the other hand, holistic VR is signified by a VR experience that is nearly indistinguishable from a real-world experience
|
||
|
||
9
|
||
|
||
P.A. Rauschnabel et al.
|
||
in the mind of the user; in these cases, it is the perception that the user feels present in the virtual world that supersedes other aspects.
|
||
Proposition 4. Atomistic VR and Holistic VR are the opposite poles of the VR continuum, and this categorization depends on the degree of telepresence perceived by the user.
|
||
Two aspects primarily emerged from our discussions with experts in this area. First, the extant literature often refers to VR experiences consistent with Hassenzahl’s (2018) classification of VR products or services based on their hedonic or pragmatic qualities. Whereas prag matic quality refers to the perceived usefulness and ease of use, hedonic quality refers to an intrinsic “joy of use” that accompanies the user’s experience. According to our conceptualization of atomistic VR, VR experiences at this end of the spectrum are likely to have a higher pragmatic quality, while holistic VR might also be characterized by a higher hedonic quality as perceived by the user. Second, this dichotomy is analogous to the concept of instrumental versus terminal use. In instrumental examples, the user employs VR technology as a means to an end as the technology is an instrument designed to accomplish some other goal. The word “terminal” signifies an end, and holistic VR ex periences are often an end in and of themselves. In the following section, we categorize properties of VR applications and explain how they impact the position of a VR experience on the VR continuum.
|
||
5.4.1. Content stability and persistence Like in AR, (portions of) content in VR can be stable with respect to
|
||
the user’s head movements (head or user stability) or the surrounding virtual environment (world stability) (Sipatchin et al., 2021). A VR experience can contain both head and world stable content at the same time. For example, the environment in a game can be world-stable, but objects belonging to the user (e.g., a map) or status information would be displayed in a head-stable manner.
|
||
Similar to AR, persistence in VR means that virtual objects are attached to a fixed location in 3D space. In contrast to AR (where objects are fixed in the physical world), persistent objects in VR are attached to a digital 3D position in the virtual world (Zielinski et al., 2015). The im plications are similar, as different users can see and potentially manip ulate these digital objects by navigating to a distinct location virtually.
|
||
5.4.2. Dimensionality of content In VR applications, the dimensionality of content refers to how the
|
||
elements forming the virtual world are rendered, typically in 3D (Zie linski et al., 2015). A specific example, where the virtual world is only rendered in 2D, is 360-degree videos. VR users require depth cues to allow them to judge the size of and distance to virtual objects and to perceive high levels of telepresence (LENA). Note that designers can deliberately design 2D or 3D content based on its purpose. For example, information on objects or status information about the user’s location could be designed using 2D objects even though the user navigates in a 3D environment, and vice versa.
|
||
5.4.3. Contextual embedding Contextual embedding is relevant for many XR applications, but it is
|
||
different in VR compared to AR. In order to embed content in context, knowledge of the context is required. In VR, the context is typically entirely virtual (e.g., a virtual room), where such a 3D map is already part of the application, whereas in AR, this is much more challenging, as the context must be obtained externally (as outlined in the AR section).
|
||
5.4.4. Technological embodiment Similar to the discussion in the AR section, advances in technology
|
||
and hardware support higher levels of technological embodiment in VR applications (Flavia´n et al., 2021). However, the nature of this embodiment differs between AR and VR applications. Whereas a high level of technological embodiment in AR is based on unobtrusiveness and registration between real and augmented content, a high-level
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
embodiment in VR depends on the individual’s perception of tele presence in the virtual world. Instead of registering the real to the augmented content, technological embodiment in VR remains in the mind of the user, yet can be enhanced by technologies such as smart gloves.
|
||
5.4.5. Interactivity How users interact with the application, the controllers, and other
|
||
users in VR strongly contributes to the perception of telepresence. This differs from AR where the typical objective is to make interaction closely resemble interactions in the physical world. Traditional interaction techniques (i.e., the way in which I/O devices are used for interactive tasks) were developed for computer-human interaction through desktop computer interfaces, and these generally do not translate well to VR interaction (cf. Flavia´n et al., 2021).
|
||
Users often interact in VR using a controller, their hands, their gaze, or a combination thereof. These approaches to interaction can pose specific challenges. Because the entire VR world is synthetic, users are not able to see their own hands, and a virtual representation of the hand (or any other part of the user) is needed. Informant LENA highlighted the importance of self-perception in the virtual world, and research has shown that the realism of projected body parts has a substantial effect on both perceptions of embodiment (Argelaguet et al., 2016) and task performance (Knierim et al., 2018). VR systems that project represen tations of the human body require motion capture systems (cf. Liebers et al., 2021; Pfeuffer et al., 2019) to track user movements, and body movements are often extrapolated from the position and orientation of the headset and/or other controllers (Yung et al., 2021). For atomistic VR applications, it might be sufficient if the user – in particular the parts of their body with which they interact (arm, finger, foot) - is shown in a very simplified way, or if only the controller is visible. For example, if users are positioning objects in VR to build a model of a product, they only need to perceive the controller and not their entire body. In contrast, holistic applications typically require a higher level of self-perception as body position may significantly influence the user’s performance. For example, VR applications modeling fine-grained motor-control like typing (Knierim et al., 2018) or playing piano (Fanger et al., 2020) require higher levels of self-perception.
|
||
Furthermore, when objects are displayed beyond arm’s reach where direct interaction with the object is not possible, more indirect inter action techniques, such as using a laser pointer metaphor (Hoppe et al., 2018) or the user’s gaze can be employed. Many headsets now employ eye tracking technology (e.g., Pico Neo or HTC Vive Pro Eye) that can be used to select objects at a distance using simple visual focus. Gaze can be paired with other techniques for operations such as translating or rotating an object (Pfeuffer et al., 2014). The choice of such techniques may influence telepresence.
|
||
A particular form of interaction in VR is navigation, and in particular, locomotion. While users in AR simply physically move through the environment, other concepts are needed in VR. A high-quality experi ence while navigating is crucial for users to experience high levels of telepresence. Depending on the degrees of freedom, locomotion in VR may be realized using simple walking patterns. VR systems track the users’ movements and map them to the virtual world. Yet the physical world in which the user experiences the VR application is usually limited in terms of space (e.g., the user’s living room). To accommodate this, treadmills or so-called VR walkers can be used. As an alternative form, locomotion in VR can be realized using controllers to proceed through the environment. However, this approach typically results in a higher cognitive load for the user.
|
||
Summarizing, the extant literature consistently finds that higher levels of interactivity increase perceived telepresence (Beck et al., 2019; Kim & Ko, 2019; Mütterlein, 2018). This perspective is also supported by a large number of our informants. For example, the focus group dis cussed telepresence for atomistic and holistic VR and indicated that “looking around” 360◦ is a very “simple form of interaction” (JOE).
|
||
|
||
10
|
||
|
||
P.A. Rauschnabel et al.
|
||
Interaction within VR must be logical, as stated by DORIAN: “It’s not only the extent or capacity of interaction that I can do and the quantity of interaction, but whether it makes sense or not. The interaction, to me, must be logical”. Supporting this reasoning, MIKE argued that “inter activity leads to high levels of [tele]presence because you are more connected to the experience when you are making a choice, when you are affecting the outcome of the experience in some way” and uses ex amples of lower (e.g., making a choice) and higher (e.g., picking up a tool and unscrewing a screw) interactions. Hence, we posit that low levels of interactivity typically indicate a more atomistic form of VR, whereas high levels of interactivity lead to holistic representations of VR.
|
||
5.4.6. Shared and social experience To enable social interaction – either with virtual or real characters –
|
||
VR experiences rely on virtual representations, commonly referred to as avatars (Schroeder, 2002). However, when employing avatars, non-verbal communication cues pose a particular challenge. For example, sophisticated user representations record and display not only overall avatar movement, but also the specific head and eye movements of these avatars. Failure to accurately animate the head and the eyes may lead to avatars being perceived as unrealistic or inattentive, and this can significantly reduce users’ perception of telepresence (Itti et al., 2003). For avatars depicting real world characters, representing their appearance and behavior in realistic ways is challenging. Furthermore, whereas creating static avatar models is less problematic, adding motion adds complexity. Typically, the more appearance and behavior coincide with what users would expect in the physical world, the higher the perceived telepresence. However, our informants pointed out that tel epresence might be negatively influenced by content that is too realistic. This is particularly true when displaying representations of humans, as this evokes strangely familiar feelings or eeriness and revulsion, an ef fect that is commonly referred to as the uncanny valley (MacDorman et al., 2009; also suggested by our informants, e.g., BILL). In addition, whereas AR is often built to accommodate multiple users, doing so in a shared VR environment poses challenges for designers. For example, when implementing conversations with multiple users, VR environ ments need to account for the distance to the user (e.g., the closer a communication partner, the louder the voice of this avatar should be). If not done properly, this might negatively influence the experience. An example where distance-based volume of other users’ voices is imple mented is Mozilla Hubs.
|
||
5.4.7. Perceptual experience To maximize the feeling of telepresence, a plausible virtual world
|
||
needs to be presented to the user (Lee, 2004) where aspects such as the quality of the graphics, dimensionality of content, and self-perception play an integral role (DORIAN). When events in the environment correlate with users’ actions and meet their expectations of how objects and people are expected to behave, plausibility increases because they feel that the events are really happening (Slater & Sanchez-Vives, 2016).
|
||
Graphics are fundamental in driving user perceptions of telepresence in VR (Slater & Sanchez-Vives, 2016). This aspect depends on the hardware, such as the resolution of the VR display device, the visual field-of-view, and the frame rate at which the graphic updates (Bowman & McMahan, 2007). For example, for holistic VR, HMDs should allow for high resolution to increase the sense of immersion (CARLA) – optimally as close as possible to the resolution of the human eye. Trackers should update just as quickly to translate user feedback instantly. The quality of the VR model plays an important role in the determination of atomistic vs. holistic VR as well. There is a spectrum ranging from rendering simple abstract geometric shapes to highly realistic objects (Bierbaum et al., 2001) with perfect texturing and shading, making the virtual world closely resemble what users expect from the real world. For many atomistic use cases (i.e., training, familiarization, etc.), simpler forms and shapes might be appropriate, but for holistic use cases, high-quality
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
graphics are more likely to result in increased levels of perceived telepresence.
|
||
Despite a frequent focus on visual output, it is important to acknowledge that multiple human senses are part of a VR experience. The feeling of telepresence is strongly influenced by how well a VR experience communicates through all of our senses (Baus & Bouchard, 2017), and is most positive when congruent (Flavia´n et al., 2021; Petit et al., 2019). Several experts referred to these communication options as modalities (visual, audio, tactile), or different feedback channels through which people can perceive the world (CARLA, LENA, PAT). For example, the lack of a haptic experience poses a challenge when inter acting with virtual objects. For atomistic VR, it might be sufficient if the VR application primarily focuses on the senses required to accomplish the main task, even though a specific combination of senses may posi tively influence task performance. To increase telepresence for holistic VR use cases, approaches like sensory substitution (e.g., replacing the lack of haptic feedback with visual feedback), providing appropriate controllers that match the physical properties of the virtual object (for example, a sphere when interacting with a globe; Englmeier et al., 2020), or using electric muscle stimulation to simulate forces (Lopes et al., 2018) may be employed.
|
||
The feeling of telepresence in VR is equally influenced by the phys ical behavior of virtual objects – i.e., the laws of physics – as mentioned by LENA: “If something is falling down, it would not fall down and stay on the floor. It would basically bump up, like physical laws would have to be integrated into the virtual object.” This may stand in contrast to AR, where the physical behavior of a virtual object is expected to match the laws of physics on earth. For example, when putting users in a VR application where they can experience being on the moon, gravity should match the physical laws on the moon. However, there are certain exceptions to this rule. For example, some virtual worlds allow users to fly or teleport (Hinsch & Bloch, 2009). These capabilities, while not in line with the expected laws of physics, are novel elements of the virtual world that are embraced by users specifically for this reason. Informants used the term “suspension of disbelief” as a component of certain virtual worlds (Steffen et al., 2019). However, we suggest that in most cases, the system’s behavior in terms of gravity, movement, or size of objects should be internally consistent to create a higher level of telepresence.
|
||
5.4.8. Motion sickness A specific challenge in VR is motion sickness, also referred to as VR
|
||
sickness (Mai & Steinbrecher, 2018). It typically results from a mismatch between users’ actual movements and the movements that they perceive through the virtual world but is also influenced by human factors (Chang et al., 2020). A common cause for this phenomenon is latency in VR rendering, i.e., the system is not capable of responding to users’ move ments in real-time (Ko¨se et al., 2020). As technology advances, sensing user motion, calculating the required changes in the virtual scene, and rendering content in real-time will improve and continue to attenuate this issue. However, it currently poses a challenge specifically in highly sophisticated VR environments which require substantial computing power. For atomistic applications where the time users spend in VR is rather short, motion sickness might present less of an issue than holistic VR applications, where users spend a considerable amount of time (Ruddle, 2004). A summary of sections 4.3 and 4.4 is given in Table 2.
|
||
6. Discussion & implications
|
||
Due to their immense opportunities in many disciplines, AR, VR, and related reality formats have recently received increased attention. En thusiasts from companies and research institutions have developed fascinating experiences both through augmentation of the real world (AR) and the creation of virtual worlds (VR). Scholars and industry practitioners, coping with a rapid evolution in this field, have defined and organized terms associated with these developments. Twenty-five years ago, Milgram outlined a technology-focused continuum leading
|
||
|
||
11
|
||
|
||
P.A. Rauschnabel et al.
|
||
|
||
Table 2 How various factors drive local presence and telepresence in common XR use cases and devices.
|
||
|
||
AR (local presence)
|
||
|
||
VR (telepresence)
|
||
|
||
Content Stability Dimensionality of
|
||
Content Contextual
|
||
Embedding
|
||
Technological Embodiment
|
||
Interactivity
|
||
Shared Social Experience
|
||
Augmentational control
|
||
Motion sickness
|
||
|
||
Head stable, world stable, persistence incl. Geocoordinates 3D leads to higher realism and thus, local presence Processual context: AR is aligned with a user’s tasks Contextual embedding: Content aligns with the realworld Contact lenses or AR smart glasses enable higher levels of embodiment Realistic interaction techniques can increase technological embodiment High levels of interactivity may not necessarily result in higher local presence Navigation is similar to the real world Real and virtual objects can interact with each other
|
||
Real-world interactions with other people are possible Avatars are only required if different users experience the same content from different physical locations More control over real objects that are augmented leads to a more realistic experience Less relevant in AR
|
||
|
||
Head stable and VR-world stable
|
||
3D leads to higher telepresence Processual context: VR is aligned with a user’s tasks Contextual embedding: Content aligns with other objects in the virtual world Emerges through telepresence
|
||
The ability to interact with content provides a more plausible and engaging environment and thus, higher levels of telepresence Navigation might be restricted due to limited physical space Requires avatars
|
||
N/A
|
||
Mismatch between users’ actual movements and the movements that they perceive through the virtual world
|
||
|
||
from the real to the virtual, with mixed reality between these poles. Other authors (e.g., Farshid et al., 2018; Flavia´n et al., 2019) defined mixed reality as a combination of AR and VR, while industry (e.g., Kunkel & Soechtig, 2017, pp. 48–63) has proposed new terms and repurposed (or, as stated by some experts, “misused”) existing ones. In short, the current literature contains inconsistent and often conflicting conceptualizations of these realities, resulting in confusion amongst academics, users, and practitioners. Meanwhile, the AR and VR in dustries have steadily matured to generate multiple billions in revenue each year.
|
||
Based on an intense review of the academic literature, industry publications, and expert input, we propose a complementary approach to define, organize, and conceptualize common reality formats. More specifically, the xReality framework separates AR from VR based on whether the physical environment plays a role in the user’s experience or not. If yes, the experience is AR; if no, and the experience is virtual, it is VR. In order to specify AR and VR in more detail, the framework provides two continua: the AR continuum ranges from assisted to mixed reality with local presence forming the core distinction between poles. The VR continuum ranges from atomistic to holistic, and the level of telepresence is the primary discriminating factor between these poles. Our findings provide a series of implications for the emerging XR discipline. Importantly, rather than separating managerial from theo retical implications, we sought to understand these perspectives and propose an approach that incorporates both. We argue that many cur rent discrepancies exist because academia and industry management are conceptually separated, and the current work attempts to consolidate
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
these divergent perspectives. XR is an overarching term used primarily by practitioners to describe
|
||
“all” forms of new realities (e.g., Dalton, 2021). XR subsumes both AR and VR, as well as their various sub forms. Contrary to extant research, we propose that the term “extended reality” might be misleading since it does not include VR (where reality is replaced, not extended). Therefore, we propose to maintain the term XR (Dwivedi et al., 2020), but use it as an abbreviation for xReality. Practically speaking, the variable x serves as a placeholder for Augmented, Assisted, Mixed, Virtual, Atomistic Virtual, Holistic Virtual, or Diminished Reality.
|
||
6.1. The xReality framework
|
||
The proposed framework presents XR as an umbrella term, with two distinct sub streams: AR and VR, which contain their own continua. This conceptualization differs from existing classifications (e.g., Farshid et al., 2018; Flavia´n et al., 2019; Milgram et al., 1995; Milgram & Kishino, 1994) in which AR and VR are located on the same continuum. Likewise, to the best of our knowledge, this framework is the first to include all commonly used terms in a coherent framework, including AR, VR, XR, mixed reality, and assisted reality. For instance, many older (e.g., Milgram & Kishino, 1994) and more recent (e.g., Farshid et al., 2018; Flavia´n et al., 2019) frameworks remain silent on some aspects of reality (e.g., assisted reality). Industry frameworks sometimes incorpo rate the term XR but it is frequently not used consistently.
|
||
However, contrary to the MR-dominant (e.g., Milgram & Kishino, 1994) and MR-centered view (e.g., Farshid et al., 2018; Flavi´an et al., 2019), our framework excludes AV. We justify this based on the view points expressed by the academic and industrial informants along with our inspection of both recent academic literature and online search trends. Moreover, following Looser et al. (2004) and Leclet-Groux et al. (2013), the AR/AV distinction is based on the proportion of virtual versus real content, and this is difficult to quantify. Other perspectives suggest that the AR/AV distinction is based on whether virtual content is augmented to the real-world, or if the real world is mapped to the digital content, but users may not perceive any difference between these two approaches. Our framework simply argues that if the physical environ ment is part of the user experience, then it is some sort of AR; if not, it is VR.
|
||
Our distinction between AR and VR is not dependent on the devices a person uses, but rather complements more technology-focused defini tions (e.g., Azuma, 1997; Zhou et al., 2008). For instance, one could consider a wearable device occluding the real-world from the user as VR, but this device may also include cameras to capture the real world for AR applications. Potentially, this would allow users to switch from a VR mode to a video see-through AR mode. However, while the hardware might accommodate both AR and VR, a user could only be in either AR or VR at any given time. Furthermore, shared experiences, often dis cussed as “metaverses” (such as spatial. io), could also fit into the framework. Here, multiple users located in different physical locations could interact in a shared experience through either AR or VR; some seeing the others as holographic avatars and some interacting from a VR environment. Our framework also acknowledges that not all AR expe riences are equal, and the same is true for VR. More specifically, we propose separate continua for AR and VR that describe how ‘sophisti cated’ the experiences are as perceived by the user.
|
||
Moreover, managers often want to solve business problems through XR. While we clearly acknowledge that XR cannot solve every problem, our framework further suggests that the distinction between AR and VR, and its sub forms in particular, is important. For instance, if a company wants to guide its production workers through a specific task, AR – in particular assisted reality – is most likely beneficial. On the other hand, mixed reality might be best for letting a customer aesthetically experi ence a product in their living space. However, if the firm seeks to provide customers with an understanding of their production environment, VR might be the best choice. Nevertheless, in all cases, the availability of
|
||
|
||
12
|
||
|
||
P.A. Rauschnabel et al.
|
||
devices to the target consumer (e.g., mixed reality devices) is crucial. For AR, we argue that user experiences can range from a very low
|
||
functional level (assisted reality) to highly interactive and realistic ex periences (mixed reality). We propose the degree of local presence of the experience as the primary distinction on this continuum. Low local presence indicates the assisted reality pole (e.g., simple text overlaid over real world data) and high levels of local presence indicate mixed reality. In true mixed reality, local presence would be so high that users may not be able to distinguish virtual from real content; they would actually experience it as being in their physical environment. However, we acknowledge that higher levels of local presence might be preferable in many, but not in all cases. There might indeed be situations where lower levels of local presence are preferred, such as when providing instructions to an employee (e.g., a worker should clearly be able to distinguish a real cable with high voltage from a virtual one).
|
||
In VR, higher levels of telepresence indicate a higher feeling of “being there” in the simulated environment. The xReality framework’s VR continuum indicates that user experiences can be described between the end poles of atomistic and holistic VR. Atomistic experiences are typically simply designed, have low levels of interaction, and generally have a more “pragmatic” purpose. Holistic VR experiences are charac terized by multi-sensory, complex, social experiences. As in AR, we argue that higher levels of telepresence might be better in many, yet not all, cases. For instance, a higher level of telepresence can lead to a flow experience of “time flying by” and longer use - which might not always be desirable (e.g., an instrumental task such as finding a specific piece of information). Likewise, our informants mentioned various practical ar guments as to why lower telepresence VR experiences are acceptable and potentially preferable in certain situations. For instance, although most experts expect rapid improvements in technology, they suggest that simpler devices (which might not enable very high levels of tele presence) will remain more practicable (e.g., not require external computing power or external tracking technology, be lower priced, more flexible in use, or even owned by a large number of people) than highly sophisticated ones.
|
||
6.2. AR and VR as technologically-mediated experiences
|
||
A substantial amount of research conceptualizes xReality through the lens of technology. For instance, Milgram and Kishino (1994) argue that in AR, “real world and virtual world objects are presented together within a single display”, and Azuma et al. (2001) define AR as ‘‘[a technology which] supplements the real world with virtual (computer generated) objects that appear to coexist in the same space as the real world”. Azuma et al. (2001) go on to add three characteristics: First, AR combines real and virtual objects into a real environment and runs interactively; second, AR is in real time, and third, it registers real and virtual objects with each other (p.34). The current research comple ments Azuma et al. by conceptualizing XR from a user experience perspective which requires certain technology.
|
||
6.2.1. Defining AR in the xReality framework Augmented Reality is a hybrid experience consisting of context-
|
||
specific virtual content that is merged into a user’s real-time percep tion of the physical environment through computing devices. AR can further be refined based on the level of local presence, ranging from assisted reality (low) to mixed reality (high).
|
||
Several elements of this definition require discussion in more detail. First, determining at what point forms of merged virtual and real objects become AR may remain at times ambiguous. For instance, the first versions of Snapchat Glasses are often discussed as AR, but according to our definition, they would not be considered AR in a strict sense. These glasses capture the real world through a camera, and subsequently (not in real-time) add virtual elements on a user’s smartphone or tablet. However, if this was in real-time (e.g., as in the Snap Spectacles AR Developer Edition), it would be AR.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Second, AR content must be associated with a user’s physical, realworld context, which might also include the user herself (e.g., through a mirror-like make-up app on a tablet). Azuma (1997) and other scholars argue that augmented content needs to be physically registered to the real world. In other words, from Azuma’s perspective, the virtual con tent must be attached to a specific location or object (“world-stable”), which means that head-stable content (a common use case for AR glasses) would not be AR. We argue that some form of context relevance is required (e.g., the processual context, such as receiving information related to certain work instructions), but world-stable context is not a specific requirement for AR.
|
||
Third, the term ‘hybrid’ echoes the general consensus that AR ex periences must consist of digital and real-content, and both need to coexist during the experience. In most cases, the core digital content is visual, but our definition does not exclude user experiences with, for instance, acoustic digital content only.
|
||
Fourth, AR is not limited to a specific hardware, but each technology has some specific characteristics that determine whether an experience is AR or not:
|
||
• In video see-through AR (including AR mirrors where users see themselves, such as in a Makeup trial app), the AR experience hap pens on an intransparent screen. Hence, a regular digital picture frame or screen in a living room (without merging real and virtual content in real-time) might influence how a person experiences the room in general, but this would not be considered AR. In contrast, a virtual TV screen (cf. Rauschnabel et al., 2020) which users experi ence through AR smart glasses with optical see-through technology only would be considered AR. The same would be true for video see-through glasses or AR in live TV programs; the fact that real-world objects are digitized is inconsequential, since they are still displayed as real (as in AR mirrors).
|
||
• In optical see-through AR, virtual elements presented on the screen must be part of the experience. That is, a (digital) street sign would not be considered AR, since the virtual content is not integrated but rather a part of the physical object (i.e., the sign).
|
||
• In projection-based AR, the projected content must be digitally controlled and be registered to the physical world. A decorative color light would not be considered AR.
|
||
• Although most AR discourses typically center around visual content (e.g., Gatter et al., 2022), similar principles may apply in non-visual AR. There, the content must also merge with a specific context in real-time (e.g., a specific person’s actions). For instance, a user wearing Amazon’s Alexa spectacles (screen-less glasses, basically an Amazon Echo in the spectacles’ frame) that react to auditory com mands or environmental triggers is considered AR, whereas a PA announcement in an airport not to leave the luggage unattended would not be considered auditory AR.
|
||
Finally, we acknowledge diminished reality as a specific sub form of AR that can occur anywhere on the AR continuum. In assisted reality, objects might be blurred out or hidden by a censor bar, whereas they are realistically erased in mixed reality. Diminished reality also works for audio (e.g., active noise reduction in headphones), but diminishing other sensual stimuli (smell or taste) might be challenging.
|
||
6.2.2. Defining VR in the xReality framework Virtual Reality is an artificial, virtual, and viewer-centered experi
|
||
ence in which the user is enclosed in an all-encompassing 3D space that is - at least visually - sealed off from the physical environment. VR ex periences can lie on a telepresence continuum ranging from atomistic (low) to holistic (high).
|
||
As with AR, several elements of this VR definition require discussion in more detail. First, the term viewer-centered implies that the content is typically built around a certain user (we acknowledge second-person VR as a specific sub form which we do not discuss further). Second, visually
|
||
|
||
13
|
||
|
||
P.A. Rauschnabel et al.
|
||
sealed off implies that a user does not see the physical environment. We reduce this notion to visual senses since temperature, loud noises, or haptics (e.g., wind, surface of the floor etc.) are not very well omittable or controllable. Furthermore, looking at 3D content on a 2D screen and browsing through it with a mouse is excluded from our conceptualiza tion of VR. Third, users see only virtual content (and not content from the physical environment) which can range from static 360-degree views to high-end multi-sensual and fully immersive VR experiences. It is important to acknowledge that we did not receive full agreement on the minimum requirements for the lowest possible level of VR. For instance, some experts suggested that VR requires the possibility for users to manipulate content (in contrast to purely consumable content, e.g., a 360-degree video displayed on an HMD), whereas others argued that looking around and switching apps qualifies as a minimum level of interactivity. Fourth, most discussions in this article revolve around common HMDs. Although our industry experts observed a declining relevance of VR caves, the general propositions of our framework might still apply to them. Fifth, we conclude that in most cases, higher levels of telepresence lead to a better user experience and, thus, are desired. This aligns with the views of our informants that VR experiences situated toward the atomistic endpoint are in many cases perceived to be more “practical” (e.g., mobile, easy-to-use, etc.), whereas user experiences toward the holistic endpoint often require stationary infrastructure and complex tracking technology.
|
||
7. Limitations and future research
|
||
This study has several limitations. Our research represents a “snap shot” of terminologies for AR and VR scenarios and devices as they are currently being used. These terminologies and uses will evolve with technological advances and user experience. Our framework and how we look at XR provides a complementary view to much of the extant work on the XR landscape without claiming that the proposed frame work is a one-size-fits-all solution. More specifically, we acknowledge that the purpose of this framework is to organize common XR use cases.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
We discussed several uncommon cases with our panel of experts, such as a person sitting on a real chair in a VR cave while using an AR device inside of it, which may be unlikely to ever take place. One could argue that these examples represent exceptions where a user experiences ele ments of both AR and VR at the same time, or where one could argue that the “AR content” is part of a VR experience. However, we clearly acknowledge that these examples require significant justification to be implemented in other frameworks, if this is even possible. With this in mind, we argue that the purpose of the xReality framework is to classify and organize current use cases and devices – which implies a call to reassess the framework later. Moreover, we identified various factors throughout our research that determine an experience’s position on the AR or VR continuum, respectively. We acknowledge that this list of factors might not be complete and thus opportunities for further research arise. Hence, our work remains largely conceptual, and future research should be conducted to validate and extend our findings.
|
||
Table 3 presents more specific suggestions for future academic research that emerged from the current research findings. Here, we distinguish between user-focused (i.e., how users react to XR) and management-focused (i.e., how practitioners should apply XR) research. On a meta level, we provide several broader implications on how aca demics can shape the future and impact of XR, including ethics and privacy (e.g., Lammerding et al., 2021; Cowan et al., 2021, Finnegan et al., 2021; Rauschnabel et al., 2022). However, this distinction may be perceived as a false dichotomy since the rapidly evolving nature of XR will blur the lines between what is currently considered “academic” or “practitioner” research. One purpose of the current research is to define XR in a way that transcends this distinction. Much like the contention that approaches from diverse disciplines will be beneficial in under standing XR, merging the perspectives of academics and practitioners will speed our understanding of this unique arena. We hope that the current research contributes here.
|
||
During the review process for this paper, Facebook, Inc. changed its name to Meta Platforms, Inc. in an effort to shift the brand’s focus from social networking to the creation of an AR/VR driven metaverse. As the
|
||
|
||
Table 3 Potential future research on XR.
|
||
|
||
User-Focused Research
|
||
|
||
Management-Focused Research
|
||
|
||
XR as an umbrella term (referencing P1)
|
||
Separation of AR from VR (referencing P2)
|
||
The scope of AR and VR, (referencing P3, P4)
|
||
On a Meta Level
|
||
|
||
Study users’ understanding of XR terminology; can help scholars identify and explore the core mechanisms driving complex consumer responses to these technologies, including their dark sides (e.g., privacy). Research could identify the differences in/influences on user behavior (e.g., success factors for pleasant user experiences, risk factors, performance and usability) when interacting through either AR or VR. The findings of the current research (e.g., Table 1) can inform a research agenda. What additional properties/features determine an experience’s position on the AR or VR continuum? How can these criteria be measured? What are the psychological constructs (e.g., user goals, information processing styles) that mediate the characteristics? How do these factors impact users?
|
||
|
||
Managing XR requires the incorporation of user perception of the technology. Define typical use cases and how they can benefit from XR. Future research could use case studies as a starting point to identify dynamic capabilities specific to XR. A framework that theorizes the feasibility of using AR or VR in different contexts. Furthermore, academic research could develop approaches to the measurement of effectiveness and efficiency of AR and VR applications (e.g., identification of KPIs like engagement, intention for repeated use, etc.). For which use cases is it necessary to focus specifically on sophisticated mixed reality or holistic VR? When is assisted reality or atomistic VR sufficient or better suited to accomplish organizational goals? How should companies react to differences in users’ perceptions? How can these differences be monitored to facilitate effective responses?
|
||
|
||
Research principles, methodologies, data, and research practices
|
||
Disciplines Ethics
|
||
|
||
Future academic research should address practical problems, such as defining the scope of an evolving industry. As is common in the humancomputer interaction field, the use of design science research could be beneficial in other disciplines when exploring these topics. Furthermore, XR offers new forms of data that can help understand users from behavioral, physiological, emotional, and attitudinal perspectives (e.g., through tracking via embedded sensors such as eye tracking and motion sensing; embedded surveys; etc.). Scholars and practitioners might also make use of the data about the surrounding physical environment that is gathered through sensors. The above also points to new ethical challenges, such as how scholars can protect respondents physically (e.g., from collisions or motion sickness in VR) and psychologically (e.g., from traumatizing content) as well as from a security (e.g., impostors manipulating the visual output or capturing the input) and privacy (e.g., by collecting user data without prior consent) perspective. Finally, exploring how knowledge can be transferred effectively between industry, academia, and public policy should be an area of future research. The current research serves as a call for more interdisciplinary research on XR, AR and VR. XR has suffered from inconsistent definition and industrial application, and few strategies have been developed for its effective implementation. Furthermore, innovative ideas from one discipline might lead to legal problems related to another. Interdisciplinary teams of researchers could tackle these challenges better than scholars from a single discipline. Very little is known about the dark side of XR. Future research should investigate how the excessive or repetitive use of XR impacts the physical and psychological wellbeing of individuals and societies at large.
|
||
|
||
14
|
||
|
||
P.A. Rauschnabel et al.
|
||
next iteration of social networking, Meta seeks to leverage the tech nologies to simulate social connection, fulfilling a purpose consistent with the Facebook legacy. Future research might apply the xReality framework to the metaverse concept as the differences between XR and VR will become increasingly important.
|
||
8. General conclusion
|
||
xReality is a dynamic and rapidly developing field that is influenced by technology companies and their marketing departments who aim to differentiate themselves from existing solutions. Consultancies, aca demics, bloggers, and journalists are shaping the discipline by commu nicating their own understanding of terms. The current research provides an attempt to structure and organize terminologies through the lenses of industry, academia, and the user. We hope that our framework supports the discipline.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Author statement
|
||
All authors contributed equally to the manuscript. Philipp Rausch nabel initiated the project.
|
||
Acknowledgements
|
||
The authors wish to acknowledge the support of all experts and discussion partners. The authors furthermore thank Charlie Fink, Martin Meiβner, Carlos Flavi´an, Petra Dahm, Katrin Schein as well as the par ticipants of the Artificial Intelligence and Robotics in Service In teractions Conference (airsi) for their valuable feedback on previous versions of this manuscript. Furthermore, we thank the students and managers enrolled in the XR marketing seminars at Universita¨t der Bundeswehr München, Otto-Friedrich Universita¨t Bamberg and MCI Innsbruck. Finally, we acknowledge the constructive feedback and guidance of the guest editors and two anonymous Computers in Human Behavior reviewers.
|
||
|
||
Appendix 1. Overview of the AR/VR/XR Experts
|
||
|
||
No Study Name
|
||
|
||
Gender, Age, years of professional XR experience
|
||
|
||
Background
|
||
|
||
Category
|
||
|
||
Focus
|
||
|
||
1 FG 2 FG 3 FG 4 FG 5 FG 6 FG 7 FG 8 QI 9 QI 10 QI 11 QI 12 QI 13 QI
|
||
|
||
JOE MAX MEL RICK CARL BEN LUKE SVEN ANNA TIM PAT DORIAN LENA
|
||
|
||
m/thirties/~5 years m/forties/~8 years f/thirties/~5 years m/forties/~15 years m/forties/~6 years m/forties/~5 years m/thirties/~5years m/fifties/~5years f/fifties/~7 years m/thirties/~5 years m/fifties/~6 years m/thirties/~8 years f/thirties/~5 years
|
||
|
||
14 QI 15 QI
|
||
16 QI 17 QI
|
||
18 QI
|
||
19 QI
|
||
20 QI
|
||
21 QI
|
||
22 QI
|
||
|
||
JAKE MIKE
|
||
|
||
m/thirties/~12 years m/thirties/~6 years
|
||
|
||
SAM ELON
|
||
|
||
m/eighties/>20 years m/thirties/~12 years
|
||
|
||
CARLA
|
||
|
||
f/thirties/~8 years
|
||
|
||
BILL
|
||
|
||
m/forties/~20 years
|
||
|
||
MARTIN m/sixties/~30 years
|
||
|
||
TRISTAN m/thirties/~8 years
|
||
|
||
GARRY
|
||
|
||
m/twenties/~4 years
|
||
|
||
Has a background in design science (PhD) and international experience as a VR expert. Researcher and
|
||
|
||
VR
|
||
|
||
Now working as a consultant and XR educator
|
||
|
||
industry expert
|
||
|
||
CEO of a company specialized on the development of mobile AR applications in
|
||
|
||
Developer
|
||
|
||
AR
|
||
|
||
marketing, sales and event management (B2C context)
|
||
|
||
XR specialist in a leading IT company where she is involved in implementing XR in Industry XR
|
||
|
||
XR
|
||
|
||
different fields (B2B); PhD in business
|
||
|
||
Manager
|
||
|
||
Developer and entrepreneur in XR, now working as an AR specialist for a leading
|
||
|
||
Developer and
|
||
|
||
AR
|
||
|
||
technology company; has a PhD in computer science
|
||
|
||
industry expert
|
||
|
||
Head of an XR association and owner of a digital content company specialized in 3D Developer and
|
||
|
||
VR
|
||
|
||
and AR/VR content
|
||
|
||
industry expert
|
||
|
||
Journalism Professor (PhD) who has been involved in numerous non-profit XR
|
||
|
||
Researcher and
|
||
|
||
AR
|
||
|
||
projects, worked in media, sociology degree
|
||
|
||
developer
|
||
|
||
Creative Director in an AR content agency specialized on B2C content, more than five Developer
|
||
|
||
AR
|
||
|
||
years of professional experience in XR
|
||
|
||
Director of an industry association specialized in enterprise AR; former consultant
|
||
|
||
Industry expert
|
||
|
||
AR
|
||
|
||
Head of a ministerial association that promotes XR businesses, law degree
|
||
|
||
Industry expert
|
||
|
||
XR
|
||
|
||
Head of XR in a recreation arena with a VR area, responsible for a variety of XR
|
||
|
||
Industry expert
|
||
|
||
VR
|
||
|
||
attractions (with customer contact)
|
||
|
||
AR consultant specialized on the implementation of AR in enterprises, multiple years Consultant
|
||
|
||
AR
|
||
|
||
in various XR hardware and software companies
|
||
|
||
XR Expert, academic researcher (PhD) and educator, involved in mobile AR, smart
|
||
|
||
Researcher
|
||
|
||
XR
|
||
|
||
glasses projects (academic and with the industry)
|
||
|
||
Academic researcher (PhD) in a university of USA, specialized in the applications of Researcher
|
||
|
||
XR
|
||
|
||
AR and VR in tourism and history communication. She is also a former marketing
|
||
|
||
manager
|
||
|
||
Assoc. Professor in a research university, specialized in VR; background in robotics
|
||
|
||
Researcher
|
||
|
||
VR
|
||
|
||
National responsibility for AR and VR in a leading consulting firm (Big Four); book Consultant
|
||
|
||
XR
|
||
|
||
author and speaker on AR and VR topics
|
||
|
||
Head of a commercial AR research company and author of an impactful book on AR Consultant
|
||
|
||
AR
|
||
|
||
Theater background, experience in projection mapping and various AR and VR
|
||
|
||
Developer
|
||
|
||
XR
|
||
|
||
projects, founder and CEO of an XR company
|
||
|
||
Academic and applied research on XR in different fields (tourism, marketing, culture Researcher
|
||
|
||
XR
|
||
|
||
etc.)
|
||
|
||
XR consultant and team lead in a fortune 100 IT/telecommunications company,
|
||
|
||
Consultant
|
||
|
||
XR
|
||
|
||
consults and develops XR solutions for clients
|
||
|
||
Researcher/Professor at a US R1 university, in a leading position for a HCI research Researcher
|
||
|
||
XR
|
||
|
||
center, author of books on AR and VR.
|
||
|
||
Engineer in a company specialized on training in VR, has a PhD in computer science Engineer
|
||
|
||
VR
|
||
|
||
with a focus on XR.
|
||
|
||
Active member in an XR association, project managers in a VR education company and Manager
|
||
|
||
VR
|
||
|
||
entrepreneur.
|
||
|
||
Note: To protect respondents’ identity, names represent pseudonyms and the age is categorized in decades. Most respondents are natively from European countries, but live(d) and work(ed) on other continents; FG = Focus groups; QI = Qualitative Interview.
|
||
|
||
15
|
||
|
||
P.A. Rauschnabel et al.
|
||
References
|
||
Alcan˜iz, M., Bign´e, E., & Guixeres, J. (2019). Virtual reality in marketing: A framework, review, and research agenda. Frontiers in Psychology, 10(1530), 1–15.
|
||
Argelaguet, F., Hoyet, L., Trico, M., & L´ecuyer, A. (2016). The role of interaction in virtual embodiment: Effects of the virtual hand representation. In 2016 IEEE virtual reality (VR) (pp. 3–10).
|
||
Azuma, R. T. (1997). A survey of augmented reality. Presence, 6(4), 355–385. Azuma, R., Baillot, Y., Behringer, R., Feiner, S., Julier, S., & MacIntyre, B. (2001). Recent
|
||
advances in augmented reality. IEEE Computer Graphics and Applications, 21(6), 34–47. Bachras, V., Raptis, G. E., & Avouris, N. M. (2019). On the use of persistent spatial points for deploying path navigation in augmented reality: An evaluation study. In D. Lamas, F. Loizides, L. Nacke, et al. (Eds.), Human-computer interaction –interact 2019 (pp. 308–319). Battarbee, K., & Koskinen, I. (2008). Co-experience: Product experience as social interaction. In H. N. J. Schifferstein, & P. Hekkert (Eds.), Product experience (pp. 461–476). Amsterdam, The Netherlands: Elsevier. Baum, F. L. (1901). The master key: An electrical fairy tale. Indianapolis. Bowen-Merrill Company Publishers. Baus, O., & Bouchard, S. (2017). Exposure to an unpleasant odour increases the sense of presence in virtual reality. Virtual Reality, 21(2), 59–74. Beck, J., Rainoldi, M., & Egger, R. (2019). Virtual reality in tourism: A state-of-the-art review. Tourism Review, 74(3), 586–612. Berg, L. P., & Vance, J. M. (2017). Industry use of virtual reality in product design and manufacturing: A survey. Virtual Reality, 21(1), 1–17. Berryman, D. R. (2012). Augmented reality: A review. Medical Reference Services Quarterly, 31(2), 212–218. Bierbaum, A., Just, C., Hartling, P., Meinert, K., Baker, A., & Cruz-Neira, C. (2001). VR juggler: A virtual platform for virtual reality application development. In Proceedings IEEE virtual reality 2001 (pp. 89–96). Billinghurst, M., Clark, A., & Lee, G. (2015). A survey of augmented reality. Foundations and Trends in Human-Computer Interaction, 8(2–3), 73–272. Bosco, F. J., & Herman, T. (2010). Focus groups as collaborative research performances. In D. Delyser, S. Herbert, S. Aitken, M. Crang, & L. McDowell (Eds.), The SAGE Handbook of qualitative Geography (pp. 193–207). Los Angeles, CA: SAGE. Bourdon, J. (2020). From correspondence to computers: A theory of mediated presence in history. Communication Theory, 30(1), 64–83. Bowman, D. A., & McMahan, R. P. (2007). Virtual reality: How much immersion is enough? Computer, 40(7), 36–43. Bown, J., White, E., & Boopalan, A. (2017). Looking for the ultimate display: A brief history of virtual reality. In J. Gackenbach, & J. Bown (Eds.), Boundaries of self and reality online: Implications of digitally constructed realities (pp. 239–259). London, United Kingdom: Elsevier. Brooks, R. A. (1990). Elephants don’t play chess. Robotics and Autonomous Systems, 6 (1–2), 3–15. Carmigniani, J., Furht, B., Anisetti, M., Ceravolo, P., Damiani, E., & Ivkovic, M. (2011). Augmented reality technologies, systems and applications. Multimedia Tools and Applications, 51(1), 341–377. Carrozzi, A., Chylinski, M., Heller, J., Hilken, T., Keeling, D. I., & de Ruyter, K. (2019). What’s mine is a hologram? How shared augmented reality augments psychological ownership. Journal of Interactive Marketing, 48, 71–88. Chang, E., Kim, H. T., & Yoo, B. (2020). Virtual reality sickness: A review of causes and measurements. International Journal of Human-Computer Interaction, 36(17), 1658–1682. Chylinski, M., Heller, J., Hilken, T., Keeling, D. I., Mahr, D., & de Ruyter, K. (2020). Augmented reality marketing: A technology-enabled approach to situated customer experience. Australasian Marketing Journal, 28(4), 374–384. Ço¨ltekin, A., Lochhead, I., Madden, M., Christophe, S., Devaux, A., Pettit, C., Lock, O., Shukla, S., Herman, L., Stachonˇ, Z., Kubíˇcek, P., Snopkov´a, D., Bernardes, S., & Hedley, N. (2020). Extended reality in spatial sciences: A review of research challenges and future directions. International Journal of Geo-Information, 9(7), 1–29. Coon, J. (2018). Assisted reality: What it is, and how it will improve service productivity. PTC. https://www.ptc.com/en/blogs/service/what-is-assisted-reality. Cowan, K., Javornik, A., & Jiang, P. (2021). Privacy concerns when using augmented reality face filters? Explaining why and when use avoidance occurs. Psychology and Marketing, 38(10), 1799–1813. Cowan, K., & Ketron, S. (2019). A dual model of product involvement for effective virtual reality: The roles of imagination, co-creation, telepresence, and interactivity. Journal of Business Research, 100, 483–492. Daassi, M., & Debbabi, S. (2021). Intention to reuse AR-based apps: The combined role of the sense of immersion, product presence and perceived realism. Information & Management, 58(4), 103453. Dalton, J. (2021). Reality check: How immersive technologies can transform your business. London, UK: Kogan Page. Deleuze, G. (1966). Le Bergsonisme. Originally published in French and translated into English by Hugh Tomlinson and Barbara Habberjam (1988). Brooklyn: Zone Books. Deloitte. (2018). Immersive technologies of the new digital reality. Delloitte Insights. https ://www2.deloitte.com/us/en/insights/focus/tech-trends/2018/immersive-tech nologies-digital-reality.html. Deng, L., Turner, D. E., Gehling, R., & Prince, B. (2010). User experience, satisfaction, and continual usage of IT. European Journal of Information Systems, 19, 60–75. Dwivedi, Y. K., Ismagilova, E., Hughes, D. L., Carlson, J., Filieri, R., Jacobson, J., & Wang, Y. (2020). Setting the future of digital and social media marketing research: Perspectives and research propositions. International Journal of Information Management, 59, 102168.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Englmeier, D., O’Hagan, J., Zhang, M., Alt, F., Butz, A., Ho¨llerer, T., & Williamson, J. (2020). Tangible sphere–interaction techniques for physical and virtual spherical displays. In Proceedings of the 11th Nordic conference on human-computer interaction: Shaping experiences (pp. 1–11). Shaping Society.
|
||
Fanger, Y., Pfeuffer, K., Helmbrecht, U., & Alt, F. (2020). PIANX–A platform for piano players to alleviate music performance anxiety using mixed reality. In 19th International conference on mobile and ubiquitous Multimedia (pp. 267–276).
|
||
Farshid, M., Paschen, J., Eriksson, T., & Kietzmann, J. (2018). Go boldly! Explore augmented reality (AR), virtual reality (VR), and mixed reality (MR) for business. Business Horizons, 61, 657–663.
|
||
Feiner, S., MacIntyre, B., Haupt, M., & Solomon, E. (1993). Windows on the world: 2D windows for 3D augmented reality. In Proceedings of the 6th annual ACM symposium on User interface software and technology (pp. 145–155).
|
||
Fertleman, C., Aubugeau-Williams, P., Sher, C., Lim, A. N., Lumley, S., Delacroix, S., & Pan, X. (2018). A discussion of virtual reality as a new tool for training healthcare professionals. Frontiers in Public Health, 6(44), 1–5.
|
||
Finnegan, D. J., Zoumpoulaki, A., & Eslambolchilar, P. (2021). Does mixed reality have a cassandra complex? Frontiers in Virtual Reality, 2, 56.
|
||
Flavia´n, C., Ib´an˜ez-Sa´nchez, S., & Orús, C. (2019). The impact of virtual, augmented and mixed reality technologies on the customer experience. Journal of Business Research, 100, 547–560.
|
||
Flavia´n, C., Iba´n˜ez-S´anchez, S., & Orús, C. (2021). Impacts of technological embodiment through virtual reality on potential guests’ emotions and engagement. Journal of Hospitality Marketing & Management, 30(1), 1–20.
|
||
Gatter, S., Hüttl-Maack, V., & Rauschnabel, P. A. (2022). Can augmented reality satisfy consumers’ need for touch? Psychology & Marketing, 39(3), 508–523.
|
||
Greene, C. M., Broughan, J., Hanlon, A., Keane, S., Hanrahan, S., Kerr, S., & Rooney, B. (2021). Visual search in 3D: Effects of monoscopic and stereoscopic cues to depth on the validity of feature integration theory and perceptual load theory. Frontiers in Psychology, 12, 748.
|
||
Guttentag, D. A. (2010). Virtual reality: Applications and implications for tourism. Tourism Management, 31(5), 637–651.
|
||
Han, D. I., Jung, T., & Gibson, A. (2013). Dublin AR: Implementing augmented reality in tourism. In Information and communication technologies in tourism 2014 (pp. 511–523).
|
||
Hassenzahl, M. (2018). A personal journey through user experience. Journal of Usability Studies, 13(4), 168–176.
|
||
Hassenzahl, M., & Tractinsky, N. (2006). User experience - a research agenda. Behaviour & Information Technology, 25(2), 91–97.
|
||
Heinrich, F., Bornemann, K., Lawonn, K., & Hansen, C. (2019). Depth perception in projective augmented reality: An evaluation of advanced visualization techniques. In Proceedings of the 25th ACM symposium on virtual reality software and technology (pp. 1–11).
|
||
Heller, J., Chylinski, M., de Ruyter, K., Keeling, D. I., Hilken, T., & Mahr, D. (2021). Tangible service automation: Decomposing the technology-enabled engagement process (TEEP) for augmented reality. Journal of Service Research, 24(1), 84–103.
|
||
Heller, J., Chylinski, M., de Ruyter, K., Mahr, D., & Keeling, D. I. (2019a). Let me imagine that for you: Transforming the retail frontline through augmenting customer mental imagery ability. Journal of Retailing, 95(2), 94–114.
|
||
Heller, J., Chylinski, M., de Ruyter, K., Mahr, D., & Keeling, D. I. (2019b). Touching the untouchable: Exploring multi-sensory augmented reality in the context of online retailing. Journal of Retailing, 95(4), 219–234.
|
||
Hilken, T., de Ruyter, K., Chylinski, M., Mahr, D., & Keeling, D. I. (2017). Augmenting the eye of the beholder: Exploring the strategic potential of augmented reality to enhance online service experiences. Journal of the Academy of Marketing Science, 45, 884–905.
|
||
Hilken, T., Keeling, D. I., de Ruyter, K., Mahr, D., & Chylinski, M. (2020). Seeing eye to eye: Social augmented reality and shared decision making in the marketplace. Journal of the Academy of Marketing Science, 48, 143–164.
|
||
Hinsch, C., & Bloch, P. H. (2009). Interaction seeking in Second Life and implications for consumer behavior. In N. T. Wood, & M. R. Solomon (Eds.), Virtual social identity and consumer behavior (pp. 43–60). New York, NY: Routledge.
|
||
Hoffman, H. G., Meyer, W. J., III, Ramirez, M., Roberts, L., Seibel, E. J., Atzori, B., Sharae, S. R., & Patterson, D. R. (2014). Feasibility of articulated arm mounted Oculus Rift Virtual Reality goggles for adjunctive pain control during occupational therapy in pediatric burn patients. Cyberpsychology, Behavior, and Social Networking, 17(6), 397–401.
|
||
Hoffman, D. L., Moreau, C. P., Stremersch, S., & Wedel, M. (2022). The rise of new technologies in marketing: A framework and outlook. Journal of Marketing, 86(1), 1–6.
|
||
Holl¨ander, K., Colley, A., Mai, C., Ha¨kkila¨, J., Alt, F., & Pfleging, B. (2019). Investigating the influence of external car displays on pedestrians’ crossing behavior in virtual reality. In Proceedings of the 21st intl. conference on human-computer interaction with mobile devices and services (MobileHCI ’19). ACM.
|
||
Hollebeek, L. D., Clark, M. K., Andreassen, T. W., Sigurdsson, V., & Smith, D. (2020). Virtual reality through the customer journey: Framework and propositions. Journal of Retailing and Consumer Services, 55, 102056.
|
||
Hoppe, A. H., Westerkamp, K., Maier, S., van de Camp, F., & Stiefelhagen, R. (2018). Multi-user collaboration on complex data in virtual and augmented reality. In HCI 2018: Communications in computer and information science (Vol. 851).
|
||
Hornecker, E. (2010). Interactions around a contextually embedded system. In Proceedings of the Fourth International conference on Tangible, embedded, and embodied interaction (pp. 169–176).
|
||
Hoyer, W. D., Kroschke, M., Schmitt, B., Kraume, K., & Shankar, V. (2020). Transforming the customer experience through new technologies. Journal of Interactive Marketing, 51, 57–71.
|
||
|
||
16
|
||
|
||
P.A. Rauschnabel et al.
|
||
Hudson, S., Matson-Barkat, S., Pallamin, N., & Jegou, G. (2019). With or without you? Interaction and immersion in a virtual reality experience. Journal of Business Research, 100, 459–468.
|
||
Itti, L., Dhavale, N., & Pighin, F. (2003). Realistic avatar eye and head animation using a neurobiological model of visual attention. In SPIE proceedings Vol. 5200: Applications and science of Neural Networks, Fuzzy systems, and evolutionary Computation VI.
|
||
Jaekl, P. M., Allison, R. S., Harris, L. R., Jasiobedzka, U. T., Jenkin, H. L., Jenkin, M. R., & Zikovitz, D. C. (2002). Perceptual stability during head movement in virtual reality. In Proceedings IEEE virtual reality 2002 (pp. 149–155).
|
||
Jeon, C. (2015). The virtual flier: The link trainer, flight simulation, and pilot identity. Technology and Culture, 56(1), 28–53.
|
||
Keil, J., Edler, D., & Dickmann, F. (2019). Preparing the HoloLens for user studies: An augmented reality interface for the spatial adjustment of holographic objects in 3D indoor environments. KN-Journal of Cartography and Geographic Information, 69(3), 205–215.
|
||
Kim, D., & Ko, Y. J. (2019). The impact of virtual reality (VR) technology on sport spectators’ flow experience and satisfaction. Computers in Human Behavior, 93, 346–356.
|
||
Knierim, P., Schwind, V., Feit, A. M., Nieuwenhuizen, F., & Henze, N. (2018). Physical keyboards in virtual reality: Analysis of typing performance and effects of avatar hands. In Proceedings of the 2018 CHI conference on human factors in computing systems (pp. 1–9). ACM.
|
||
Ko¨se, A., Tepljakov, A., & Petlenkov, E. (2020). Real time data communication for intelligent extended reality applications. In 2020 IEEE International conference on Computational intelligence and virtual environments for Measurement systems and applications (CIVEMSA) (pp. 1–6).
|
||
Krasonikolakis, I., Vrechopoulos, A., Pouloudi, A., & Dimitriadis, S. (2018). Store layout effects on consumer behavior in 3D online stores. European Journal of Marketing, 52 (5/6), 1223–1256.
|
||
Kunkel, N., & Soechtig, S. (2017). Mixed reality: Experiences get more intuitive, immersive and empowering. Deloitte University Press.
|
||
Lammerding, L., Hilken, T., Mahr, D., & Heller, J. (2021). Too real for comfort: Measuring consumers’ augmented reality information privacy concerns. In M. C. tom Dieck, T. H. Jung, & S. M. C. Loureiro (Eds.), Augmented reality and virtual reality: New trends in immersive technology (pp. 95–108). Cham, Switzerland: Springer.
|
||
Lebeck, K., Ruth, K., Kohno, T., & Roesner, F. (2018). Towards security and privacy for multi-user augmented reality: Foundations with end users. In 2018 IEEE symposium on Security and privacy (SP) (pp. 392–408).
|
||
Leclet-Groux, D., Caron, G., Mouaddib, E., & Anghour, A. (2013). A Serious Game for 3D cultural heritage. In 2013 digital Heritage International congress (DigitalHeritage) (Vol. 1, pp. 409–412).
|
||
Lee, K. M. (2004). Presence, explicated. Communication Theory, 14(1), 27–50. Lee, H., Jung, T. H., Tom Dieck, M. C., & Chung, N. (2020). Experiencing immersive
|
||
virtual reality in museums. Information & Management, 57(5), 103229. Liebers, J., Gruenefeld, U., Mecke, L., Saad, A., Auda, J., Alt, F., Abdelaziz, M., &
|
||
Schneegass, S. (2021). Understanding user identification in virtual reality through behavioral biometrics and the effect of body normalization. In Proceedings of the 2021 CHI conference on human factors in computing. ACM. Lim, J., & Ayyagari, R. (2018). Investigating the determinants of telepresence in the ecommerce setting. Computers in Human Behavior, 85, 360–371. Lin, C. H., & Hsu, P. H. (2017). Integrating procedural modelling process and immersive VR environment for architectural design education. In MATEC Web of Conferences (Vol. 104). Lombard, M., & Ditton, T. (1997). At the heart of it all: The concept of presence. Journal of Computer-Mediated Communication, 3(2). Looser, J., Billinghurst, M., & Cockburn, A. (2004). Through the looking glass: The use of lenses as an interface tool for augmented reality interfaces. In Proceedings of the 2nd intl. Conference on computer graphics and interactive techniques in Australasia and South East Asia (pp. 204–211). Lopes, P., You, S., Ion, A., & Baudisch, P. (2018). Adding force feedback to mixed reality experiences and games using electrical muscle stimulation. In Proceedings of the 2018 CHI conference on human factors in computing systems (pp. 1–13). Lu, Y., & Smith, S. (2009). GPU-based real-time occlusion in an immersive augmented reality environment. Journal of Computing and Information Science in Engineering, 9 (2), 1–4. MacDorman, K. F., Green, R. D., Ho, C.-C., & Koch, C. T. (2009). Too real for comfort? Uncanny responses to computer generated faces. Computers in Human Behavior, 25 (3), 695–710. Mai, C., & Steinbrecher, T. (2018). Evaluation of visual discomfort factors in the context of HMD usage. In Proceedings of IEEE VR 2018. Mann, S., & Fung, J. (2002). EyeTap devices for augmented, deliberately diminished, or otherwise altered visual perception of rigid planar patches of real-world scenes. Presence: Teleoperators and Virtual Environments, 11(2), 158–175. Mantovani, G., & Riva, G. (1999). Real” presence: How different ontologies generate different criteria for presence, telepresence, and virtual presence. Presence, 8(5), 540–550. Milgram, P., & Kishino, A. F. (1994). Taxonomy of mixed reality visual displays. IEICE Transactions on Info and Systems, E77-D(12), 1321–1329. Milgram, P., Takemura, H., Utsumi, A., & Kishino, F. (1995, December). Augmented reality: A class of displays on the reality-virtuality continuum. In Telemanipulator and telepresence technologies (Vol. 2351, pp. 282–292). International Society for Optics and Photonics. Miller, M. R., Jun, H., Herrera, F., Yu Villa, J., Welch, G., & Bailenson, J. N. (2019). Social interaction in augmented reality. PLoS One, 14(5). https://doi.org/10.1371/ journal.pone.0216290
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Mollen, A., & Wilson, H. (2010). Engagement, telepresence and interactivity in online consumer experience: Reconciling scholastic and managerial perspectives. Journal of Business Research, 63(9–10), 919–925.
|
||
Montero, A., Zarraonandia, T., Diaz, P., & Aedo, I. (2019). Designing and implementing interactive and realistic augmented reality experiences. Universal Access in the Information Society, 18, 49–61.
|
||
Mura, M. D., Dini, G., & Failli, F. (2016). An integrated environment based on augmented reality and sensing device for manual assembly workstations. Procedia Cirp, 41, 340–345.
|
||
Mütterlein, J. (2018). The three pillars of virtual reality? Investigating the roles of immersion, presence, and interactivity. In Proceedings of the 51st Hawaii International conference on system sciences (Vol. 9, pp. 1407–1415).
|
||
Nowak, K. (2001). Defining and differentiating copresence, social presence and presence as transportation. In Paper presented at the 4th International Workshop on presence, Philadelphia, PA.
|
||
Oxford English Dictionary. (2021a). Mix. https://www.oxfordify.com/meaning/mix. Oxford English Dictionary. (2021b). Tele-, comb. form. https://www.oed.
|
||
com/view/Entry/198666?rskey=5TzsRx&result=7&isAdvanced=false#eid. Pan, X., & Hamilton, A. F. D. C. (2018). Why and how to use virtual reality to study
|
||
human social interaction: The challenges of exploring a new research landscape. British Journal of Psychology, 109(3), 395–417. Park, M., & Yoo, J. (2020). Effects of perceived interactivity of augmented reality on consumer responses: A mental imagery perspective. Journal of Retailing and Consumer Services, 52, 101912. Petit, O., Velasco, C., & Spence, C. (2019). Digital sensory marketing: Integrating new technologies into multisensory online experience. Journal of Interactive Marketing, 45, 42–61. Pfeuffer, K., Alexander, J., Chong, M. K., & Gellersen, H. (2014). Gaze-touch: Combining gaze with multi-touch for interaction on the same surface. In Proceedings of the 27th annual ACM symposium on User interface software and technology (pp. 509–518). Pfeuffer, K., Geiger, M. J., Prange, S., Mecke, L., Buschek, D., & Alt, F. (2019). Behavioural biometrics in vr: Identifying people from body motion and relations in virtual reality. In Proceedings of the 2019 CHI conference on human factors in computing systems (pp. 1–12). Poushneh, A., & Vasquez-Parraga, A. (2017). Discernible impact of augmented reality on retail customer’s experience, satisfaction and willingness to buy. Journal of Retailing and Consumer Services, 34, 229–234. Raghavan, R., & Rao, P. (2018). Immersive learning for the future workforce. Available at: https://www.accenture.com/_acnmedia/pdf-86/accenture-extended-reality-imm ersive-training.pdf. Rauschnabel, P. A. (2018). Virtually enhancing the real world with holograms: An exploration of expected gratifications of using augmented reality smart glasses. Psychology and Marketing, 35(8), 557–572. Rauschnabel, P. A. (2021). Augmented reality is eating the real-world! the substitution of physical products by holograms. International Journal of Information Management, 57, 102279. Rauschnabel, P. A., Felix, R., & Hinsch, C. (2019). Augmented reality marketing: How mobile AR-apps can improve brands through inspiration. Journal of Retailing and Consumer Services, 49, 43–53. Rauschnabel, P. A., He, J., & Ro, Y. K. (2018). Antecedents to the adoption of augmented reality smart glasses: A closer look at privacy risks. Journal of Business Research, 92, 374–384. Rauschnabel, P. A., Babin, B. J., Tom Dieck, M. C., Krey, N., & Jung, T. (2022). What is augmented reality marketing? Its definition, complexity, and future. Journal of Business Research, 142, 1140–1150. Raymundo, O. (2016). Tim Cook: Augmented reality will be an essential part of your daily life, like the iPhone. Macworld. October 3, 2016 https://www.macworld.com/articl e/3126607/tim-cook-augmented-reality-will-be-an-essential-part-of-your-daily-life-l ike-the-iphone.html. Rodríguez-Ardura, I., & Martínez-Lo´pez, F. J. (2014). Another look at ‘being there’ experiences in digital media: Exploring connections of telepresence with mental imagery. Computers in Human Behavior, 30, 508–518. Ruddle, R. A. (2004). The effect of environment characteristics and user interaction on levels of virtual environment sickness. In Proceedings of IEEE virtual reality. Saunders, B., Sim, J., Kingstone, T., Baker, S., Waterfield, J., Bartlam, B., Burroughs, H., & Jinks, C. (2018). Saturation in qualitative research: Exploring its conceptualization and operationalization. Quality and Quantity, 52(4), 1893–1907. Schein, K. E., & Rauschnabel, P. A. (2021). Augmented reality in manufacturing: Exploring workers’ perceptions of barriers. IEEE Transactions on Engineering Management (in press). Scholz, J., & Duffy, K. (2018). We ARe at home: How augmented reality reshapes mobile marketing and consumer-brand relationships. Journal of Retailing and Consumer Services, 44, 11–23. Schroeder, R. (2002). The social life of avatars: Presence and interaction in shared virtual environments. London, UK: Springer. Shahab, M. H., Ghazali, E., & Mohtar, M. (2021). The role of elaboration likelihood model in consumer behaviour research and its extension to new technologies: A review and future research agenda. International Journal of Consumer Studies, 45(4), 664–689. Sheridan, T. B. (1992). Musings on telepresence and virtual presence. Presence: Teleoperators and Virtual Environments, 1(1), 120–126. Sipatchin, A., Wahl, S., & Rifai, K. (2021). Eye-tracking for clinical ophthalmology with virtual reality (VR): A case study of the HTC vive Pro eye’s usability. In Healthcare, 9 (2). Skarbez, R., Smith, M., & Whitton, M. C. (2021). Revisiting Milgram and Kishino’s reality-virtuality continuum. Frontiers in Virtual Reality, 2(647997), 1–8.
|
||
|
||
17
|
||
|
||
P.A. Rauschnabel et al.
|
||
Slater, M., & Sanchez-Vives, M. V. (2016). Enhancing our lives with immersive virtual reality. Frontiers in Robotics and AI, 3(73), 1–47.
|
||
Smink, A. R., van Reijmersdal, E. A., van Noort, G., & Neijens, P. C. (2020). Shopping in augmented reality: The effects of spatial presence, personalization and intrusiveness on app and brand responses. Journal of Business Research, 118, 474–485.
|
||
Spagnolli, A., Lombard, M., & Gamberini, L. (2009). Mediated presence: Virtual reality, mixed environments and social networks. Virtual Reality, 12, 137–139.
|
||
Speicher, M., Hall, B. D., & Nebeling, M. (2019). What is mixed reality?. In CHI conference on human factors in computing systems proceedings (pp. 1–15). Glasgow, Scotland, UK: ACM.
|
||
Spiggle, S. (1994). Analysis and interpretation of qualitative data in consumer research. Journal of Consumer Research, 21(3), 491–503.
|
||
Stadler, S., Cornet, H., Theoto, T. N., & Frenkler, F. (2019). A tool, not a toy: Using virtual reality to evaluate the communication between autonomous vehicles and pedestrians. In Augmented reality and virtual reality (pp. 203–216). Cham, Switzerland: Springer.
|
||
Steffen, J. H., Gaskin, J. E., Meservy, T. O., Jenkins, J. L., & Wolman, I. (2019). Framework of affordances for virtual reality and augmented reality. Journal of Management Information Systems, 36(3), 683–729.
|
||
Steuer, J. (1992). Defining virtual reality: Dimensions determining telepresence. Journal of Communication, 42(4), 73, 73.
|
||
Szymanski, D. M., & Henard, D. H. (2001). Customer satisfaction: A meta-analysis of the empirical evidence. Journal of the Academy of Marketing Science, 29(1), 16–35.
|
||
Tamura, H., Yamamoto, H., & Katayama, A. (2001). Mixed reality: Future dreams seen at the border between real and virtual worlds. IEEE Computer Graphics and Applications, 21(6), 64–70.
|
||
Tan, Y. C., Chandukala, S. R., & Reddy, S. K. (2021). Augmented reality in retail and its impact on sales. Journal of Marketing, Article 0022242921995449.
|
||
Tussyadiah, I. P., Jung, T. H., & tom Dieck, M. C. (2017). Embodiment of wearable augmented reality technology in tourism experiences. Journal of Travel Research, 57 (5), 597–611.
|
||
Tussyadiah, I. P., Wang, D., Jung, T. H., & tom Dieck, M. C. (2018). Virtual reality, presence, and attitude change: Empirical evidence from tourism. Tourism Management, 66, 140–154.
|
||
|
||
Computers in Human Behavior 133 (2022) 107289
|
||
Verhagen, T., Vonkeman, C., Feldberg, F., & Verhagen, P. (2014). Present it like it is here: Creating local presence to improve online product experiences. Computers in Human Behavior, 39, 270–280.
|
||
Vonkeman, C., Verhagen, T., & Van Dolen, W. (2017). Role of local presence in online impulse buying. Information & Management, 54(8), 1038–1048.
|
||
Wedel, M., Bign´e, E., & Zhang, J. (2020). Virtual and augmented reality: Advancing research in consumer marketing. International Journal of Research in Marketing, 37, 443–465.
|
||
Weibel, D., Schmutz, J., Pahud, O., & Wissmath, B. (2015). Measuring spatial presence: Introducing and validating the pictorial presence SAM. Presence: Teleoperators and Virtual Environments, 24(1), 44–61.
|
||
Wei, W., Qi, R., & Zhang, L. (2019). Effects of virtual reality on theme park visitors’ experience and behaviors: A presence perspective. Tourism Management, 71, 282–293.
|
||
Yaoyuneyong, G. S., Pollitte, W. A., Foster, J. K., & Flynn, L. R. (2018). Virtual dressing room media, buying intention and mediation. The Journal of Research in Indian Medicine, 12(1), 125–144.
|
||
Yim, M. Y.-C., Chu, S.-C., & Sauer, P. L. (2017). Is augmented reality technology an effective tool for e-commerce? An interactivity and vividness perspective. Journal of Interactive Marketing, 39, 89–103.
|
||
Yin, C. Z. Y., Jung, T., Tom Dieck, M. C., & Lee, M. Y. (2021). Mobile augmented reality heritage applications: Meeting the needs of heritage tourists. Sustainability, 13(5), 2523.
|
||
Yung, R., Khoo-Lattimore, C., & Potter, L. E. (2021). VR the world: Experimenting with emotion and presence for tourism marketing. Journal of Hospitality and Tourism Management, 46, 160–171.
|
||
Zhou, F., Duh, H. B. L., & Billinghurst, M. (2008). Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR. In 2008 7th IEEE/ACM International symposium on mixed and augmented reality (pp. 193–202).
|
||
Zielinski, D. J., Rao, H. M., Sommer, M. A., & Kopper, R. (2015). Exploring the effects of image persistence in low frame rate virtual environments. In 2015 IEEE virtual reality (pp. 19–26).
|
||
|
||
18
|
||
|