I. INTRODUCTION
Personal computing has evolved from desktop computers to mobile devices such as smartphones to increase the mobility of computing activities. The next upcoming paradigm shift of personal computing is actively occurring in wearable computers such as smartwatches and smartglass as shown in Fig. 1.
Fig. 1. Wearable computing as the next level of personal computing.
Wearable computing is known for always-on, alwaysaccessible, and always-connected properties of computing that enable more intimate and frequent use regardless of the user’s whereabouts and intention. Commercial products such as Google Glass Enterprise Edition 2, Microsoft HoloLens 2, and Facebook Oculus Quest 2 have proven feasibility and effectiveness in the development of niche application domains such as in the medical and industrial sectors. Moreover, the recent outbreak of the COVID-19 pandemic has accelerated the use of non-contact remote collaboration technology with various configurations of augmented reality (AR), virtual reality (VR), mixed reality (MR), and extended reality (XR). The concept of remote collaboration has been explored in specialized medical and surgical fields [1], [2], [3]. Going beyond targeted and highly specialized niche wearable applications in the healthcare and entertainment domains, we seek to identify the underlying principles that can rapidly develop smart glasses-assisted killer applications for everyday use. We believe that smartglass-assisted interactive remote collaboration is a promising candidate for killer wearable applications when this concept is extended and deployed for more broad everyday applications and casual users.
In this paper, we analyze intermediary roles in remote collaboration and propose technical components for realizing smartglass-assisted interactive remote collaboration [4]. More specifically, we define various intermediary roles of remote collaboration participants and elaborate technical components between those interfaces.
II. RELATED WORK
The proliferation of mobile devices has introduced various frameworks specifically designed for mobile computing environments. For example, cameras on mobile devices are actively used in Personal Companion [5] and camera phone-based interaction techniques [6] to interact with the smart environment. The camera-based frameworks [5], [6] enabled the mobile devices to recognize visual codes or fiducial markers for interaction where a relative distance and angle between camera and visual tags are used as a set of commands.
Research on mobile AR has introduced sensor fusion that naturally connected wearable devices for extended usage and better performance. For these extended use cases, Yoon et al. introduced an augmented smart coach that benefited from using wearable devices for collecting personal big data and generating a personalized quantifiedself representation [7]. Moreover, developments in wearable devices have increased non-academic people to quantify their lives in quantified-self movements [8] aiming for personalized tracking [9]. Choe and Lee defined personalized tracking as, “an act of tailoring a tracking regimen to suit an individual’s specific situations, preferences, and needs” [9].
More recent studies used commercial smartglass such as Google Glass Enterprise Edition 2 for health professional training [10] and Vuzix Blade for nursing education [11]. Ong et al. discussed the potentials for XR telehealth for health care providers during the COVID-19 pandemic [12], while traditional “telementoring” in surgical and medical domains is summarized in survey and review papers [1], [2].
In comparison to related works, we attempt to polish previous remote collaboration scenarios to support three intermediary roles beyond medical and surgical contexts by including smartglass and their characteristics. The proposed intermediary roles and technical components can be used as a reference implementation to explore smartglass-assisted interactive remote collaboration applications.
III. INTERMEDIARY ROLES AND TASKS
In remote collaboration scenarios, we can define intermediary roles based on who takes initiatives with varying controllability and interaction expressivity (CIE). We propose three possible intermediary roles for remote collaboration as authoritative supervisor, synergetic collaborator, and speculative explorer as illustrated in Fig. 2.
Fig. 2. Possible intermediary roles for remote collaboration.
3.1. Authoritative Supervisor
The first intermediary role is authoritative supervisor where one participant has significantly greater CIE than other parties. The party with this authoritative CIE acts as a supervisor to other participants. This relationship is common in telementoring scenarios where mentors have greater knowledge and expertise than mentees [4]. The smartglasses of the supervisor displays visual data from other participants that help the supervisor to give concrete instructions or guidance. Non-supervisory participants basically play two functions. They play the role of passive sensors to deliver details captured with their smartglasses to their supervisor on the remote site. Furthermore, they play the role of the designated actuator to physically carry out a task on behalf of the supervisor. This intermediary role is depicted in the left figure of Fig. 2.
3.2. Synergetic Collaborator
The second intermediary role involves participants with approximately the same level of CIE. The smartglasses on these peers display the same visualization and they can zoom in or out visual data on one’s own device. However, the CIE of one device is on equal terms with the other device. Since collaborators or peers have a similar level of knowledge and expertise, their collaboration can create synergy. This intermediary role is depicted in the center figure of Fig. 2.
3.3. Speculative Explorer
The third intermediary role involves participants carrying out tasks in unknown or unfamiliar domains. In this role, smartglasses are used as an exploration device that recognizes, interprets, and augments additional visualization on the smartglass. A notable application on this role includes social and gaming applications that do not require prior knowledge or experience from the users. In this role, users are explorative in previously unknown or unexperienced tasks. This intermediary role is depicted in the right figure of Fig. 2.
3.4. Tasks for Different Intermediary Roles
The aforementioned three intermediary roles have already been examined in previous remote collaboration systems and applications. We have classified tasks and applications found in remote collaboration studies according to the three intermediary roles in Table 1.
Table 1. Tasks and applications found in various roles.
Surgery and medical tasks [10], [13], [14], [15] in remote collaboration scenarios include intermediary roles of authoritative supervisor (AS) as well as synergetic collaborator (SC). Since surgical and medical tasks require an expertise, speculative explorer (SE) is not common. In education and training tasks [11], [16], [17], all three intermediary roles are addressed. In AS, we have scenarios involving teachers and students. We can have a group of peers or team-based activities with prior knowledge (i.e., SC) or for completely different or new domain (i.e., SE). Similar to surgery and medica tasks, industrial and physical tasks [3], [18], [19], [20] also typically involve AS and SC roles. Lastly, in social and gaming [21], [22], [23] tasks, all three intermediate roles addressed where there exist various solutions to carry out such tasks.
IV. TECHNICAL COMPONENTS
We have previously defined a specialized use case of telementoring with three distinguishing characteristics of “(1) mentors and mentees of different levels, (2) sharing of first-person and eye-level view via smartglass, and (3) extended use cases with bi-directional interaction [3]”. To realize the three characteristics of remote collaboration, the following functional components are designed and included in our reference framework.
4.1. Components for Personal Tracking
Due to the always-(on, accessible, connected) characteristics of wearable devices, embedded sensors on these devices are used to monitor and analyze users physically via various physiological sensors and IMU (inertial measurement unit). A notable example includes location-based services where the user’s location is tracked for providing location-based notifications. Personal big data collected from smartglass and smartwatches are important sources for machine learning approaches. Moreover, quantified-self [6] and personal tracking [7] approaches have already been effectively deployed in healthcare, sleep, and food journaling. For example, users' activities can be tracked with motion sensors and physiological sensors to measure physical fatigue as well as a cognitive load of a task of interest. The lessons and insights from these studies pave the way for preparing successful remote collaboration. Power-consuming personal tracking components for wearable devices should be carefully designed to deal with the limited battery capacity on these compact and portable devices.
4.2. Components for Audio-visual Teleconference
By the definition of AS, SC, and SE, participants (including mentors, mentees, and peers) communicate with audio and video in real-time. At least one side (i.e., mentor or mentee) needs to wear smartglasses, but the remaining side can use non-smartglasses computers. To effectively communicate in terms of real-time feedback and communicative instructions, audio-visual teleconference functions are required. It is critical to provide and share users’ views when collaborating, while the actual resolution of the streaming video may differ in varioustasks of interest (i.e., medical/surgical, educational, and industrial).
Due to the COVID-19 pandemic, audio-visual teleconference functions are used globally in several platforms such as Zoom, Google MeetTM, Microsoft Teams and implemented with open sources such as Jitsi Meet (https://github.com/jitsi/jitsi-meet), BigBlueButton (https://bigbluebutton.org/), OpenVidu (https://openvidu.io/), Element (https://element.io/), and Wire (https://wire.com/en/) as well as using WebRTC API [3]. Audio-visual components should be designed and implemented considering various display sizes and display resolution.
4.3. Components for Supporting Various Interaction Styles
Participants in remote collaboration need tools and user interfaces for pointing or marking problematic areas that needs the other participants’ feedback. Mentors also need tools and user interfaces to guide and instruct the mentees. To support the needs of mentors, mentees, and collaborators, various interaction styles including touchbased, voice-based, gesture-based, and telestration (i.e., pointing, annotating, drawing) are required.
Wearable devices’ various embedded sensors can be used to improve the user experience and usability of the intermediary roles of AS, SC, and SE. For example, tactile physical feedback generated via vibration can be used to alert users in situations where the users’ attention level is low (i.e., performing some activities), hard-of-hearing users (i.e., physically, and noisy ambient sound) are involved, and multimodality are exploited for emphasis.
Fig. 3 shows our proposed technical components for the smartglass-based remote collaboration in three categories of personal tracking components, teleconference components, and interaction components. Personal tracking components are employed to track the wearer’s physiological, cognitive load, physical fatigue, and activity levels while managing and analyzing personal big data generated from the smartglass. Teleconference components are used to live-stream video and audio to mentors and mentees, respectively. Interaction components allow the user to pick and choose various interaction styles (i.e., touch-based, voice-based, gesture-based, telestration) that suit the user’s task and environment. All three components rely on embedded sensors such as motion sensors [24] and smartglass hardware. For example, on a Google Glass Enterprise Edition 2, activity levels in personal tracking components, video and audio streaming in teleconference components, and touch-based UI, voicebased UI, and gesture-based UI in interaction components can be exploited to implement remote collaboration applications. Similar to personal tracking components, power-consuming and computationally intensive interaction components such as voice-based UIs and gesture-based UIs must be carefully integrated to handle limited battery capacity.
Fig. 3. Technical components for SAIT.
V. DISCUSSION
This section illustrates how the previously mentioned technical components can or are being used in smartglassassisted applications. Table 2 shows three remote collaboration applications of VIRNECT Remote by VIRNECT (https://virnect.com), Spatial 4.0 by Spatial (https://spatial.io), and metaverse platform by MAXST (https://maxst.com). Personal tracking components are used by a metaverse platform by MAXST, for tracking the user’s physical location within the office space. All three applications use teleconference components to share camera views for remote collaboration. Lastly, various interaction styles including touch-based UI, voice-based UI, and gesture-based UI are used in these examples as well.
Table 2. Smartglass-assisted application examples.
VI. CONCLUSION
Wearable devices including smartglasses are getting popular among technical and healthcare professionals. More casual use cases involving gaming and work-fromhome applications are being developed and explored. Remote collaboration, a popular topic studied in medical and surgical domains, can be tailored to include smartglasses and various common functionalities to realize a killer application for wearables. In this paper, we analyzed intermediary roles in remote collaboration and proposed technical components for realizing smartglass-assisted interactive remote collaboration. More specifically, we defined various intermediary roles of remote collaboration participants and elaborated technical components between those interfaces that include three main components of personal tracking, teleconference, and interaction. Further studies are deserved for the practical development and deployment of the proposed reference components using commercially available smartglasses to iteratively gain insights and improve smartglasses-based interactive remote collaboration applications.
Acknowledgement
This work was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF2021R1F1A1063340). Photos used in Fig. 1 are from Pexels with the free license to use (https://www.pexels.com/license). Icons used in Fig. 2 are Organisation Behaviour, Supervising, and Value by Massupa Kaewgahya from NounProject.com. Examples of smartglass-assisted application videos are from YouTube which can be found on the provided links.
References
- D. T. Bui, T. Barnett, H. T. Hoang, and W. Chinthammit, "Tele-mentoring using augmented reality technology in healthcare: a systematic review," Australasian Journal of Education Technology, vol. 37, no. 4, pp. 68-88, May 2021.
- S. Erridge, D. K. T. Yeung, H. R. H. Patel, and S. Purkayastha, "Telementoring of surgeons, a systematic review," Surgical Innovation, vol. 26, pp.95-111, 2018. https://doi.org/10.1177/1553350618813250
- Y. Lee, B. Yoo, and S.-H. Lee, "Sharing ambient objects using real-time point cloud streaming in web-based XR remote collaboration," in Proceedings of the 26th International Conference on 3D Web Technology, Article 4, Pisa, Italy, pp. 1-9, Nov. 2021.
- H. Yoon, "Opportunities and challenges of smartglass-assisted interactive telementoring," Applied System Innovation, vol. 4, no. 3, Article 56, Aug. 2021.
- H. Yoon, H. Kim, and W. Woo, "Personal companion: Personalized user interface for u-service discovery, selection and interaction," in Proceedings of International Conference on Universal Access in Human-Computer Interaction, Beijing, China, pp. 1052-1061, Jul. 2007.
- M. Rohs and P. Zweifel, "A conceptual framework for camera phone-based interaction techniques," in Proceedings of International Conference on Pervasive Computing, Munich, Germany, pp. 171-189, May 2005.
- H. Yoon, Y. Y. Doh, M. Y. Yi, and W. Woo, "A conceptual framework for augmented smart coach based on quantified holistic self," in Proceedings of International Conference on Distributed, Ambient, and Pervasive Interactions, Heraklion, Crete, Greece, pp. 498-508, Jun. 2014.
- E. K. Choe, N. B. Lee, B. Lee, W. Pratt, and J. A. Kientz, "Understanding quantified-selfers' practices in collecting and exploring personal data," in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Toronto, Ontario, Canada, pp. 1143-1152, April 2014.
- E. K. Choe and B. Lee, "Toward supporting personalized tracking experiences in healthcare," Interactions, vol. 27, no. 1, pp. 84-87, Jan.-Feb. 2020. https://doi.org/10.1145/3371287
- H. Yoon, S. K. Kim, Y. Lee, and J. Choi, "Google glass-supported cooperative training for health professionals: A case study based on using remote desktop virtual support," Journal of Multidisciplinary Healthcare, vol. 14, pp.1451-1462, Jun. 2021. https://doi.org/10.2147/JMDH.S311766
- S. K. Kim, Y. Lee, H. Yoon, and J. Choi, "Adaptation of extended reality smart glasses for core nursing skill training among undergraduate nursing students: Usability and feasibility study," Journal of Medical Internet Research, vol. 23, no. 3, e24313, Mar. 2021. https://doi.org/10.2196/24313
- T. Ong, H. Wilczewski, S. Paige, H. Soni, B. Welch, and B. Bunnell, "Extended reality for enhanced telehealth during and beyond COVID-19: Viewpoint," JMIR Serious Games, vol. 9, no. 3, e26520, Jul.-Sep. 2021. https://doi.org/10.2196/26520
- K. Klinker, M. Wiesche, and H. Krcmar, "Digital transformation in health care: Augmented reality for hands-free service innovation," Information Systems Frontiers, vol. 22, pp. 1419-1431, 2020. https://doi.org/10.1007/s10796-019-09937-7
- E. Rojas-Munoz, M. E. Cabrera, C. Lin, D. Andersen, V. Popescu, K. Anderson, B. L. Zarzaur, B. Mullis, and J. P. Wachs, "The system for telementoring with augmented reality (STAR): A head-mounted display to improve surgical coaching and confidence in remote areas," Surgery, vol. 167, pp. 724-731, 2020. https://doi.org/10.1016/j.surg.2019.11.008
- A. Semsar, H. McGowan, Y. Feng, H. R. Zahiri, A. Park, A. Kleinsmith, and H. M. Mentis, "Quality of and attention to instructions in telementoring," Proceedings of the ACM Human-Computer Interaction, vol. 4, pp. 1-21, 2020.
- L. R. Rochlen, R. Levine, A. R. Tait, "First-person point-of-view-augmented reality for central line insertion training," Simulation in Healthcare, vol. 12, no. 1, pp. 57-62, Feb. 2017. https://doi.org/10.1097/SIH.0000000000000185
- J. A. Greenberg, E. Schwarz, J. Paige, J. Dort, and S. Bachman, "At-home hands-on surgical training during COVID19: Proof of concept using a virtual telementoring platform," Surgical Endoscopy, vol. 35, pp. 1963-1969, 2021. https://doi.org/10.1007/s00464-021-08470-6
- P. Wang, S. Zhang, X. Bai, M. Billinghurst, L. Zhang, S. Wang, D. Han, H. Lv, and Y. Yan, "A gesture- and head-based multimodal interaction platform for MR remote collaboration," The International Journal of Advanced Manufacturing Technology, vol. 105, pp. 3031-3043, 2019. https://doi.org/10.1007/s00170-019-04434-2
- P. Wang, X. Bai, M. Billinghurst, S. Zhang, W. He, D. Han, Y. Wang, H. Min, W. Lan, an S. Han, "Using a head pointer or eye gaze: The effect of gaze on spatial AR remote collaboration for physical tasks," Interacting with Computers, vol. 32, no. 1, pp. 153-169, Aug. 2020. https://doi.org/10.1093/iwcomp/iwaa012
- S. Kim, G. Lee, M. Billinghurst, and W. Huang, "The combination of visual communication cues in mixed reality remote collaboration," Journal on Multimodal User Interfaces, vol. 14, no. 4, pp. 321-335, 2020. https://doi.org/10.1007/s12193-020-00335-x
- Y. Yuan, J. Cao, R. Wang, and S. Yarosh, "Tabletop games in the age of remote collaboration: Design opportunities for a socially connected game experience," in Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, Article 436, 2021.
- F. Kleiman, S. Meijer, and M. Janssen, "Gaming for meaningful interactions in teleworking lessons learned during the COVID-19 pandemic from integrating gaming in virtual meetings," Digital Government: Research and Practice, vol. 1, no. 4, Article 31, Dec. 2020.
- N. Osmers, M. Prilla, O. Blunk, G. George Brown, M. Janssen, and N. Kahrl, "The role of social presence for cooperation in augmented reality on head mounted devices: A literature review," in Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, Article 457, May 2021.
- J.-H. Kim, G.-S. Hong, B.-G. Kim, and D. P. Dogra, "deepGesture: Deep learning-based gesture recognition scheme using motion sensors," Displays, vol. 55, pp. 38-45, Dec. 2018. https://doi.org/10.1016/j.displa.2018.08.001