{"id":290,"date":"2024-12-31T14:33:00","date_gmt":"2024-12-31T14:33:00","guid":{"rendered":"https:\/\/vrst.acm.org\/vrst2025\/?page_id=290"},"modified":"2025-12-12T05:28:49","modified_gmt":"2025-12-12T05:28:49","slug":"proceedings","status":"publish","type":"page","link":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/proceedings\/","title":{"rendered":"Proceedings"},"content":{"rendered":"\n<html xmlns:bkstg=\"http:\/\/www.atypon.com\/backstage-ns\" xmlns:urlutil=\"java:com.atypon.literatum.customization.UrlUtil\" xmlns:pxje=\"java:com.atypon.frontend.services.impl.PassportXslJavaExtentions\">\n   <head>\n      <meta http-equiv=\"Content-Type\" content=\"text\/html; charset=UTF-8\">\n      <meta http-equiv=\"Content-Style-Type\" content=\"text\/css\">\n      <style type=\"text\/css\">\n            #DLtoc {\n            font: normal 12px\/1.5em Arial, Helvetica, sans-serif;\n            }\n\n            #DLheader {\n            }\n            #DLheader h1 {\n            font-size:16px;\n            }\n\n            #DLcontent {\n            font-size:12px;\n            }\n            #DLcontent h2 {\n            font-size:14px;\n            margin-bottom:5px;\n            }\n            #DLcontent h3 {\n            font-size:12px;\n            padding-left:20px;\n            margin-bottom:0px;\n            }\n\n            #DLcontent ul{\n            margin-top:0px;\n            margin-bottom:0px;\n            }\n\n            .DLauthors li{\n            display: inline;\n            list-style-type: none;\n            padding-right: 5px;\n            }\n\n            .DLauthors li:after{\n            content:\",\";\n            }\n            .DLauthors li.nameList.Last:after{\n            content:\"\";\n            }\n\n            .DLabstract {\n            padding-left:40px;\n            padding-right:20px;\n            display:block;\n            }\n\n            .DLformats li{\n            display: inline;\n            list-style-type: none;\n            padding-right: 5px;\n            }\n\n            .DLformats li:after{\n            content:\",\";\n            }\n            .DLformats li.formatList.Last:after{\n            content:\"\";\n            }\n\n            .DLlogo {\n            vertical-align:middle;\n            padding-right:5px;\n            border:none;\n            }\n\n            .DLcitLink {\n            margin-left:20px;\n            }\n\n            .DLtitleLink {\n            margin-left:20px;\n            }\n\n            .DLotherLink {\n            margin-left:0px;\n            }\n\n        <\/style>\n      <title>VRST &#8217;25: Proceedings of the 2025 31st ACM Symposium on Virtual Reality Software and Technology<\/title>\n   <\/head>\n   <body>\n      <div id=\"DLtoc\">\n         <div id=\"DLheader\">\n            <h1>VRST &#8217;25: Proceedings of the 2025 31st ACM Symposium on Virtual Reality Software and Technology<\/h1><a class=\"DLcitLink\" title=\"Go to the ACM Digital Library for additional information about this proceeding\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/proceedings\/10.1145\/3756884\"><img decoding=\"async\" class=\"DLlogo\" alt=\"Digital Library logo\" height=\"30\" src=\"https:\/\/dl.acm.org\/specs\/products\/acm\/releasedAssets\/images\/footer-logo1.png\">\n               Full Citation in the ACM Digital Library\n               <\/a><\/div>\n         <div id=\"DLcontent\">\n            <h2>SESSION 1: Interaction Design and Input Techniques I<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766047\">SpatialMouse: A Hybrid Pointing Device for Seamless Interaction Across 2D and 3D Spaces<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Sebastian Hubenschmid<\/li>\n               <li class=\"nameList\">Johannes Zagermann<\/li>\n               <li class=\"nameList\">Robin Erb<\/li>\n               <li class=\"nameList\">Tiare Feuchtner<\/li>\n               <li class=\"nameList\">Jens Grubert<\/li>\n               <li class=\"nameList\">Markus Tatzgern<\/li>\n               <li class=\"nameList\">Dieter Schmalstieg<\/li>\n               <li class=\"nameList Last\">Harald Reiterer<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We introduce the <em>SpatialMouse<\/em>, a hybrid pointing device that combines the capabilities of a desktop mouse with\n                     the spatial input of a virtual reality (VR) controller, enabling seamless transitions\n                     between 2D and 3D interaction spaces in immersive mixed reality environments. Holistic\n                     usage scenarios in mixed reality involve tasks suited alternately to 2D or 3D information\n                     spaces. Yet, existing input devices excel in either 2D or 3D, but not both, making\n                     it necessary to switch between multiple input devices (e.g., mouse and VR controller).\n                     Our <em>SpatialMouse<\/em> addresses this issue, offering the affordances of a desktop mouse for indirect 2D\n                     pointing and the spatial capabilities of VR controllers with six degrees of freedom.\n                     In a user study with 12&nbsp;participants, our prototype significantly reduced perceived\n                     task load and improved user experience compared to switching between separate devices.\n                     We extract design recommendations to further support such hybrid input approaches.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765982\">Gated Temporal Shifts with Depth-Efficient Channel Attention for Real-Time Hand-Gesture\n                  Interaction<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Salah eddine Laidoudi<\/li>\n               <li class=\"nameList\">Madjid Maidi<\/li>\n               <li class=\"nameList Last\">Samir Otmane<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We introduce a compact video-classification pipeline for real-time dynamic hand-gesture\n                     recognition in mixed-reality (MR) settings. The network marries a MobileNetV3 backbone\n                     with two purpose-built temporal components: (1) a Gated Discriminative Temporal Shift\n                     Module (G-DiTSM) that inserts first-order motion differences and learns channel-wise\n                     gates to fuse them adaptively, and (2) a lightweight Depth-Efficient Channel Attention\n                     (DepthECA) block that recalibrates spatial features on the fly. Operating on eight\n                     sparsely sampled frames per clip (Temporal Segment Network paradigm), the resulting\n                     model contains 2.65 M parameters and requires only 0.084 GFLOPs per inference. Evaluated\n                     on the RGB-only 20BN Jester benchmark (148k clips spanning 27 gesture classes) recorded\n                     from front-facing viewpoints. The system reaches 95.34% Top-1 and 99.80% Top-5 accuracy,\n                     surpassing recent 3D CNNs and transformer baselines while being an order of magnitude\n                     lighter. Ablations confirm that DepthECA and G-DiTSM provide complementary gains (+18.78%\n                     and +0.93% Top-1, respectively, over the MobileNetV3 baseline). Because all components\n                     are plug-and-play and introduce minimal overhead, the architecture is well suited\n                     to the tight latency and power budgets of standalone MR headsets, paving the way for\n                     natural grab, rotate, and command interactions using only on-board RGB cameras<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765994\">Enhancing the Sensation of Depth in Mid-Air Image Interactions with Pictorial Depth\n                  Cues<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Saki Kominato<\/li>\n               <li class=\"nameList Last\">Naoya Koizumi<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>In virtual reality, visual information plays a critical role, and head-mounted displays\n                     are widely recognized as the primary means of presentation. However, non-wearable\n                     approaches such as projection mapping and mid-air images have also been explored.\n                     Mid-air images present content near real objects without screens, making them promising\n                     for mixed reality. Yet, their lack of physicality weakens depth perception and diminishes\n                     the sensation of pressing buttons. We tested whether pictorial cues (shading, shadow,\n                     size) enhance depth perception and button-press sensation in mid-air image UIs. Each\n                     experiment involved 14\u201316 participants. These cues increased perceived depth and improved\n                     pressing sensation. These findings suggest that pictorial cues can compensate for\n                     the absence of physical sensation and enhance the usability of mid-air image UIs.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765975\">Enhancing Freehand VR Interaction Using Fingertip Deformation on User Performance<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Kosuke Morimoto<\/li>\n               <li class=\"nameList\">Nobuhito Kasahara<\/li>\n               <li class=\"nameList\">Shota Yamanaka<\/li>\n               <li class=\"nameList\">Homei Miyashita<\/li>\n               <li class=\"nameList Last\">Keita Watanabe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This study investigated the use of the deformation of the \ufb01ngertips of a virtual hand\n                     to enhance depth perception during freehand interaction in a virtual reality (VR).\n                     Artificial fingertip deformation may generate mapping of real hand position and pseudo-haptics,\n                     improving UI usability. We conducted two experiments focusing on depth manipulation\n                     in both pointing and steering tasks. Our results revealed that changes in fingertip\n                     shape reduced operation time in pointing tasks and improved accuracy in steering tasks.\n                     Additionally, we conducted subjective evaluation surveys for both experiments, which\n                     showed improvements in pseudo-haptics, spatial perception, and user experience. Based\n                     on these results, we propose several applications and demonstrate that fingertip deformations\n                     in virtual hands can contribute to better 3D UI design.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765990\">Trade-offs in Virtual Grasping: The Interplay of Interaction Fidelity and Object Affordance<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jemin Lee<\/li>\n               <li class=\"nameList\">Jeonghyeon Kim<\/li>\n               <li class=\"nameList\">Hyeongjun Kang<\/li>\n               <li class=\"nameList\">Hoon Ryu<\/li>\n               <li class=\"nameList Last\">Youngwon Kim<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>In Virtual Reality (VR), object grasping is a core interaction that critically influences\n                     both user immersion and task performance. While contemporary systems offer both high-precision\n                     controllers and intuitive hand tracking, they present a trade-off between performance\n                     and naturalness. However, empirical guidance for selecting an optimal grasping method\n                     for object grasping remains limited. In particular, how object shape and size (as\n                     affordance-related factors) modulate this trade-off within a standardized pick-and-place\n                     paradigm is underexplored.<\/p>\n                  <p>We investigate the interplay between interaction fidelity and object shape\/size and\n                     its impact on user performance and experience in a controlled pick-and-place task.\n                     We conducted a within-subjects study with a 3 (grasping modality: controller, pinch,\n                     plausible gesture) \u00d7 5 (object shape: cube, sphere, cylinder, handled mug, complex\n                     model) \u00d7 3 (object size) factorial design. We measured objective performance (task\n                     completion time, placement accuracy) and subjective experience (NASA-TLX workload,\n                     IPQ presence).<\/p>\n                  <p>Our findings provide evidence-based answers to the scoped question: \u201cWhich grasping\n                     method is best suited for an object of a given shape and size in a pick-and-place\n                     task?\u201d Ultimately, this work offers actionable guidelines to help VR developers design\n                     effective and satisfying object-grasping interactions tailored to users\u2019 task goals\n                     and virtual environments, without claiming a single universally \u201cbest\u201d method.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766042\">Rethinking Gesture Recognition: Toward Fatigue-Aware sEMG Gesture Recognition for\n                  VR Interaction<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Kirti Lakra<\/li>\n               <li class=\"nameList\">Chaitanya Garg<\/li>\n               <li class=\"nameList\">Pranav Jain<\/li>\n               <li class=\"nameList\">Rudra Jyotirmay<\/li>\n               <li class=\"nameList Last\">Pushpendra Singh<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Advances in virtual reality (VR) are transforming interaction paradigms by shifting\n                     towards gesture-based control driven by physiological sensing, enabling more intuitive\n                     and embodied experiences. Surface electromyography (sEMG) is emerging as a reliable\n                     modality for this hands-free and expressive gesture recognition in VR. However, prolonged\n                     mid-air gestures can lead to muscle fatigue and physiological changes that degrade\n                     overall recognition performance. Further, this degradation is not uniform across gestures\n                     which can impact user performance and experience in VR applications. While existing\n                     literature has shown that fatigue alters sEMG signals, its effects during extended\n                     immersive interaction and across various gestures remain underexplored. We conducted\n                     a 35-participant study in which each participant continuously performed five gesture\n                     in VR for 20 minutes each, while we collected high-resolution sEMG data from eight\n                     forearm sensors and real-time subjective fatigue ratings using the Borg CR10 scale.\n                     Further, we evaluate how gesture recognition models behave under fatigue and explore\n                     the impact of incorporating both objective (signal-derived) and subjective (user-reported)\n                     fatigue features into classification models. Our results show that integrating fatigue\n                     signals enhances model robustness and improves recognition accuracy during extended\n                     use.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766043\">Freehand Sketch-Based 3D Reconstruction with Contour Constraints via Elastic Metrics<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Gaoyang Liu<\/li>\n               <li class=\"nameList\">Chunyang Huo<\/li>\n               <li class=\"nameList\">Zhentong Xu<\/li>\n               <li class=\"nameList\">Junli Zhao<\/li>\n               <li class=\"nameList\">Yishan Dong<\/li>\n               <li class=\"nameList\">Baodong Wang<\/li>\n               <li class=\"nameList\">Xinbin Sun<\/li>\n               <li class=\"nameList\">Yingying Jiang<\/li>\n               <li class=\"nameList Last\">Yuxuan Zhao<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Sketch-based 3D reconstruction enables intuitive content creation through freehand\n                     drawings, yet generating high-fidelity 3D models from geometrically ambiguous, structurally\n                     simplified, and sparse sketches remain challenging. To overcome existing methods\u2019\n                     limitations in sketch-style generalization, contour accuracy, and suboptimal texture\n                     effects, we propose an end-to-end framework that generates textured 3D models directly\n                     from a single freehand sketch and semantic labels. To address the scarcity of paired\n                     freehand sketch training data, we introduce a 3D model-based automated sketch generation\n                     method for extracting mesh contours via a 3D mesh-to-sketch pipeline and synthesizing\n                     freehand-style sketches employing a Transformer-based stroke generator to construct\n                     a paired dataset of hand-drawn sketches and 3D models. Meanwhile, we design a contour\n                     constraint mechanism that jointly optimizes projection-space Chamfer distances and\n                     elastic metrics, significantly enhancing the reconstruction accuracy of complex geometries.\n                     Furthermore, we integrate a semantic-guided texture generation module using Text2Tex\n                     with depth-aware diffusion models and dynamic view-optimization strategies, achieving\n                     a complete geometry-appearance integrated modeling pipeline. Finally, extensive experimental\n                     results demonstrate that our method outperforms existing structural reconstruction\n                     and texture synthesis approaches, exhibiting strong generalization capabilities and\n                     practical applicability.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 2: Locomotion and Wayfinding<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765970\">Not All WIP Are Perceived Equally: Different Speed Expectations in Seated Walk-in-Place\n                  Locomotion<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Yusuke Kitaura<\/li>\n               <li class=\"nameList\">Keigo Hattori<\/li>\n               <li class=\"nameList\">Fumihiko Nakamura<\/li>\n               <li class=\"nameList\">Yuta Kataoka<\/li>\n               <li class=\"nameList\">Fumihisa Shibata<\/li>\n               <li class=\"nameList\">Asako Kimura<\/li>\n               <li class=\"nameList Last\">Shohei Mori<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Gesture-based locomotion enhances immersion in virtual reality (VR), with seated motion\n                     being crucial for accessibility and prolonged use. However, existing techniques often\n                     apply uniform gesture-to-walking speed mappings, ignoring the fact that different\n                     gestures involve varying levels of physical effort and subjective impressions. This\n                     mismatch can degrade the user experience. This study investigates how three seated\n                     gestures with different physical loads\u2014Tap-in-Place (TIP), Swing-in-Place (SIP), and\n                     Grip-in-Place (GIP)\u2014influence users\u2019 expected walking speed. While the evaluations\n                     revealed unique experiential trade-offs for each gesture, our primary finding is a\n                     consistent perceptual pattern in the expectation of walking speed: Users expected\n                     to walk fastest with SIP, followed by GIP, then TIP (SIP &gt; GIP &gt; TIP). These results\n                     demonstrate that a one-size-fits-all approach is insufficient and provide empirical\n                     recommendations for designing more intuitive seated VR locomotion systems that align\n                     walking speed with user perception.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765977\">VisionPort: Enhancing Building-Scale Indoor Navigation through Obstacle-Removing Point-and-Teleport\n                  Techniques<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jia-Jun Wang<\/li>\n               <li class=\"nameList\">Tzu-Wei Mi<\/li>\n               <li class=\"nameList\">Ting-Han Wu<\/li>\n               <li class=\"nameList\">Chih-Jou Li<\/li>\n               <li class=\"nameList Last\">Liwei Chan<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We present VisionPort, an enhanced Point-and-Teleport technique designed for navigating\n                     building-scale indoor virtual environments, specifically addressing the challenges\n                     posed by obstacles including walls, ceilings, and floors. VisionPort is available\n                     in two versions: VisionPort-essential and VisionPort-full. VisionPort-essential removes\n                     only the necessary portion of an obstacle targeted by the pointer, revealing the landing\n                     position behind it. In contrast, VisionPort-full allows for the complete removal of\n                     obstacles. Both versions enable users to seamlessly pass through barriers during the\n                     Point-and-Teleport locomotion process, while preserving the natural flow of navigation\n                     within the virtual building. Our evaluation, conducted in a multi-floor building setting,\n                     demonstrates that VisionPort improves navigation by reducing the time, head movement,\n                     and distance required to reach destinations. While VisionPort-full enhances efficiency,\n                     VisionPort-essential provides users with a greater sense of control, reflecting diverse\n                     preferences among participants.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766023\">Tunnels vs. Wires: A Comparative Analysis of Two 3D Steering Tasks in Virtual Environments<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Mohammadreza Amini<\/li>\n               <li class=\"nameList\">Wolfgang Stuerzlinger<\/li>\n               <li class=\"nameList\">Shota Yamanaka<\/li>\n               <li class=\"nameList\">Hai-Ning Liang<\/li>\n               <li class=\"nameList Last\">Anil Ufuk Batmaz<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Steering involves continuous movement along constrained paths, well-studied in 2D.\n                     The extensions to 3D using the Ring-and-Wire and Ball-and-Tunnel tasks were often\n                     treated as interchangeable in previous work. In this paper, we directly compare these\n                     two tasks through a within-subjects user study (<em>n<\/em> = 18) with varying 3D path orientations. The results show that Ring-and-Wire significantly\n                     outperformed Ball-and-Tunnel, with 17.17% lower task time, 21.65% higher throughput,\n                     and 21.52% faster average speed. Participants also preferred Ring-and-Wire and reported\n                     lower workload. Visual ambiguity, especially near the tunnel\u2019s rear surface, complicated\n                     spatial perception in the Ball-and-Tunnel task. We thus recommend that future studies\n                     choose 3D steering tasks carefully for experiments, as the two tasks are not interchangeable.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766033\">IGUANA: Immersive Guidance, Navigation, and Control for Consumer UAV<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Victor Victor<\/li>\n               <li class=\"nameList\">Tania Krisanty<\/li>\n               <li class=\"nameList\">Matthew McGinity<\/li>\n               <li class=\"nameList\">Stefan Gumhold<\/li>\n               <li class=\"nameList Last\">Uwe A\u00dfmann<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>As the markets for unmanned aerial vehicles (UAVs) and mixed reality (MR) headsets\n                     continue to grow, recent research has increasingly explored their integration, which\n                     enables more intuitive, immersive, and situationally aware control systems. We present\n                     IGUANA, an MR-based immersive guidance, navigation, and control system for consumer\n                     UAVs. IGUANA introduces three key elements beyond conventional control interfaces:\n                     (1)&nbsp;a 3D terrain map interface with draggable waypoint markers and live camera preview\n                     for high-level control, (2)&nbsp;a novel spatial control metaphor that uses a virtual ball\n                     as a physical analogy for low-level control, and (3)&nbsp;a spatial overlay that helps\n                     track the UAV when it is not visible with the naked eye or visual line of sight is\n                     interrupted. We conducted a user study to evaluate our design, both quantitatively\n                     and qualitatively, and found that (1)&nbsp;the 3D map interface is intuitive and easy to\n                     use, relieving users from manual control and suggesting improved accuracy and consistency\n                     with lower perceived workload relative to conventional dual-stick controller, (2)&nbsp;the\n                     virtual ball interface is intuitive but limited by the lack of physical feedback,\n                     and (3)&nbsp;the spatial overlay is very useful in enhancing the users\u2019 situational awareness.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766038\">Beyond Parabolas: Linear Pointer Teleportation for Vertical Navigation in VR<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Hibiki Kirihata<\/li>\n               <li class=\"nameList Last\">Tomokazu Ishikawa<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual reality teleportation using hand tracking faces significant challenges in\n                     vertical navigation, with conventional parabolic methods requiring users to mentally\n                     calculate trajectories and landing points. We present two novel linear pointer-based\n                     teleportation methods\u2014SphereBackcast and Penetration\u2014that enable intuitive vertical\n                     movement through direct pointing and collision handling strategies. Through two experiments\n                     involving 34 participants total, we evaluated these methods against traditional parabolic\n                     teleportation across diverse environments ranging from flat terrain to multi-level\n                     structures with 4m height differences. Results demonstrate that the Penetration method\n                     significantly outperforms parabolic teleportation, achieving 53% faster completion\n                     times (13.85s vs 29.38s) and 35% lower path deviation (0.552m vs 0.851m) in environments\n                     with 2m+ vertical elements. Controller input provided 12-18% performance improvements\n                     over hand tracking while maintaining consistent relative advantages of linear methods.\n                     Both proposed methods received superior usability ratings (SUS: 69.2 and 68.75 vs\n                     64.66) and reduced cognitive workload (NASA-TLX: 25.35 and 25.08 vs 33.05), with 50%\n                     of participants preferring the Penetration method. These findings establish efficient\n                     teleportation techniques that address critical limitations in current VR navigation,\n                     offering practical solutions for applications requiring vertical movement such as\n                     adventure games and architectural visualization.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766040\">You Have Arrived&#8230; Kind of: Investigating the Limits of Undetectable Destination\n                  Displacement During Teleportation<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Taylor Laird<\/li>\n               <li class=\"nameList\">Jasmine Joyce DeGuzman<\/li>\n               <li class=\"nameList\">Gerd Bruder<\/li>\n               <li class=\"nameList\">Carolina Cruz-Neira<\/li>\n               <li class=\"nameList Last\">Dirk Reiners<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Teleportation has become a popular locomotion method for virtual reality due to lesser\n                     demands on physical space and decreased levels of motion sickness compared to other\n                     methods. However, prior work has shown that these advantages come at the cost of impaired\n                     spatial perception and awareness, the extent to which is still largely unknown. In\n                     this work, we present a within-subjects study (<em>N<\/em> = 29) that explores the effects of teleportation on spatial perception by investigating\n                     how much humans can be unknowingly displaced relative to their intended destination\n                     during teleportation. After teleporting to the specified location, participants indicated\n                     the direction and magnitude (small, medium, large) of the perceived shift or rotation.\n                     Displacement from the target happened either as a translation in the forward- or strafe-axis,\n                     or a rotation about the up-axis at the intended target. Each displacement condition\n                     included eleven offsets that were repeated six times. Our results indicate points\n                     of subjective equality, which show a significant perceptual shift along the forward-direction,\n                     as well as detection thresholds, which indicate a comparatively wide range in which\n                     humans are unable to detect induced shifts. Furthermore, our results show that even\n                     if humans are able to detect these shifts, larger ones can be introduced before their\n                     magnitudes are rated as medium or large, which provides ample opportunities for interface\n                     designers who want to leverage these results in virtual reality.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 3: Cybersickness, Health, and Digital Twins<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766010\">Unmanned Aerial Vehicles Control in a Digital Twin: Exploring the Effect of Different\n                  Points of View on User Experience in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Francesco Vona<\/li>\n               <li class=\"nameList\">Mohamed Amer<\/li>\n               <li class=\"nameList\">Omar Abdellatif<\/li>\n               <li class=\"nameList\">Michelle Celina Hallmann<\/li>\n               <li class=\"nameList\">Maximilian Warsinke<\/li>\n               <li class=\"nameList\">Adriana-Simona Mihaita<\/li>\n               <li class=\"nameList Last\">Jan-Niklas Voigt-Antons<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Controlling Unmanned Aerial Vehicles (UAVs) is a cognitively demanding task, with\n                     accidents often arising from insufficient situational awareness, inadequate training,\n                     and bad user experiences. Providing more intuitive and immersive visual feedback\u2014particularly\n                     through Digital Twin technologies\u2014offers new opportunities to enhance pilot awareness\n                     and the overall experience quality. In this study, we investigate how different virtual\n                     points of view (POVs) influence user experience and performance during UAV piloting\n                     in Virtual Reality (VR), utilizing a digital twin that faithfully replicates the real-world\n                     flight environment. We developed a VR application that enables participants to control\n                     a physical DJI Mini 4 Pro drone while immersed in a digital twin with four distinct\n                     camera perspectives: Baseline View (static external), First Person View, Chase View,\n                     and Third Person View. Nineteen participants completed a series of ring-based obstacle\n                     courses from each perspective. In addition to objective flight data, we collected\n                     standardized subjective assessments of user experience, presence, workload, cybersickness,\n                     and situational awareness. Quantitative analyses revealed that the First Person View\n                     was associated with significantly higher mental demand and effort, greater trajectory\n                     deviation, but smoother control inputs compared to the Third Person and Chase perspectives.\n                     Complementing these findings, preference data indicated that the Third Person View\n                     was most consistently favored, whereas the First Person View elicited polarized reactions.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765997\">An In-the-Wild Accessibility Evaluation of Apple Vision Pro for Deaf or Hard of Hearing\n                  Users<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Roshan Mathew<\/li>\n               <li class=\"nameList\">Kai Wu<\/li>\n               <li class=\"nameList Last\">Wendy Dannels<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Extended Reality (XR) technologies, including Mixed Reality (MR), Augmented Reality\n                     (AR), and Virtual Reality (VR), are blurring the lines between physical and digital\n                     environments, transcending the limitations of traditional two-dimensional (2D) interfaces.\n                     This shift toward embodied, often context-aware spatial interaction offers broad potential\n                     benefits, yet also introduces unique challenges, especially for certain user groups.\n                     For people who are Deaf or hard of hearing (DHH), XR\u2019s immersive and multi-sensory\n                     environments provide unique opportunities to improve accessibility. However, design\n                     principles that work well in 2D interfaces may not always translate seamlessly into\n                     immersive contexts, creating new accessibility barriers. The launch of the Apple Vision\n                     Pro marks a significant moment in the mainstream adoption of spatial computing, yet\n                     little is known about its accessibility implications for deaf users. To explore this\n                     emerging area, we conducted an in-the-wild, open-ended study with five deaf participants\n                     who have diverse communication preferences, evaluating the Apple Vision Pro in everyday\n                     situations. Based on this exploratory evaluation, we identify key accessibility challenges\n                     and opportunities and provide practical recommendations to make spatial computing\n                     more inclusive for deaf users in the future.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766020\">Design and Evaluation of a Mixed Reality Biofeedback System for Home-Based Physiotherapy\n                  Exercises<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Nikolai Hepke<\/li>\n               <li class=\"nameList\">Moritz Scherer<\/li>\n               <li class=\"nameList\">Steffen M\u00fcller<\/li>\n               <li class=\"nameList\">Benjamin Weyers<\/li>\n               <li class=\"nameList Last\">J\u00f6rg Lohscheller<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Home-based exercise programs are a cornerstone in managing chronic non-specific back\n                     pain. However, their effectiveness is often limited by low adherence and incorrect\n                     exercise execution. This study presents and evaluates a Mixed Reality (MR) biofeedback\n                     system that tracks body motion using a multi-Kinect setup and provides real-time feedback\n                     via the Microsoft HoloLens 2. The evaluation focuses on whether the proposed real-time\n                     biofeedback enables participants to perform physiotherapy exercises more accurately\n                     and in closer alignment with prescribed guidance in a home-based setting, while also\n                     assessing system usability as well as cognitive and emotional workload experienced\n                     by users.<\/p>\n                  <p>Thirty-two healthy adults (16 female, 16 male) participated in two sessions in a counterbalanced\n                     cross-over design. In the first session, participants were introduced to the rehabilitation\n                     exercises and instructed in the use of the MR-based biofeedback system. On the second\n                     session, which took place 2\u20133 days later, they performed the exercises with and without\n                     the system in a simulated home-based scenario. Of a total of 23 observed parameters,\n                     17 showed improvements, including 8 with notably strong progress. The findings demonstrate\n                     that MR biofeedback improves the accuracy of exercise execution in home-based physiotherapy.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766058\">A Systematic Mapping Study on the Joint Use of AI and VR in Stroke Care<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">David Ferrufino<\/li>\n               <li class=\"nameList\">Lauren Viado<\/li>\n               <li class=\"nameList\">Felipe Fronchetti<\/li>\n               <li class=\"nameList\">Daniel Falcao<\/li>\n               <li class=\"nameList Last\">Rodrigo Spinola<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p><em>Context<\/em>. Stroke remains a leading cause of long-term disability, prompting growing interest\n                     in emerging technologies like artificial intelligence (AI) and virtual reality (VR)\n                     to improve treatments. The combination of AI\u2019s adaptability and VR\u2019s immersive environments\n                     holds promise for personalized, engaging, and scalable stroke care, though research\n                     in this area remains fragmented. <em>Objective<\/em>. This study provides an overview of current research on the combined use of AI and\n                     VR in stroke care, focusing on system types, clinical validation, technologies employed,\n                     and autonomy levels. <em>Method<\/em>. We conducted a systematic mapping study of papers published between 2014 and 2024.\n                     <em>Results<\/em>. We identified 73 relevant studies. Most systems are still in early prototype or\n                     usability-testing stages, with limited clinical validation and frequent human oversight.\n                     Technologies used are diverse, and longitudinal evaluations are rare. <em>Conclusion<\/em>. Significant research gaps persist, including limited validation, lack of pre-stroke\n                     applications, and fragmented tools. These findings offer guidance for developing more\n                     robust, clinically viable, and interoperable AI and VR systems for stroke care.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766030\">The Impact of Sensory Levels on Presence and Cybersickness in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Hannah Champury<\/li>\n               <li class=\"nameList Last\">Sharif Mohammad Shahnewaz Ferdous<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This study investigates how varying sensory inputs\u2014visual, auditory, and tactile\u2014influence\n                     both presence and cybersickness in a virtual reality (VR) environment. Grounded in\n                     sensory conflict theory, which posits that mismatched multisensory input can cause\n                     cybersickness symptoms, we evaluated participant responses across three sensory configurations:\n                     video-only (V), video with audio (VA), and video with audio and directional wind (VAF).\n                     Fifty-six participants experienced a VR roller coaster simulation in either increasing\n                     or decreasing sensory order. Presence was assessed through self-reported realism,\n                     movement perception, and speed, while cybersickness was measured using the Simulator\n                     Sickness Questionnaire (SSQ), heart rate monitoring, Fast Motion Sickness (FMS) scale,\n                     and discomfort ratings. Results showed that increased sensory input significantly\n                     enhanced realism and speed perception, indicating improved presence. Similarly, the\n                     highest sensory condition (VAF) yielded the lowest cybersickness indicators across\n                     FMS, heart rate, and discomfort ratings. These findings suggest that carefully integrated\n                     multisensory stimuli can improve VR user experience by increasing immersion and reducing\n                     discomfort.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766031\">Ghost in the VR Shell: Capturing Spectral Cardio-Respiratory Rates from Subtle VR\n                  Device Movements<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ivo Roupa<\/li>\n               <li class=\"nameList\">Jo\u00e3o Raposo<\/li>\n               <li class=\"nameList\">Camila Abreu<\/li>\n               <li class=\"nameList\">Maria Ribeiro dos Santos<\/li>\n               <li class=\"nameList\">Pedro F. Campos<\/li>\n               <li class=\"nameList\">Carlo Massaroni<\/li>\n               <li class=\"nameList\">Hugo Pl\u00e1cido da Silva<\/li>\n               <li class=\"nameList Last\">Daniel S Lopes<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This work examines the estimation of heart rate and respiratory rate using only the\n                     kinematic data captured by a consumer-grade standalone VR devices. The high-resolution\n                     motion tracking offered by these devices creates an opportunity for indirect vital\n                     sign detection through spectral analysis of subtle VR device movement data. In our\n                     study, kinematic data were collected from a Meta Quest 3 head-mounted display, controllers\n                     and MX Ink pen across multiple posture configurations (e.g., seated, standing, lying\n                     down), both at rest and after moderate exercise. These postures emulate real-world\n                     XR scenarios for rest, fitness, and meditation. The collected data was processed using\n                     what we refer to as the <em>Ghost<\/em> approach, a simple yet effective method that applies a Fast Fourier Transform to\n                     capture the spectral components associated with respiratory and cardiac rhythms. Ground-truth\n                     biosignals were simultaneously recorded using wearable physiological sensors for validation.\n                     Results clearly reveal that both heart rate and respiratory rate can be reliably estimated\n                     from subtle micro-movements in the head-mounted display, VR controllers, or VR pen,\n                     revealing the potential for non-contact physiological monitoring within immersive\n                     environments. Finally, we demonstrate a use case of a VR stethoscope, where a standard\n                     VR controller is repurposed to estimate heart and respiratory rates.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766034\">See It and Hear It: Multimodal Guidance in MR-Based Neurosurgical Simulation for Skill\n                  Retention<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Pasquale Cascarano<\/li>\n               <li class=\"nameList\">Andrea Loretti<\/li>\n               <li class=\"nameList\">Luca Zanuttini<\/li>\n               <li class=\"nameList\">Daniele Giunchi<\/li>\n               <li class=\"nameList\">Riccardo Bovo<\/li>\n               <li class=\"nameList\">Shirin Hajahmadi<\/li>\n               <li class=\"nameList\">Giacomo Vallasciani<\/li>\n               <li class=\"nameList\">Matteo Martinoni<\/li>\n               <li class=\"nameList Last\">Gustavo Marfia<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>External Ventricular Drain (EVD) placement is a complex neurosurgical task that requires\n                     identifying a target point within the brain and accurately positioning a catheter\n                     at the appropriate angle. While Mixed Reality (MR) technologies have seen limited\n                     adoption in the operating room, they offer significant potential for developing training\n                     systems that enhance skill acquisition and retention in unaided conditions. A current\n                     gap in research concerns the effectiveness of multimodal guidance systems that incorporate\n                     both visual and audio-based MR cues. In this paper, we present an MR-based simulator\n                     for EVD placement training and evaluate the impact of three MR-guided training modalities:\n                     (1) a baseline condition using only 2D CT scans and a 2D catheter projection; (2)\n                     a visual guidance modality incorporating a 3D trajectory overlay; and (3) an embodied-audio\n                     guidance modality featuring a virtual agent delivering spoken instructions and feedback.\n                     Participants underwent a digital training phase using one of the three modalities,\n                     followed by an unaided EVD placement on a physical phantom with a real catheter to\n                     evaluate skill transfer and retention. Results indicate that both advanced MR modalities\n                     significantly improve procedural accuracy, execution speed and receive higher scores\n                     in usability and technology acceptance compared to the baseline. Notably, training\n                     with 3D visual trajectory guidance led to significantly higher unaided placement accuracy,\n                     indicating stronger skill retention. However, multimodal guidance demonstrated equivalent\n                     execution speed, while showing a trend toward lower overall cognitive load.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766046\">A Feasibility and Impact Investigation of Continuous Subjective Cybersickness Feedback\n                  Reporting<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ville Cantory<\/li>\n               <li class=\"nameList\">Jangyoon Kim<\/li>\n               <li class=\"nameList\">Courtney Hutton Pospick<\/li>\n               <li class=\"nameList\">Koorosh Vaziri<\/li>\n               <li class=\"nameList\">Evan Suma Rosenberg<\/li>\n               <li class=\"nameList Last\">Victoria Interrante<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This paper quantitatively investigates the feasibility of, and merit in, soliciting\n                     continuous subjective cybersickness ratings as participants passively engage in an\n                     immersive VR experience. The main research questions addressed are: (1) Feasibility:\n                     To what extent will participants successfully engage, unprompted, in continuous cybersickness\n                     reporting while engaging with a secondary task? and (2) Merit: To what extent do continuously\n                     reported subjective cybersickness ratings offer valuable insights beyond what can\n                     be obtained from less frequent querying?<\/p>\n                  <p>Participants used a physical slider device, in conjunction with discreet visual feedback,\n                     to continuously report their instantaneous motion sickness state as they rode nineteen\n                     consecutive rounds of a virtual roller coaster ride and performed a simple visual\n                     counting task. We analyzed the reported sickness ratings in the context of pre-post\n                     SSQ scores, #rounds endured before quitting, tonic skin conductance levels (SCL),\n                     optical flow of the visual stimulus, and rotational and translational velocity of\n                     the virtual viewpoint, as well as in comparison to previously-obtained data from different\n                     participants who underwent the same exposure but only verbally reported a single FMS\n                     score at the end of each round (every 65s).<\/p>\n                  <p>We found that most participants used the slider actively, and that, averaged across\n                     participants, the reported sickness scores not only increased over time but also varied\n                     up and down in conjunction with the intensity of the ride. We found a statistically\n                     significant positive correlation between instantaneous reported sickness levels and\n                     tonic electrodermal activity in 76% of our participants, as well as a statistically\n                     significant positive correlation with optical flow magnitude and viewpoint rotational\n                     velocity. We observed no significant differences in #rounds completed, <em>\u0394<\/em> SSQ scores, and average maximum or average last-reported sickness levels between\n                     the continuous and discrete reporting groups.<\/p>\n                  <p>Altogether, our results (1) demonstrate the feasibility of collecting valid self-reported\n                     ratings of cybersickness on a continuous basis during a passive VR experience, and\n                     (2) suggest that such data has the potential to be useful for better understanding\n                     cybersickness evolution in the context of potentially transient triggers.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 4: Interaction Design and Input Techniques II<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765992\">The Importance of Cueing While Visually Searching a 360 Degree Environment for Multiple\n                  Targets in the Presence of Distractors<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Brendan Kelley<\/li>\n               <li class=\"nameList\">Ryan P. McMahan<\/li>\n               <li class=\"nameList\">Christopher D Wickens<\/li>\n               <li class=\"nameList\">Benjamin A. Clegg<\/li>\n               <li class=\"nameList Last\">Francisco R Ortega<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Visually searching for objects is an everyday task. In many contexts, people must\n                     visually search for multiple objects at the same time while avoiding distractor objects,\n                     such as triage during a mass casualty incident. While many prior augmented reality\n                     (AR) and virtual reality (VR) studies have investigated cues to aid in visual search\n                     tasks, few have investigated cues in contexts involving multiple targets and distractors\n                     with a full 360\u00b0 effective field of regard (EFOR). Individually, multiple targets,\n                     distractors, and a full 360\u00b0 EFOR each add complexity to visual search; when combined,\n                     they compound the difficulty even further. In this paper, we present such a study\n                     that compares three common types of visual cues (2D Wedge, 3D Arrow, and Gaze Line)\n                     to a baseline condition with no cueing for a 360\u00b0 visual search task. Our results\n                     reinforce the importance of providing some type of cue, with the Gaze Line design\n                     being particularly beneficial. We discuss the potential implications of these findings\n                     for designing cues specifically for such complex visual search tasks.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765972\">Perceiving Multilingual Text in Virtual Reality: Glyph Complexity and Font Effects\n                  on Preferred Viewing Distance<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Huidan Zhang<\/li>\n               <li class=\"nameList\">Daisuke Sakamoto<\/li>\n               <li class=\"nameList Last\">Tetsuo Ono<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Differences in text perception among users of diverse language backgrounds may be\n                     accentuated in immersive environments. To investigate this, we measured preferred\n                     perceptual viewing distances in virtual reality (VR) as a function of language familiarity,\n                     glyph complexity, font weight and font type. 30 native readers each of Chinese, English,\n                     and Japanese adjusted text panels\u2014initially placed at 0.5&nbsp;m, 2.5&nbsp;m, 5&nbsp;m, and 10&nbsp;m\u2014to\n                     the distance they perceived as most appropriate for reading. Stimuli varied in visual\n                     complexity (simple vs. complex characters or words) and in font style (serif vs. sans-serif,\n                     light vs. bold). Our results show that at the farthest distances, native English and\n                     Japanese readers positioned text significantly farther away than non-natives, indicating\n                     a top-down perceptual compensation effect; this advantage was not observed for native\n                     Chinese participants. Moreover, at the closest distances, native English readers also\n                     required slightly farther viewing distances across all language conditions. Across\n                     all groups, simple glyphs and bold fonts supported greater perceptual distances, whereas\n                     complex glyphs and light fonts required closer viewing. These findings suggests that\n                     how language background and font variables shape text perception in VR and provide\n                     a theoretical basis for adaptive rendering to optimize display parameters for diverse\n                     user populations.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765979\">Assessing Redundant Interface Designs for Precise Number Input in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Pedro Miguel Matono<\/li>\n               <li class=\"nameList\">Ivo Roupa<\/li>\n               <li class=\"nameList\">Pedro Campos<\/li>\n               <li class=\"nameList Last\">Daniel Sim\u00f5es Lopes<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Typing and editing precise numerical input, in particular floating-point values, are\n                     essential tasks in Spatial Computing for applications such as 3D precision modeling,\n                     object measurement, object dimensioning, mathematical visualization, and immersive\n                     media creation. Yet the adopted interfaces and interaction techniques in VR\/AR\/MR\n                     are often replicas of flat, palm-sized numpads such as those found in physical calculators\n                     or their WIMP counterparts. To move beyond such conventional confines, this paper\n                     explores redesigning the numpad by leveraging the spatial freedom of VR with a specific\n                     focus on introducing redundancy in the input of floating point values. To do so, we\n                     took inspiration from mechanisms such as combination dials, movable numbers, and a\n                     mechanical calculator that offer a larger and multi-column number layout. We assess\n                     how redundant interfaces can enhance user experience and efficiency when it comes\n                     to precise number editing of floating point values. Through a user study (N=30), we\n                     compared four numpads where participants engage in inputting a list of target numbers\n                     within a virtual environment. Our findings reveal that the redesigned numpads, which\n                     utilize redundant design elements, were preferred by users over the conventional numpad\n                     design as they provided clearer and more efficient number input methods in VR.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765996\">SnapSteer: A Bimanual 3D Manipulation Interface with Limitable Motion Degrees of Freedom<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Yujian Fang<\/li>\n               <li class=\"nameList\">Manato Abe<\/li>\n               <li class=\"nameList\">Kazuyuki Fujita<\/li>\n               <li class=\"nameList Last\">Yoshifumi Kitamura<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We propose SnapSteer, a bimanual 3D manipulation interface using common VR controllers,\n                     which allows restriction of motion degrees of freedom (DoFs) as needed. This interface\n                     is based on the conventional one-handed 6-DoF manipulation interface called Robot\n                     Telekinesis, and assigns the other hand the role of controlling whether and in which\n                     direction the DoFs are restricted. This enables users to quickly switch between unconstrained\n                     6-DoF operation and precise 1-DoF operation according to the task. We designed and\n                     implemented a prototype of this interface in VR, and conducted a user study (N=12)\n                     comparing its performance in a straight 3D steering task with two baseline interfaces\n                     (i.e., a 6-DoF individual control interface and Robot Telekinesis). The results showed\n                     that our interface outperformed the other two in task efficiency. On the other hand,\n                     there was no significant difference in subjective workload or usability compared to\n                     Robot Telekinesis, and which derives discussion of improvements to visual feedback\n                     during the direction adjustment phase.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766025\">Beyond the Portal: Enhancing Recognition in Virtual Reality Through Multisensory Cues<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Siyeon Bak<\/li>\n               <li class=\"nameList\">Dongyun Han<\/li>\n               <li class=\"nameList\">Inho Jo<\/li>\n               <li class=\"nameList\">Sun-Jeong Kim<\/li>\n               <li class=\"nameList Last\">Isaac Cho<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>While Virtual Reality (VR) systems have become increasingly immersive, they still\n                     rely predominantly on visual input, which can constrain perceptual performance when\n                     visual information is limited. Incorporating additional sensory modalities, such as\n                     sound and scent, offers a promising strategy to enhance user experience and overcome\n                     these limitations. This paper investigates the contribution of auditory and olfactory\n                     cues in supporting perception within the portal metaphor, a VR technique that reveals\n                     remote environments through narrow, visually constrained transitions. We conducted\n                     a user study in which participants identified target scenes by selecting the correct\n                     portal among alternatives under varying sensory conditions. The results demonstrate\n                     that integrating visual, auditory, and olfactory cues significantly improved both\n                     recognition accuracy and response time. These findings highlight the potential of\n                     multisensory integration to compensate for visual constraints in VR and emphasize\n                     the value of incorporating sound and scent to enhance perception, immersion, and interaction\n                     within future VR system designs.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766048\">CONTEXT-GAD: A Context-Aware Gaze Adaptive Dwell model for Gaze-based Selections in\n                  XR Environments<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">George Ramiotis<\/li>\n               <li class=\"nameList Last\">Katerina Mania<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Gaze-based selection, via techniques such as gaze dwell, is one of the most common\n                     hands-free interaction performed by users in eXtended Reality (XR) environments. However,\n                     selecting a small constant dwell threshold to activate a target might lead to miss-interactions,\n                     also known as the Midas Touch problem, while a large threshold leads to eye fatigue.\n                     Prior research has proposed methodologies to adapt dwell thresholds based on the probability\n                     of the user activating a certain target considering past interactions or predicting\n                     intent based on gaze features. However, utilizing past inputs or gaze features leads\n                     to a heavily biased system towards individual strategy or physiology and cannot be\n                     generalized to other XR scenarios or users. In this work, we propose a novel context-aware\n                     system that leverages visual features of the task environment and user behavioral\n                     features such as the frequency of interactions, gaze speed variance, and head rotation\n                     velocity to adapt dwell thresholds across three distinct levels. We conducted a data\n                     collection experiment with twenty participants performing gaze dwell interactions\n                     in a general User Interface (UI) navigation task, and a visual search task. We trained\n                     a hierarchical machine learning model to predict and adapt dwell thresholds into three\n                     levels based on the induced cognitive load. We evaluated our system by utilizing standard\n                     machine learning metrics and by conducting a user study (n=17) based on quantitative\n                     and qualitative measures. Our system achieves a classification accuracy of <span class=\"equationTd\">\\(70.72\\%\\)<\/span><span class=\"formulaLabel\"><\/span> on the first level and <span class=\"equationTd\">\\(85.43\\%\\)<\/span><span class=\"formulaLabel\"><\/span> on the second. In addition, the system significantly reduces task completion time\n                     in less complex tasks and improves error rates in more cognitive intensive scenes.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766051\">Saccaidance: Saccade-Aware Pattern Embedding for Gaze Guidance on High-Speed Displays<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Masahiro Nara<\/li>\n               <li class=\"nameList\">Ryusuke Miyazaki<\/li>\n               <li class=\"nameList\">Yuichi Hiroi<\/li>\n               <li class=\"nameList\">Takefumi Hiraki<\/li>\n               <li class=\"nameList\">Yuta Itoh<\/li>\n               <li class=\"nameList Last\">Shio Miyafuji<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Gaze guidance is essential for directing user attention to specific areas of interest.\n                     However, conventional visual cues generate persistent visual noise that hinders concentration\n                     during tasks. We propose Saccaidance, a gaze-guidance method that appears only when\n                     users move their gaze. Saccaidance employs temporal additive color mixing and 480\n                     Hz high-speed displays to shift the color phase of guidance patterns. This renders\n                     the patterns barely visible during fixation and makes them appear transiently when\n                     users move their gaze as a color-breaking effect. This intermittent gaze guidance\n                     appears only during gaze transitions, providing effective guidance without interfering\n                     with focused work or requiring eye-tracking hardware. We conducted experiments with\n                     24 participants under four conditions that involved search tasks: an unmodified baseline,\n                     conventional explicit guidance, and our proposed method using oval and radial patterns.\n                     The results show that our approach effectively constrains the exploration area while\n                     preserving subjective naturalness. We also outline application scenarios of our method,\n                     including document highlighting.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 5: Human Factors<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765976\">Five-day research-in-the-wild observation of notifications on smartglasses: A double\n                  edged sword<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Lucas Plabst<\/li>\n               <li class=\"nameList\">Lena Plabst<\/li>\n               <li class=\"nameList\">Florian Niebling<\/li>\n               <li class=\"nameList Last\">Francisco R. Ortega<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Notifications are a fundamental aspect of daily computing, whether on desktops, laptops,\n                     smartphones, or smartwatches. On average, adults receive around 200 notifications\n                     per day\u2014approximately one every five minutes during waking hours. As Extended Reality\n                     (XR) headsets advance, they may become the primary medium for digital interactions,\n                     making notification management a crucial factor in their usability. While notifications\n                     are known to be disruptive on smartphones, their impact could be even more pronounced\n                     on head-worn devices. To investigate this, we conducted an exploratory five-day study\n                     with eight participants wearing display-equipped smartglasses that delivered notifications\n                     from their smartphones. Participants used the glasses throughout their day for at\n                     least 2 hours receiving on average 62% of all notifications on the glasses, submitted\n                     daily journal entries, and participated in post-study interviews. We also logged notification\n                     sources and timestamps throughout the study. Our findings reveal both practical advantages\n                     and significant challenges of head-worn notification delivery. While participants\n                     appreciated the convenience and immediacy of glanceable alerts, concerns about privacy,\n                     social acceptability, and distraction emerged as key barriers to adoption.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765985\">Exploring How Prior Knowledge and Presence Shape Transfer of a Reversed Size-Weight\n                  Illusion From Virtual to Real<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Seonghyeon Nam<\/li>\n               <li class=\"nameList\">Minseong Kim<\/li>\n               <li class=\"nameList\">Bogwan Kim<\/li>\n               <li class=\"nameList\">Deokyong Kim<\/li>\n               <li class=\"nameList\">Min-Ho Seo<\/li>\n               <li class=\"nameList Last\">Myungho Lee<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual Reality (VR) can create experiences that conflict with a user\u2019s prior knowledge;\n                     however, how such conflicts influence subsequent real-world behavior remains unclear.\n                     This study explores how a virtual experience that contradicts real-world expectations\n                     affects later perception and motor actions, using the size-weight illusion\u2014where people\n                     expect larger objects to be heavier than smaller ones. We conducted a 2 (internal\n                     model robustness: reinforced vs. weakened) by 2 (presence: high vs. low) mixed-design\n                     experiment. Participants first received real-world training to either strengthen or\n                     weaken their size-weight expectations, then experienced a reversed size-weight mapping\n                     in VR under varying levels of presence. We assessed how this virtual experience influenced\n                     real-world weight estimation and object lifting behavior. Results showed that participants\n                     with weakened prior knowledge exhibited lower confidence in their weight judgments\n                     and greater motor instability when lifting objects. However, the level of presence\n                     in VR did not significantly affect transfer outcomes. These findings suggest that\n                     the strength of prior knowledge modulates how conflicting virtual experiences influence\n                     real-world behavior, underscoring the need for careful VR design, particularly for\n                     younger users with less stable internal models.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765987\">AR-TMT: Investigating the Impact of Distraction Types on Attention and Behavior in\n                  AR-based Trail Making Test<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Sihun Baek<\/li>\n               <li class=\"nameList\">Zhehan Qu<\/li>\n               <li class=\"nameList Last\">Maria Gorlatova<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Despite the growing use of AR in safety-critical domains, the field lacks a systematic\n                     understanding of how different types of distraction affect user behavior in AR environments.\n                     To address this gap, we present AR-TMT, an AR adaptation of the Trail Making Test that spatially renders targets for sequential\n                     selection on the Magic Leap 2. We implemented distractions in three categories: top-down,\n                     bottom-up, and spatial distraction based on Wolfe\u2019s Guided Search model, and captured\n                     performance, gaze, motor behavior, and subjective load measures to analyze user attention\n                     and behavior. A user study with 34 participants revealed that top-down distraction\n                     degraded performance through semantic interference, while bottom-up distraction disrupted\n                     initial attentional engagement. Spatial distraction destabilized gaze behavior, leading\n                     to more scattered and less structured visual scanning patterns. We also found that\n                     performance was correlated with attention control (<em>R<\/em><sup>2<\/sup> =.20\u2013.35) under object-based distraction conditions, where distractors possessed\n                     task-relevant features. The study offers insights into distraction mechanisms and\n                     their impact on users, providing opportunities for generalization to ecologically\n                     relevant AR tasks while underscoring the need to address the unique demands of AR\n                     environments.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766032\">How a task-blind adaptive VR system can improve users&#8217; task performance: an assisted\n                  immersive analytics use case<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Simon Besga<\/li>\n               <li class=\"nameList\">Nancy Rodriguez<\/li>\n               <li class=\"nameList\">Arnaud Sallaberry<\/li>\n               <li class=\"nameList Last\">Pascal Poncelet<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Recently, some works have built adaptive systems providing assistance to the user\n                     in virtual reality (VR), with little or no knowledge of the user\u2019s task. These task-blind\n                     help systems can influence behaviours and exploration strategies; however, their ability\n                     to significantly improve users\u2019 performance on their tasks is still unclear. In this\n                     study, we aim to clarify the impact of task-blind help systems on user performance.\n                     We also explore two avenues that could provide a better understanding of why these\n                     systems can be effective and interesting to study. Our controlled user study involved\n                     56 participants in an immersive analytics environment and compared four VR help-system\n                     configurations, including three task-blind systems and a no-assistance baseline. Results\n                     showed significant task performance improvements with one task-blind system, highlighting\n                     user control as a key factor of efficiency. This work demonstrates the potential of\n                     task-blind help systems, offering a flexible framework for adaptive design and raising\n                     questions about their broader applications.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766049\">Guiding Attention in VR: Comparing the Effect of Peripheral and Central Cues on Presence\n                  and Workload<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ricardo Pinto<\/li>\n               <li class=\"nameList\">Teresa Matos<\/li>\n               <li class=\"nameList\">Daniel Mendes<\/li>\n               <li class=\"nameList Last\">Rui Rodrigues<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual Reality applications increasingly require methods to effectively guide users\n                     to important elements within the virtual environment. Central visual cues are the\n                     most common method, which have proven effective for directing attention, yet often\n                     compromise on level of immersion. This work explored whether peripheral visual cues\n                     could serve as an alternative approach that supports attention guidance while preserving\n                     sense of presence. We performed a user study with 24 participants to compare four\n                     visual cues: two central cues (Floating Text and Floating Arrow) and two peripheral\n                     cues (Edge Lighting and Swarm). Users completed a visual search task of 7 objects\n                     for each visual cue, with data collected on performance through reaction time, round\n                     time, and total errors. Additionally, presence and workload were evaluated through\n                     the IGROUP Presence Questionnaire and NASA Task Load Index, respectively. No statistically\n                     significant differences were found between peripheral and central cues for presence,\n                     however performance and workload varied significantly based on specific cue implementation\n                     rather than type of positioning. Our findings indicate that peripheral positioning\n                     does not inherently provide attention guidance advantages over central placement.\n                     Instead, thoughtful cue design, with a simple yet clear appearance and behavior appears\n                     to be the critical factor for achieving effective attention guidance while preserving\n                     presence in IVEs. These results provide valuable insights for VR content creators\n                     to facilitate the design process of VR experiences.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766053\">PatchFusionVR: Multitask Prediction of User Gaze, Reaction Time, and Cognitive Load\n                  in Virtual Reality from Multimodal Signals<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Md Irfan Pavel<\/li>\n               <li class=\"nameList\">M Rasel Mahmud<\/li>\n               <li class=\"nameList\">Jyotirmay Nag Setu<\/li>\n               <li class=\"nameList\">Kevin Desai<\/li>\n               <li class=\"nameList Last\">John Quarles<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Enhancing user experience and performance, including task load in immersive environments,\n                     requires accurate prediction of user gaze point, reaction time, and mental and physical\n                     load uptake. Current gaze prediction approaches focus primarily on motion-based information,\n                     lacking physiological data, which leads to poor prediction accuracy in highly dynamic\n                     virtual reality (VR) environments. Traditional cognitive load measurements rely on\n                     post-task analysis without proper multimodal data integration and fail to capture\n                     the real-time dynamics of user states during interaction. Likewise, reaction time\n                     or attention load are often assessed only after the interaction, without using real-time\n                     immersive sensor data, which limits adaptive responsiveness. To tackle these limitations,\n                     we leveraged a comprehensive multimodal dataset &#8211; VRWalking, which recorded timestamped\n                     eye-tracking metrics, physiological signals (heart rate and galvanic skin response),\n                     and behavioral performance data during real-time engagement in a VR environment. We\n                     developed a unified multitask model based on the MultiPatchFormer architecture, which\n                     processes multimodal VR signals through dual patch projection branches for gaze and\n                     classification inputs. The model employs multiscale patch embeddings, cross-attention\n                     between gaze and classification pathways, channel attention, and transformer encoders\n                     to jointly predict continuous user gaze and classify reaction time, cognitive load\n                     (mental load and physical load). Our methodology achieved excellent predictive performance:\n                     95.64% for reaction time, 98.01% for mental load, and 97.45% for physical load, with\n                     a MAPE (Mean Absolute Percentage Error) of 15.24% for gaze prediction. We applied\n                     Shapley Additive explanations (SHAP) analysis to interpret the model\u2019s behavior across\n                     all features, including eye-tracking, head-tracking, and physiological signals. The\n                     analysis revealed which features most influenced the predictions of user gaze, reaction\n                     time, mental load, and physical load. Our methods, while based only on the VRWalking\n                     dataset, demonstrated strong performance across all tasks, suggesting promising potential\n                     for real-world VR applications such as interactive training systems that respond to\n                     user attention lapses, educational platforms that adapt to cognitive load, and performance\n                     assessments that consider physiological indicators.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 6: Multimodal Experiences<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766028\">TangiAR: Markerless Tangible Input for Immersive Augmented Reality with Everyday Objects<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Xu Fan<\/li>\n               <li class=\"nameList\">Xincheng Huang<\/li>\n               <li class=\"nameList Last\">Robert Xiao<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Tangible interactions with everyday objects have been shown to be fast, accurate,\n                     and natural, and have shown promise when combined with immersive augmented reality.\n                     However, implementing tangible controls presents considerable challenges. Previous\n                     works in the field either rely on additional tracking markers on objects, inadvertently\n                     shifting the difficulty to users, or are too computationally demanding for real-time\n                     operation on a head-mounted display (HMD). We propose <em>TangiAR<\/em>, a tangible control system which tracks everyday objects without the need for fiducial\n                     trackers, enabling them as passive controllers and virtual proxies in AR applications.\n                     <em>TangiAR<\/em> additionally enables hand and finger proximity interactions with tangibles, further\n                     expanding the interaction space. <em>TangiAR<\/em> can run on an unmodified Microsoft HoloLens 2, making it immediately practical. We\n                     evaluated the performance of <em>TangiAR<\/em> through a technical evaluation, including occlusion robustness and tracking accuracy\n                     tests, and a user study which examined the usability of our markerless object tracking\n                     system in various AR interactions.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765988\">Manipulating Stiffness Perception of Compliant Objects While Pinching in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Yannick Weiss<\/li>\n               <li class=\"nameList\">Steeven Villa<\/li>\n               <li class=\"nameList\">Moritz Ziarko<\/li>\n               <li class=\"nameList Last\">Florian M\u00fcller<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Providing users with realistic sensations of object stiffness in virtual environments\n                     remains challenging due to the intricacies of our haptic sense. We investigate the\n                     use of a visuo-haptic illusion to alter the perceived stiffness of hand-held objects\n                     in virtual reality. We manipulate the Control-to-Display ratio of the index finger\n                     and thumb movements during pinching to make virtual objects feel softer or harder.\n                     We evaluated this approach on a variety of haptic representations and visualizations\n                     we selected through a pre-study survey (N=24). Results of our user study (N=20) demonstrate\n                     that this method effectively and reliably modifies stiffness perception, bridging\n                     gaps of 50% in physical stiffness without adversely affecting the visuo-haptic experience.\n                     Our findings offer insights into how different visual and haptic presentations impact\n                     stiffness perception, contributing to more effective and adaptable future haptic feedback\n                     systems.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765999\">Impact of passive haptics on task performance: of the effect of technological evolution<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Robin Cherix<\/li>\n               <li class=\"nameList\">Elena Mugellini<\/li>\n               <li class=\"nameList Last\">Denis Lalanne<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Since its early development in the 1990s, Virtual Reality (VR) technology, particularly\n                     head-mounted displays (HMDs), has seen significant advancements. In 1999, an empirical\n                     study demonstrated that passive haptics could significantly improve both user performance\n                     and preference in 2D tasks. In this paper, we replicate this experiment using modern\n                     VR hardware to investigate the influence of technological evolution on the relevance\n                     of passive haptics in similar scenarios. Our findings show that, for the tasks examined,\n                     performance in non-haptic conditions with current VR systems is comparable to that\n                     in haptic conditions from 1999, challenging the relevance of passive haptics for such\n                     tasks for nowadays standards. Our results imply that enhancements in visual fidelity,\n                     tracking and interaction design may have reduced the performance gap that passive\n                     haptics were previously used to address.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766050\">Understanding Latency Sensitivity in Thermal and Tactile Feedback for Multimodal Haptics\n                  in VR<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ayush Bhardwaj<\/li>\n               <li class=\"nameList\">Ashish Pratap<\/li>\n               <li class=\"nameList\">Abbas Khawaja<\/li>\n               <li class=\"nameList\">Yatharth Singhal<\/li>\n               <li class=\"nameList\">Hyunjae Gil<\/li>\n               <li class=\"nameList Last\">Jin Ryong Kim<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Low-latency multimodal feedback is essential for maintaining a high-quality user experience\n                     in VR; however, unpredictable network conditions can introduce latency that negatively\n                     impacts user experience. This work investigates how users perceive multimodal haptic\n                     feedback\u2014specifically thermal (hot\/cold) and tactile stimuli\u2014and how latency in such\n                     feedback affects user experience. We first measured users\u2019 response times for thermal,\n                     tactile, and combined thermal-tactile stimuli. Subsequently, we conducted a psychophysical\n                     study to identify delay thresholds for each modality by examining temporal congruency\n                     between visual and haptic cues. We designed a haptic delay network simulator to emulate\n                     a realistic network environment. Results highlighted that combined thermal-tactile\n                     feedback has higher latency tolerance than thermal-only feedback, indicating that\n                     multimodal integration can buffer the negative effects of latency. Using these thresholds,\n                     we designed controlled latency conditions and assessed user experience. Based on our\n                     findings, we propose design recommendations for haptic data transmission in networked\n                     VR systems.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 7: Immersive Visualization and Interaction<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765974\">Animated Transitions for Abstract and Concrete Immersive Visualizations: A Design\n                  Space and Experiment<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ambre Assor<\/li>\n               <li class=\"nameList\">Michael McGuffin<\/li>\n               <li class=\"nameList\">Arnaud Prouzeau<\/li>\n               <li class=\"nameList\">Pierre Dragicevic<\/li>\n               <li class=\"nameList Last\">Martin Hachet<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>While data visualizations are typically abstract, there is a growing body of work\n                     around concrete visualizations, which use familiar objects to convey data. Concrete\n                     visualizations can complement abstract ones, especially in immersive analytics, but\n                     it is unclear how to design smoothly animated transitions between these two kinds\n                     of representations. We investigate a design space of abstract and concrete visualizations,\n                     where animated transitions are pathways through the design space. The design space\n                     is defined with four axes, each corresponding to a different transformation. We consider\n                     different ways to design animated transitions by staging and ordering the transformations\n                     along these axes. In a controlled experiment conducted in virtual reality with 16\n                     participants, we compared four types of animated transitions and found quantitative\n                     and qualitative evidence of the superiority of a specific staging approach over the\n                     simultaneous application of all transformations. Our study pre-registration is available\n                     at https:\/\/osf.io\/8mu73.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765980\">A Low-Latency Volumetric Display and Its Application to an Augmented Reality Mirror<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Shota Tsuchiya<\/li>\n               <li class=\"nameList Last\">Shingo Kagami<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Most conventional swept volumetric displays do not allow for direct physical interaction\n                     because their rigid, high-speed sweep screens make touch dangerous or impractical.\n                     However, several existing approaches enable direct interaction with volumetric content\n                     through the use of devices such as re-imaging plates and flexible screens. A major\n                     challenge in implementing direct interaction systems lies in the high demand for latencies,\n                     but it is, in general, difficult to achieve low latencies with swept volumetric displays\n                     because their refresh rates are limited by the physical sweep periods.<\/p>\n                  <p>This paper reports yet another type of interactive swept volumetric display using\n                     a half mirror to realize a truly 3D augmented reality mirror system. It enables a\n                     &#8220;pseudo-direct&#8221; interaction by aligning the volumetric content with the user\u2019s mirror\n                     image without occlusion problems inherent in the existing direct-interactive systems.\n                     This configuration, which presents the reflected real body and the displayed content\n                     closely together without occlusion, faces a more severe demand on latencies. To address\n                     this, we propose a new low-latency control method of a swept volumetric display to\n                     make the displayed content swiftly track the mirrored target movement. The proposed\n                     method dynamically updates each slice of the volumetric content in response to the\n                     latest pose of the tracked target, without increasing the sweeping rate of the screen.<\/p>\n                  <p>Experiments demonstrate that the proposed method effectively maintains image fidelity\n                     at a moderate speed of target movement while significantly reducing perceived latency,\n                     enabling smooth and natural pseudo-direct interaction with volumetric content.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765983\">Investigating Resolution Strategies for Workspace-Occlusion in Augmented Virtuality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Nico Feld<\/li>\n               <li class=\"nameList\">Pauline Bimberg<\/li>\n               <li class=\"nameList\">Michael Feldmann<\/li>\n               <li class=\"nameList\">Matthias W\u00f6lwer<\/li>\n               <li class=\"nameList\">Eike Langbehn<\/li>\n               <li class=\"nameList\">Benjamin Weyers<\/li>\n               <li class=\"nameList Last\">Daniel Zielasko<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Augmented Virtuality integrates physical content into virtual environments, but the\n                     occlusion of physical by virtual content is a challenge. This unwanted occlusion may\n                     disrupt user interactions with physical devices and compromise safety and usability.\n                     This paper investigates two resolution strategies to address this issue: Redirected\n                     Walking, which subtly adjusts the user\u2019s movement to maintain physical-virtual alignment,\n                     and Automatic Teleport Rotation, which realigns the virtual environment during travel.\n                     A user study set in a virtual forest demonstrates that both methods effectively reduce\n                     occlusion. While in our testbed, Automatic Teleport Rotation achieves higher occlusion\n                     resolution, it is suspected to increase cybersickness compared to the less intrusive\n                     Redirected Walking approach.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765984\">Investigating Seamless Transitions Between Immersive Computational Notebooks and Embodied\n                  Data Interactions<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Sungwon In<\/li>\n               <li class=\"nameList\">Eric Krokos<\/li>\n               <li class=\"nameList\">Kirsten Whitley<\/li>\n               <li class=\"nameList\">Chris North<\/li>\n               <li class=\"nameList Last\">Yalong Yang<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>A growing interest in Immersive Analytics (IA) has led to the extension of computational\n                     notebooks (e.g., Jupyter Notebook) into an immersive environment to enhance analytical\n                     workflows. However, existing solutions rely on the WIMP (windows, icons, menus, pointer)\n                     metaphor, which remains impractical for complex data exploration. Although embodied\n                     interaction offers a more intuitive alternative, immersive computational notebooks\n                     and embodied data exploration systems are implemented as standalone tools. This separation\n                     requires analysts to invest considerable effort to transition from one environment\n                     to an entirely different one during analytical workflows. To address this, we introduce\n                     ICoN, a prototype that facilitates a seamless transition between computational notebooks\n                     and embodied data explorations within a unified, fully immersive environment. Our\n                     findings reveal that unification improves transition efficiency and intuitiveness\n                     during analytical workflows, highlighting its potential for seamless data analysis.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765993\">Interacting Beyond Reach: Multi-Perspective Augmented Reality for Precise Virtual\n                  Border Definition in Constrained Spaces<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Malte Riechmann-Thom<\/li>\n               <li class=\"nameList Last\">Jan Rexilius<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>In spatially constrained environments, such as warehouses or industrial workspaces,\n                     users often face difficulties in defining virtual regions due to occlusions, physical\n                     barriers, or limited accessibility. This paper presents a multi-perspective Augmented\n                     Reality (AR) system designed to support the precise placement of 3D virtual borders\n                     in such scenarios. The approach integrates spatially aligned remote camera perspectives\n                     into a mobile AR application, allowing users to view and interact with virtual content\n                     from otherwise unreachable positions. A loosely coupled system architecture enables\n                     dynamic integration and removal of remote cameras, ensuring scalability and adaptability\n                     to diverse setups. We evaluate the system in a user study (N=17), assess its impact\n                     on physical and cognitive workload and to analyze the usage and effect of multiple\n                     perspectives during virtual object manipulation in constrained environments. Participants\n                     reported improved spatial understanding and ease of interaction, though occasional\n                     misplacement errors occurred when relying solely on static views. These findings suggest\n                     that integrating additional perspectives into AR interfaces can effectively enhance\n                     interaction in complex and constrained environments.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766024\">Enhancing Spatial Understanding in Mixed-Reality Presentations<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Nam-Dang Vo<\/li>\n               <li class=\"nameList\">Van-Vinh Thai<\/li>\n               <li class=\"nameList\">Nam Hoai Do<\/li>\n               <li class=\"nameList\">Viet-Tham Huynh<\/li>\n               <li class=\"nameList\">Anthony Tang<\/li>\n               <li class=\"nameList Last\">Khanh-Duy Le<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p><span style=\"color:#000000\">Mixed reality (MR) presentations often involve a presenter wearing a head-mounted\n                        display (HMD) and an audience watching via a large display, making it difficult for\n                        audiences to perceive spatial relationships between the presenter and virtual objects.\n                        We report two experiments testing three design variations: (1) scene camera placement\n                        (audience-aligned vs. opposite), (2) overlaying the presenter\u2019s first-person view,\n                        and (3) highlighting objects in the presenter\u2019s view. Results show that audience-aligned\n                        cameras and object highlighting improve spatial understanding, while combining third-\n                        and first-person views can further aid perception. We derive design guidelines for\n                        configuring MR presentations to better support audience comprehension.<\/span><\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766035\">Towards Understanding how Changing Translation Gain Affects Detection Thresholds<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Md Azizul Hakim<\/li>\n               <li class=\"nameList Last\">Jerald Thomas<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Redirected Walking (RDW) enables users to explore expansive virtual environments within\n                     limited physical spaces by subtly manipulating the mapping between their physical\n                     and virtual movements. One such manipulation, translation gain, alters the scale of\n                     user\u2019s virtual forward movement relative to their physical forward movement. The primary\n                     objective of the presented study (<em>n<\/em> = 35) was to understand how changing the user\u2019s translation gain in a constant manner\n                     affects their ability to detect the manipulation. Specifically, the study presented\n                     users with three different rates of change (slow, moderate, and fast), as well as\n                     two directions (increasing and decreasing) for the applied translation gain. The study\n                     was conducted using a \u201cMethod of Limits\u201d psychometric technique, which allows for\n                     much quicker collection of the user\u2019s detection threshold when compared to other psychometric\n                     techniques used in prior RDW literature. Our results show that both rate of change\n                     and direction had a significant effect on the participants\u2019 detection thresholds,\n                     but also suggest that time of exposure to noticeable translation gain manipulations\n                     may have an impact on detection thresholds as well. Finally, we discuss these findings,\n                     their potential implications, and relevant future work.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766044\">Enhancing Immersive Virtual Reality Experiences with Multiple Tasks Prediction Using\n                  Pre-Trained Large Foundation Models<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ripan Kumar Kundu<\/li>\n               <li class=\"nameList\">Istiak Ahmed<\/li>\n               <li class=\"nameList Last\">Khaza Anuarul Hoque<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Immersive virtual reality (VR) environments pose significant cognitive and physical\n                     challenges as users engage in multitasking scenarios involving attention management\n                     and working memory, often leading to increased cognitive load, sensory conflicts,\n                     and cybersickness, diminishing users\u2019 performance and immersion. While traditional\n                     machine learning (ML) and deep learning (DL) methods have been employed to predict\n                     individual factors such as cybersickness or attention, they often fail to capture\n                     the interconnected and dynamic nature of these cognitive and physiological demands.\n                     Moreover, these methods typically require large volumes of labeled data, extended\n                     training times, and struggle to generalize across diverse VR contexts. To address\n                     these limitations, we propose an innovative method for predicting multiple tasks,\n                     i.e., cybersickness, cognitive load, working memory, and attention by leveraging the\n                     knowledge of pre-trained large foundation models, namely TimeGPT and Chronos. We apply\n                     two learning mechanisms, zero-shot and few-shot learning, for adapting these foundation\n                     models for multiple task predictions. We validate our approach on the open-source\n                     VRWalking dataset, utilizing multimodal data fusion and participant-specific grouping\n                     (based on age and gender), and compare it against traditional DL-based methods trained\n                     from scratch. Results show that our few-shot-based fine-tuned TimeGPT and Chronos\n                     models significantly outperform traditional DL models in multiple tasks. Specifically,\n                     the fine-tuned TimeGPT model achieves significantly lower RMSE values for predicting\n                     cybersickness, cognitive physical load, cognitive mental load, working memory, attention\n                     success rate, and reaction time, respectively, outperforming the traditional transformer.\n                     Furthermore, the fine-tuned TimeGPT model achieves a 4.52 \u00d7 reduction in training\n                     time compared to a conventional Transformer model for the same prediction tasks. Moreover,\n                     we deploy the fine-tuned TimeGPT model on the HTC VIVE Pro VR headset, enabling real-time\n                     prediction of multiple task severity levels from streaming VR simulation data during\n                     gameplay.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 8: Affective, Collaborative, and Social Interaction<\/h2>\n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766037\">Effects of Co-speech Gesture Size of Virtual Agents on Persuasive Communication<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Gayun Suh<\/li>\n               <li class=\"nameList\">Gun A. Lee<\/li>\n               <li class=\"nameList\">Hyung-Jeong Yang<\/li>\n               <li class=\"nameList\">Soo-Hyung Kim<\/li>\n               <li class=\"nameList\">Ji-eun Shin<\/li>\n               <li class=\"nameList\">Jaejoon Jeong<\/li>\n               <li class=\"nameList\">Sei Kang<\/li>\n               <li class=\"nameList Last\">Seungwon Kim<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Co-speech gestures are crucial for enriching both human-human and human-agent communications.\n                     Yet, the specific impacts of gesture size\u2014especially when being generated by advanced\n                     data-driven techniques\u2014remain underexplored. This study investigates how varying gesture\n                     sizes affect human-agent interactions across two distinct persuasive contexts (informational\n                     and emotional), with a focus on social outcomes such as persuasion and empathy. We\n                     conducted two controlled experiments, each involving 36 participants, comparing three\n                     gesture conditions: Minimal gesture, Small gesture, and Large gesture conditions.\n                     Experiment 1, set in an informational sales context, showed that small and large gestures\n                     significantly enhanced persuasive effectiveness, social presence, and communication\n                     quality compared to the minimal gesture condition, although no meaningful differences\n                     emerged between small and large gestures. In contrast, Experiment 2, situated in an\n                     emotionally charged context, revealed that larger gestures progressively amplified\n                     both persuasive impact and perceived empathy. These findings highlight that gesture\n                     size matters in emotionally intensive communications and the substantial social benefits\n                     of deep-learning techniques for gesture generation.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765991\">A Silent Negotiator? Cross-cultural VR Evaluation of Smart Pole Interaction Units\n                  in Dynamic Shared Spaces<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Vishal Chauhan<\/li>\n               <li class=\"nameList\">Anubhav Anubhav<\/li>\n               <li class=\"nameList\">Robin Sidhu<\/li>\n               <li class=\"nameList\">Yu Asabe<\/li>\n               <li class=\"nameList\">Kanta Tanaka<\/li>\n               <li class=\"nameList\">Chia-Ming Chang<\/li>\n               <li class=\"nameList\">Xiang Su<\/li>\n               <li class=\"nameList\">Ehsan Javanmardi<\/li>\n               <li class=\"nameList\">Takeo Igarashi<\/li>\n               <li class=\"nameList\">Alex Orsholits<\/li>\n               <li class=\"nameList\">Kantaro Fujiwara<\/li>\n               <li class=\"nameList Last\">Manabu Tsukada<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>As autonomous vehicles (AVs) enter pedestrian-centric environments, existing vehicle-mounted\n                     external human\u2013machine interfaces (eHMIs) often fall short in shared spaces due to\n                     line-of-sight limitations, inconsistent signaling, and increased decision latency\n                     on pedestrians. To address these challenges, we introduce the Smart Pole Interaction\n                     Unit (SPIU), an infrastructure-based eHMI that decouples intent signaling from vehicles\n                     and provides context-aware, elevated visual cues. We evaluate SPIU using immersive\n                     VR-AWSIM simulations in four high-risk urban scenarios: four-way intersections, autonomous\n                     mixed traffic, blindspots, and nighttime crosswalks. The experiment was developed\n                     in Japan and replicated in Norway, where forty participants engaged in 32 trials each\n                     under both SPIU-present and SPIU-absent conditions. Behavioral (response time) and\n                     subjective (acceptance scale) data were collected. Results show that SPIU significantly\n                     improves pedestrian decision-making, with reductions ranging from 40% to over 80%\n                     depending on scenario and cultural context, particularly in complex or low-visibility\n                     scenarios. Cross-cultural analyses highlight SPIU\u2019s adaptability across differing\n                     urban and social contexts. We release our open-source Smartpole-VR-AWSIM framework\n                     to support reproducibility and global advancement of infrastructure-based eHMI research\n                     through reproducible and immersive behavioral studies.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765998\">Toward Multimodal Asynchronous Collaboration in VR Artistic Creation with S.P.A.R.K<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Nathan Salin<\/li>\n               <li class=\"nameList\">Val\u00e9rie Gouranton<\/li>\n               <li class=\"nameList\">Florent Berthaut<\/li>\n               <li class=\"nameList Last\">Ronan Gaugne<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Recent artistic explorations in VR environments have explored how users engage with\n                     virtual spaces, sounds, and bodies whether, as artists, coders, or spectators. While\n                     many applications focus primarily on musical interaction or motion capture, few allow\n                     users to actively compose spatial and temporal relationships within a multimodal collaborative\n                     creation context merging altogether different artistic modalities. In this project,\n                     we present a Virtual Reality application that enables novel forms of collaboration\n                     between dancers and musicians. The application allows dancers to record full-body\n                     performances, which musicians can then use as the basis for sonic composition by spatially\n                     mapping sound triggers onto the dancer\u2019s recorded movement. Rather than relying on\n                     live capture or real-time gesture tracking, our approach treats movement as a timeline\n                     for interaction blending choreography with sound design in an asynchronous workflow.\n                     We employ an iterative design process to ensure usability among experts. This paper\n                     details the first implementation and study involving 10 participants recruited from\n                     professional and amateur artists with electronic music backgrounds, highlighting positive\n                     reception of the application\u2019s creative potential and usability.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766036\">Exploring Bichronous Collaboration in Virtual Environments<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Alexander Giovannelli<\/li>\n               <li class=\"nameList\">Shakiba Davari<\/li>\n               <li class=\"nameList\">Cherelle Connor<\/li>\n               <li class=\"nameList\">Fionn Chamberlain Murphy<\/li>\n               <li class=\"nameList\">Trey Davis<\/li>\n               <li class=\"nameList\">Haichao Miao<\/li>\n               <li class=\"nameList\">Vuthea Chheang<\/li>\n               <li class=\"nameList\">Brian Giera<\/li>\n               <li class=\"nameList\">Peer-Timo Bremer<\/li>\n               <li class=\"nameList Last\">Doug Bowman<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual environments (VEs) empower geographically distributed teams to collaborate\n                     on a shared project regardless of time. Existing research has separately investigated\n                     collaborations within these VEs at the same time (i.e., synchronous) or different\n                     times (i.e., asynchronous). In this work, we highlight the often-overlooked concept\n                     of bichronous collaboration and define it as the seamless integration of archived\n                     information during a real-time collaborative session. We revisit the time-space matrix\n                     of computer-supported cooperative work (CSCW) and reclassify the time dimension as\n                     a continuum. We describe a system that empowers collaboration across the temporal\n                     states of the time continuum within a VE during remote work. We conducted a user study\n                     using the system to discover how the bichronous temporal state impacts the user experience\n                     during a collaborative inspection. Findings indicate that the bichronous temporal\n                     state is beneficial to collaborative activities for information processing, but has\n                     drawbacks such as changed interaction and positioning behaviors in the VE.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766039\">ArithMotion: Peer-Relative Motion Generation for Social VR via Arithmetic Metaphor<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jaewoong Jang<\/li>\n               <li class=\"nameList\">Sungjae Cho<\/li>\n               <li class=\"nameList\">Yeseul Shin<\/li>\n               <li class=\"nameList Last\">Inseok Hwang<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>In social VR, bodily motions are a major nonverbal channel for expressing intent or\n                     emotion. However, freely making bodily motions is not always possible due to unaffordability\n                     of rich tracking devices, physical disabilities, or social\/spatial constraints. While\n                     current social VR platforms provide methods like emotes, expressions are limited to\n                     a finite preset. To facilitate open-ended and socially-aligned motion in constrained\n                     environments, our insight is <em>peer-relativity<\/em> found in everyday interaction. Specifically, we propose ArithMotion, an end-to-end\n                     system to generate peer-relative motions by combining generative models with arithmetic-inspired\n                     interaction. We fully implemented and iteratively refined the system. User studies\n                     show participants experienced novel, open-ended expressions closely tied to social\n                     context.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766057\">Enhancing the Audience Experience for VR and AR Theatre with AI-generated Subtitles<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Irene Viola<\/li>\n               <li class=\"nameList\">Moonisa Ahsan<\/li>\n               <li class=\"nameList\">Olga Chatzifoti<\/li>\n               <li class=\"nameList\">Atanas Yonkov<\/li>\n               <li class=\"nameList\">Eleni Oikonomou<\/li>\n               <li class=\"nameList\">Ioannis Radin<\/li>\n               <li class=\"nameList\">Pawe\u0142 M\u0105ka<\/li>\n               <li class=\"nameList\">Abderrahmane Issam<\/li>\n               <li class=\"nameList Last\">Pablo Cesar<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Recent technological developments on AI and immersive media are transforming the artistic\n                     landscape, providing novel mechanisms for artists and audiences. Following a human-centric\n                     approach, together with a theatre company in Greece, this paper investigates how subtitle\n                     placement affects user experience and cognitive load in a live theatre performance\n                     enhanced by AR glasses. To <span style=\"color:#000000\">do<\/span> so, we design and develop a system for displaying subtitles in VR and AR. We evaluated\n                     the system in two conditions (<em>N<\/em> = 19;<em>N<\/em> = 12), both in a controlled environment (VR) and an actual theatre (AR). In the latter,\n                     we integrate AI solutions to provide automatic captioning and translation in real\n                     time, <span style=\"color:#000000\">and VFX to further augment the experience<\/span>. Our quantitative and qualitative results showed no difference between subtitle placements\n                     in terms of cognitive load and user experience, with users equally liking the two\n                     proposed approaches. Results also highlighted the perceived usefulness of AR to enhance\n                     theatre performances, indicating new paths for wider accessibility and further immersion.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION 9: Avatars, Agents, and Embodiment<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765986\">The 2\u00d72 of Being Me and You: How the Combination of Self and Other Avatars and Movements\n                  Alters How We Reflect on Ourselves in VR<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Dennis Dietz<\/li>\n               <li class=\"nameList\">Samuel Benjamin Rogers<\/li>\n               <li class=\"nameList\">Julian Rasch<\/li>\n               <li class=\"nameList\">Sophia Sakel<\/li>\n               <li class=\"nameList\">Nadine Wagener<\/li>\n               <li class=\"nameList\">Andreas Martin Butz<\/li>\n               <li class=\"nameList Last\">Matthias Hoppe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Effective self-reflection is crucial for motor skill acquisition, yet it is challenging\n                     to facilitate in single-user VR training environments. We investigate this through\n                     a method where users are embodied as a virtual trainer and prompted to actively evaluate\n                     a recorded performance. In an empirical study, we systematically varied the trainee\u2019s\n                     appearance and their movements. Our mixed-methods analysis reveals that confronting\n                     one\u2019s own performance triggers a fundamental role conflict between the user\u2019s identity\n                     as the performer and their new role as the evaluator. Most importantly, this conflict\n                     challenges a binary view of embodiment. Participants experienced a multi-faceted sense\n                     of self, oscillating between identifying with the trainee and detaching as the trainer.\n                     Our work contributes a novel characterization of embodied self-evaluation, revealing\n                     a psychological duality at its core and offering clear design implications for VR\n                     systems that foster self-insight in training and therapy.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765989\">Investigating How to Control Virtual Spiders While Embodying Them in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Philipp Thayer<\/li>\n               <li class=\"nameList Last\">Martin Kocur<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual Reality (VR) enables users to embody avatars with vastly different appearances\n                     and anatomies. Embodying virtual spiders with their alien morphology could offer exciting\n                     experiences for immersive VR in gaming or education. While prior research has explored\n                     embodiment of human-like avatars and even non-human forms such as animals, it still\n                     remains unclear how best to control anatomically distinct avatars such as spiders.\n                     In this <span style=\"color:#000000\">exploratory<\/span> study, we systematically compared four control methods\u2014standard VR controller, hand\n                     control, half-body control, and full-body control\u2014while embodying a spider in VR.\n                     Using a repeated-measures design with 20 participants, we assessed each control method\n                     in terms of embodiment, usability, and perceived exertion. Results indicate that half-body\n                     control offered the best overall balance, with the highest usability and lowest exertion,\n                     while still maintaining a comparable level of embodiment to other methods. Full-body\n                     control was rated significantly lower in usability and higher in perceived exertion.\n                     These findings suggest that half-body control may provide a good balance between realism\n                     and usability for embodying spiders in VR.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766021\">VReflect: Evaluating the Impact of Perspectives, Mirrors and Avatars in Virtual Reality\n                  Movement Training<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Dennis Dietz<\/li>\n               <li class=\"nameList\">Fabian Berger<\/li>\n               <li class=\"nameList\">Changkun Ou<\/li>\n               <li class=\"nameList\">Francesco Chiossi<\/li>\n               <li class=\"nameList\">Giancarlo Graeber<\/li>\n               <li class=\"nameList\">Andreas Martin Butz<\/li>\n               <li class=\"nameList Last\">Matthias Hoppe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual reality training systems require the careful design of content presentation,\n                     user embodiment, and overall user experience. We explore the impact of different perspectives\n                     (first-person and third-person) and virtual self-visualization techniques (VSVTs:\n                     mirrors and external avatars) on user embodiment, performance and experience. In a\n                     study with 28 participants learning karate movements, we tested four combinations\n                     of these factors. Results indicate that perspective influences visual focus and embodiment,\n                     while VSVTs affect movement execution, particularly in the third-person avatar condition.\n                     Measurements of physiological activity, workload, presence, and enjoyment found no\n                     significant overall advantages for any of the conditions. Interviews revealed that\n                     most participants preferred the familiar first-person mirror combination, although\n                     participants in third-person perspective focused more on their own body and noted\n                     the helpfulness of this viewpoint. The study demonstrates that alternative perspectives\n                     and visualization techniques offer valuable training options, as these conditions\n                     did not produce significant differences in measured cognitive load when compared with\n                     each other. Future VR training systems should incorporate interactive feedback and\n                     customization options to accommodate individual preferences and optimize learning\n                     experiences.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766022\">Joining the Circle: Human Entry Behavior in a Mixed Reality F-Formation with Agent,\n                  Avatar, and Human Partners<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Junyeong Kum<\/li>\n               <li class=\"nameList\">Sunghun Jung<\/li>\n               <li class=\"nameList\">Hyeongil Nam<\/li>\n               <li class=\"nameList\">Kangsoo Kim<\/li>\n               <li class=\"nameList Last\">Myungho Lee<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>According to Hall\u2019s theory, the space individuals maintain between one another depends\n                     on relational closeness and situational context. Prior research suggests that interpersonal\n                     distance (IPD) varies not only between virtual and real humans, but also among virtual\n                     humans depending on their perceived agency. However, little is known about how people\n                     spatially negotiate entry into mixed groups comprising different types of agents in\n                     extended reality (XR) settings. In this study, we examine participants\u2019 entry behavior\n                     as they join a circular F-formation composed of three distinct entities: an agent,\n                     an avatar, and a real human. Specifically, we investigate how participants position\n                     themselves relative to each entity, analyzing their preferences and behaviors in terms\n                     of IPD and entry dynamics. Our findings reveal that participants maintained the greatest\n                     IPD from the real human, followed by the avatar and the agent, suggesting nuanced\n                     social distinctions among these three entities. Furthermore, when the real human was\n                     absent, participants tended to maintain a greater distance from the avatar compared\n                     to the agent. These results offer valuable insights for the design of XR collaboration\n                     environments and for understanding social dynamics in multi-agent interactions.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766029\">How Avatar User Visual Incongruities Impact the Sense of Embodiment in Virtual Reality:\n                  A Systematic Review<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Dion Deng<\/li>\n               <li class=\"nameList\">Mila Buji\u0107<\/li>\n               <li class=\"nameList Last\">Juho Hamari<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Virtual reality (VR) is considered a technological megatrend that is driving the digitization\n                     of all aspects of human life. Avatars, virtual bodies controlled by users, play an\n                     important role in VR. Many scholars consider the sense of embodiment as a key affordance\n                     of avatars. However, our understanding of how to optimize embodiment through avatar\n                     representation in VR remains underdeveloped. This study systematically reviewed a\n                     body of 43 studies from 41 research papers that investigated the manipulations of\n                     avatars in VR. The corpus was coded with head-mounted display (HMD) models, avatar\n                     creation tools, mirror use, task context, and avatar manipulations. Based on these\n                     experiment-based studies, we discuss how different types of avatar representations\n                     affect users\u2019 embodiment. Based on the findings, we indicate practical implications\n                     of avatar design in VR applications.<\/p>\n                  <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766041\">I feel you: The Impact of Emotional Virtual Characters on Emotional State, Player\n                  Experience, and Connectedness in VR Games<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Linda Graf<\/li>\n               <li class=\"nameList\">Gian Luca Dossena<\/li>\n               <li class=\"nameList Last\">Maic Masuch<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Engaging in digital games, in particular, using highly immersive VR technology, has\n                     emerged as a coping mechanism to escape negative feelings and create positive emotions.\n                     It is, therefore, interesting to investigate which mechanisms in digital games influence\n                     players\u2019 emotions. One potential mechanism is using virtual characters as companions,\n                     as they can positively affect the emotional state and the player experience. However,\n                     the impact of the virtual character\u2019s specific emotions on the players has been studied\n                     less. The presence of others can positively influence people and improve their well-being\n                     when in a bad emotional state. According to the <em>Emotional Contagion<\/em> theory, the presence of positive-minded others can achieve these positive effects\n                     or, following the <em>Emotional Similarity<\/em> theory, also by the presence of negative-minded others. Therefore, we want to investigate\n                     how virtual characters in different emotional states affect the players when they\n                     are sad and immerse themselves in a game. Hence, we posed the following research question:\n                     &#8220;<em>How do different emotions of a virtual character affect emotional state, player experience,\n                        and connectedness when players are in a sad emotional state?<\/em>&#8220;. Our lab study with 75 participants put in a sad emotional state, revealed significant\n                     differences between playing a cooperative VR game with either a happy or a sad virtual\n                     character regarding participants\u2019 emotional state and connectedness, but not regarding\n                     player experience. We discuss the results and implications of emotional virtual characters\n                     to enhance well-being.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766054\">Co-embodied Mirroring: Investigating the Effects of Movement Blending on Partner Impressions\n                  in Virtual Environments<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Yusuke Koseki<\/li>\n               <li class=\"nameList\">Kizashi Nakano<\/li>\n               <li class=\"nameList\">Takuji Narumi<\/li>\n               <li class=\"nameList\">Hideaki Kuzuoka<\/li>\n               <li class=\"nameList Last\">Tomohiro Amemiya<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>The mirroring effect has been demonstrated to facilitate smooth social interactions.\n                     Digital technology can automatically implement mirroring to achieve more natural interactions\n                     in contexts where spontaneous mirroring is challenging, such as remote human communication\n                     or dialogue with non-human agents. However, many existing automatic mirroring systems\n                     replicate a user\u2019s movements with a time delay, and when applied to full-body interactions\n                     between human users, such systems often result in excessive and unnatural mimicry.\n                     To address this issue, this study examines the effectiveness of Co-embodiment, a technique\n                     that blends the movements of a user and their partner in real-time, as a method for\n                     achieving both the benefits of mimicry and the naturalness of interaction. We examined\n                     how varying the blending ratios (0%, 25%, 50%) affects social impressions in a two-person\n                     interaction in the VE. Results showed that moderate blending (25%, 50%) enabled natural\n                     and comfortable interactions, comparable to no blending (0%). However, we did not\n                     observe significant improvements in perceived closeness or trustworthiness. Semi-structured\n                     interviews helped explain these results, revealing that participants differed in their\n                     interpretations of subtle synchronized movements. Some viewed them as signs of mutual\n                     understanding or cooperation, while others experienced discomfort. In some cases,\n                     this discomfort seemed specific to virtual contexts, as subtle similarities made participants\n                     doubt the presence of a real human behind the avatar. These findings highlight that\n                     impressions depend on whether blended movements feel socially responsive or merely\n                     imitative, stressing the need for interaction designs that enhance the sense of social\n                     presence.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765995\">Compensating Motion-Induced Errors in Smartphone-Based VR Avatar Reconstruction<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Friedemann Runte<\/li>\n               <li class=\"nameList\">Timo Menzel<\/li>\n               <li class=\"nameList\">Ulrich Schwanecke<\/li>\n               <li class=\"nameList Last\">Mario Botsch<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Recent developments in smartphone-based avatar reconstruction have made the creation\n                     of personalized and realistic avatars significantly more accessible. However, relying\n                     on one smartphone camera leads to capturing images sequentially, which introduces\n                     new challenges; particularly longer capture times increase the susceptibility to subject\n                     motion, which results in degraded reconstructions.<\/p>\n                  <p>We present a novel approach for smartphone-based avatar reconstruction that combines\n                     photogrammetry, silhouette constraints, and inverse rendering to produce high-fidelity,\n                     realistic avatars free of motion-induced artifacts. By using short, motion-resilient\n                     image sequences, referred to as <em>sub-scans<\/em>, we considerably reduce motion-induced artifacts. Our pipeline achieves high visual\n                     quality while offering improved robustness and outperforms current state-of-the-art\n                     methods in terms of computation time and accuracy.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            <h2>SESSION 10: Security and Systems<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766012\">Synthesizing Evidence-Based AR Design Recommendations and Identifying Gaps in Practice<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Md Mehedi Hasan Jibon<\/li>\n               <li class=\"nameList\">Ngu Quoc Truong<\/li>\n               <li class=\"nameList\">Tanzila Roushan Milky<\/li>\n               <li class=\"nameList\">Felicia Rose Drysdale<\/li>\n               <li class=\"nameList Last\">Julia Woodward<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>From handheld devices to head-mounted displays, augmented reality (AR) technologies\n                     are becoming commonplace in everyday settings, supporting tasks in education, healthcare,\n                     gaming, and beyond. Prior research has developed a number of evidence-based design\n                     recommendations for AR apps. However, these recommendations are often scattered across\n                     academic literature and differ in scope and focus. In addition, there are still open\n                     research questions about the degree to which existing guidelines are applied in practice,\n                     particularly in handheld AR contexts. To address these gaps, we synthesized AR design\n                     recommendations from academic literature and organized them into an integrated set\n                     of guidelines. We then empirically analyzed 52 commercial handheld AR apps to assess\n                     how well they align with these guidelines. We found that while most apps follow basic\n                     usability guidelines, such as using familiar UI layouts, many apps do not adopt context-aware\n                     features, offer limited support for multimodal interaction and feedback, and overlook\n                     key usability practices such as onboarding and navigational aids. In addition, we\n                     saw very few guidelines related to data privacy, collaborative AR, safety and accessibility.\n                     We contribute a synthesis of evidence-based AR recommendations and identify key areas\n                     of disconnect between recommendations and practice for handheld AR apps, which aids\n                     future designers and developers.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766045\">Beyond the Headset: A Systematization of Knowledge on Extended Reality Privacy and\n                  Security in Healthcare<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Nafisa Anjum<\/li>\n               <li class=\"nameList Last\">M Rasel Mahmud<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Extended reality (XR) systems offer transformative potential for healthcare in domains\n                     ranging from surgical planning to remote rehabilitation and mental\u2010health therapy.\n                     The rich streams of sensor, biometric, and environmental data that enable these applications,\n                     however, also create novel and poorly understood privacy and security vulnerabilities:\n                     adversaries can exploit unencrypted signaling, sensor side\u2010channels, and application\u2010layer\n                     flaws to infer sensitive patient information or disrupt clinical workflows. Nevertheless,\n                     there aren\u2019t many thorough Systematization of Knowledge (SoK) that examine XR for\n                     healthcare at the moment. In this SoK, we survey 65 peer\u2010reviewed works published\n                     between 2017 and 2024 across leading XR, security, and privacy venues, synthesizing\n                     a unified threat taxonomy that spans device, network, user and cloud layers. We introduce\n                     a quantitative evaluation framework <em>XR-PRISM<\/em> (Privacy and Risk Impact Scoring Metric), drawing on adapted risk scores, detection\n                     performance, and usability assessments to rigorously assess the level of security\n                     and privacy risks. Our analysis reveals critical gaps: over 70% of countermeasures\n                     lack standardized risk evaluations, fewer than 15% include high prerequisites to launch\n                     an attack, and reproducibility is hampered by scarce artifact releases. Finally, we\n                     chart a research roadmap advocating for open benchmark suites with shared datasets,\n                     artifact disclosure policies, cloud\u2010layer protections, and robust detection and recovery\n                     mechanisms. By quantifying \u201cwhat works\u2014and by how much,\u201d this SoK provides a data\u2010driven\n                     foundation for developing secure, privacy\u2010preserving, and usable XR healthcare technologies.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765978\">User Identification in Virtual Reality through Behavioral Biometrics and the Influence\n                  of Colocated Interactions<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jonathan Liebers<\/li>\n               <li class=\"nameList\">Frieder Sykora<\/li>\n               <li class=\"nameList\">Niklas Pf\u00fctzenreuter<\/li>\n               <li class=\"nameList\">Uwe Gruenefeld<\/li>\n               <li class=\"nameList\">David Goedicke<\/li>\n               <li class=\"nameList Last\">Stefan Schneegass<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Behavioral Biometrics in Virtual Reality (VR) allow for implicit user identification,\n                     as the head- and hand-movements that can be captured from the head-mounted display\n                     and the controllers are highly descriptive of the user\u2019s true identity. Such body\n                     movements have been explored in the past; however, to date, it is unclear how they\n                     perform in settings where more than one person interacts in a shared virtual environment.\n                     In this work, we explored through a user study (N=40) how behavioral biometrics in\n                     VR change when one or more persons interact with each other in a shared virtual environment\n                     and whether this is influenced by the nature of the interaction itself. We find that\n                     user identification is possible with up to 83.38&nbsp;% by applying deep learning models,\n                     and that particularly cooperative interactions between multiple VR users lead to highly\n                     identifiable body movements. Our results help in advancing behavioral biometrics for\n                     seamless user identification in VR, as a viable alternative to using PINs and passwords.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766052\">Motion Forecasting Attacks on Behavioral Biometric Authentication Systems in Virtual\n                  Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Mingjun Li<\/li>\n               <li class=\"nameList\">Ashutosh Shivakumar<\/li>\n               <li class=\"nameList\">Natasha Kholgade Banerjee<\/li>\n               <li class=\"nameList Last\">Sean Banerjee<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Inspired by behavioral biometrics for keystroke and touch-based systems, a large body\n                     of work has emerged over the past decade on using user behavior in VR applications\n                     as a signature of the genuine user. Recent work on forecasting approaches for behavioral\n                     biometrics for VR helps address a key challenge in existing approaches where complete\n                     user movement signatures are needed to authenticate the user. Forecasting-based approaches\n                     enable VR authentication systems to use limited user behavior data and forecast future\n                     movement trajectories. However, forecasting-based approaches present a new concern\n                     where malicious users can exploit the predictability of user motions to launch an\n                     attack. In this paper, we present the first forecasting-based attack model against\n                     VR authentication systems that rely on behavioral biometrics. We propose a two-phase\n                     approach to assess authentication performance and adversarial risk. Phase 1 develops\n                     a Fully Convolutional Network for authentication using VR motion data, evaluating\n                     stochastic gradient descent (SGD) and Adam optimizers with Equal Error Rate (EER)\n                     as the primary metric. Phase 2 introduces a forecasting attack, where partial motion\n                     sequences of an impostor\u2019s motion are fed to a Transformer model to generate future\n                     trajectories that represent genuine user behavior for an authenticator enabling an\n                     impostor to deceive the authentication system. Experimental results demonstrate the\n                     attack\u2019s effectiveness, achieving an EER as low as 0.0346, exposing security risks\n                     in motion-based authentication. These findings underscore the urgent need for robust\n                     countermeasures to defend against predictive motion attacks in VR environments. Our\n                     code is shared at: https:\/\/bit.ly\/4n1GtxG.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3766056\">Not Just Who You Are, but Where and How: Modeling XR Authentication Scenarios<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Christina Katsini<\/li>\n               <li class=\"nameList\">Gregory Epiphaniou<\/li>\n               <li class=\"nameList Last\">Carsten Maple<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Authentication in extended reality (XR) presents unique challenges due to embodied\n                     interaction, spatial immersion, and variable environmental conditions. As XR systems\n                     become more prevalent, secure and usable authentication mechanisms are critical. However,\n                     current research often overlooks the scenarios in which these mechanisms operate,\n                     limiting comparability, reproducibility, and real-world applicability. This paper\n                     addresses this gap by presenting a structured model of XR authentication scenarios.\n                     We conducted semi-structured interviews with experts in the Usable Security and Privacy\n                     domain to identify key scenario dimensions influencing the design and evaluation of\n                     XR authentication mechanisms. Through thematic analysis, we identified dimensions\n                     related to contextual parameters, environmental conditions, and XR-specific properties.\n                     The resulting scenario model was validated through literature mapping and demonstrated\n                     via a realistic use case. Our work provides a foundation for context-aware design\n                     and more rigorous evaluation of authentication mechanisms across diverse XR environments.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765973\">Predictability-Aware Motion Prediction for Edge XR via High-Order Error-State Kalman\n                  Filtering<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ziyu Zhong<\/li>\n               <li class=\"nameList\">Bj\u00f6rn Landfeldt<\/li>\n               <li class=\"nameList\">G\u00fcnter Alce<\/li>\n               <li class=\"nameList Last\">H\u00e9ctor Caltenco<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>As 6G networks evolve, offloading extended reality (XR) applications emerges as a\n                     key use case, leveraging reduced latency and edge processing to migrate computationally\n                     intensive tasks, such as rendering, from user devices to the network. This enables\n                     lower battery consumption and smaller device form factors in cellular environments.<\/p>\n                  <p>However, offloading incurs delays from network transmission and edge server queuing,\n                     particularly under multi-user concurrency, resulting in elevated motion-to-photon\n                     (MTP) latency that degrades user experience. Motion prediction techniques, including\n                     deep learning and Kalman filter (KF), have been proposed to compensate, but deep learning\n                     struggles with scalability at resource-constrained edges amid growing user loads,\n                     while traditional KF exhibits vulnerability in handling complex motions and packet\n                     loss in 6G\u2019s high-frequency interfaces.<\/p>\n                  <p>To address these challenges, we introduce a context-aware error-state Kalman filter\n                     (ESKF) framework for forecasting user head motion trajectories in remote XR, integrating\n                     a motion classifier that categorizes movements by predictability to minimize prediction\n                     errors across classes. Our results show that this optimized ESKF outperforms conventional\n                     Kalman filters in positional and orientational accuracy, while demonstrating superior\n                     robustness and resilience to packet loss.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3765981\">When AR Hinders Performance: The Hidden Costs of Video-See-Through Displays<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Hamraz Javaheri<\/li>\n               <li class=\"nameList\">Vitor Fortes Rey<\/li>\n               <li class=\"nameList\">David Kariem Habusch<\/li>\n               <li class=\"nameList\">Jakob Karolus<\/li>\n               <li class=\"nameList Last\">Paul Lukowicz<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Head-mounted displays (HMDs) are increasingly used in safety-critical fields such\n                     as surgery, aviation, and industrial manufacturing. As major manufacturers shift toward\n                     video-see through (VST) designs to deliver unified AR and VR experiences, they also\n                     replace direct visual access to the real world with a video feed. This design choice\n                     raises concerns about its impact on user performance. This study investigates the\n                     isolated impacts of VST and optical-see through HMDs on user real-world perceptual\u2013motor\n                     performance by comparing two leading HMDs, Apple Vision Pro (AVP) and HoloLens 2 (MHL2)\n                     against unencumbered vision using the Purdue Pegboard Test (PPT), a standard assessment\n                     of manual dexterity. Twenty participants completed tasks across three conditions (AVP,\n                     MHL2, and Baseline), while we recorded dexterity scores, cognitive load, system usability,\n                     VR sickness, and subjective feedback. Movement data were also collected via Apple\n                     Watches. Study results with 20 participants revealed that dexterity scores significantly\n                     declined under the AVP condition across all subtests. This was accompanied by significantly\n                     higher cognitive load and a notable drop in RMS acceleration values (observed in the\n                     RMS analysis of a subset of 13 participants). The analysis on dexterity score yielded\n                     a significant difference between MHL2 and Baseline only for a single subtest of the\n                     PPT (Left Hand). Post-task interviews revealed greater discomfort, visual fatigue,\n                     and reduced task confidence with AVP. These findings suggest that current VST HMDs\n                     impose a hidden ergonomic cost undermining user performance in tasks where precision,\n                     and comfort are essential. For AR applications designed to enhance user performance,\n                     such as assistive tools, training systems, or task guidance interfaces, designers\n                     must account for and mitigate this performance degradation through counterbalancing\n                     strategies to offset the visual and cognitive burden introduced by VST HMDs.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: VRST 2025 Poster Abstracts: Interaction Design and Input Techniques I<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770550\">VR Eye Tracking Data for Gender Identification: A Look at Same-Domain and Cross-Domain\n                  Scenarios<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Sarker Monojit Asish<\/li>\n               <li class=\"nameList Last\">Arijet Sarker<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Prior research has shown that cross-domain gender identification (GI) in VR is challenging,\n                     often due to limited overlapping features and a lack of shared users across datasets.\n                     In this work, we examine two distinct VR environments\u2014a solar panel task and a biological\n                     exploration task\u2014using a consistent feature set and eye-tracking (ET) data from common\n                     users. Our results confirm that cross-domain classification is substantially harder\n                     than domain-specific tasks and highlight head position as a key feature. Importantly,\n                     we show that incorporating common users improves model performance, emphasizing the\n                     role of user overlap in enhancing the generalizability of GI models in VR.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h2>SESSION: Locomotion and Wayfinding<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770542\">Visual Constraints Impact on Steering in VR Driving Simulation<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Obaida Alrazouk<\/li>\n               <li class=\"nameList\">Hichem Arioui<\/li>\n               <li class=\"nameList Last\">Amine Chellali<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Human vision guides lane keeping and hazard anticipation during driving. However,\n                     isolating how visual field constraints affect steering is difficult in real driving.\n                     This study used immersive VR with a depth-aligned aperture to restrict vision while\n                     participants drove curved roads under several conditions. Results revealed that misaligned\n                     restrictions impaired steering, while tangent point alignment partly improved performance.\n                     Results highlight how VR can probe visual\u2013motor mechanisms in driving.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768398\">Detection of Translation Gain is Decreased When Virtual Reality Users Are Unaware\n                  of Its Presence<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Brett Benda<\/li>\n               <li class=\"nameList\">Jennifer Cieliesz Cremer<\/li>\n               <li class=\"nameList\">John Fang-Wu<\/li>\n               <li class=\"nameList Last\">Eric Ragan<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>The prevalent evaluation methods used to estimate detection of redirected walking\n                     are based on methods from psychophysics that require users to know their virtual movements\n                     are being manipulated. However, this higher-than-normal level of attention toward\n                     their movements yields conservative detection thresholds. We find that participants\n                     who were unaware that redirected walking (translation gain) was applied detected the\n                     technique at a significantly higher gain than users who were aware (at gains of 1.73\n                     and 1.38, respectively). We provide evidence that redirected walking-based navigation\n                     solutions may be able to leverage gain values that are larger than the current threshold\n                     guidelines would suggest.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770543\">Visualizing Time-Dependent Navigation Zones in Mixed Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Tim Waechter<\/li>\n               <li class=\"nameList\">Matthias Koenig<\/li>\n               <li class=\"nameList Last\">Krzysztof Tarasiuk<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Using Mixed Reality (MR) to plan the motion of robots, for example the path of a drone,\n                     is an effective way to optimize their movements in a collaborative working environments.\n                     Path planning using MR allows for the incorporation of motion constraints arising\n                     from real-world obstacles, virtual objects, and time scheduling considerations.<\/p>\n                  <p>We address time-dependent path planning of a drone in a three-dimensional dynamic\n                     indoor environment using MR. We evaluate visualization methods which convey the planned\n                     path of the drone and which show time-dependent safety zones around real and virtual\n                     obstacles. Our results with 51 participants show the effectiveness of the visualization\n                     to support interactive path planning tasks. The contribution of our research is a\n                     novel MR visualization method for three-dimensional and time-dependent (4D) path planning.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770549\">Vibrotactile Feedback to Make Real Walking in Virtual Reality More Accessible for\n                  People With and Without Mobility Impairments<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">M. Rasel Mahmud<\/li>\n               <li class=\"nameList\">Michael Stewart<\/li>\n               <li class=\"nameList\">Alberto Cordova<\/li>\n               <li class=\"nameList Last\">John Quarles<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This research aims to examine the effects of various vibrotactile feedback techniques\n                     on gait (i.e., walking patterns) in virtual reality (VR). Prior studies have demonstrated\n                     that gait disturbances in VR users are significant usability barriers. However, adequate\n                     research has not been performed to address this problem. In our study, 39 participants\n                     (with mobility impairments: 18, without mobility impairments: 21) performed timed\n                     walking tasks in a real-world environment and identical activities in a VR environment\n                     with different forms of vibrotactile feedback (spatial, static, and rhythmic). Within-group\n                     results revealed that each form of vibrotactile feedback improved gait performance\n                     in VR significantly compared to the no vibrotactile condition in VR for individuals\n                     with and without mobility impairments. Moreover, spatial vibrotactile feedback increased\n                     gait performance significantly in both participant groups compared to other vibrotactile\n                     conditions.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770536\">Seeing With Sound in Safe Virtual Environments: A Walk-In-Place VR Training System\n                  for Users With Visual Impairment Using the vOICe Algorithm<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Myung Jin (MJ) Kim<\/li>\n               <li class=\"nameList\">Mooseop Kim<\/li>\n               <li class=\"nameList\">HyeonBeom Yi<\/li>\n               <li class=\"nameList Last\">Chi Yoon Jeong<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We present a virtual reality (VR) training system that supports safe mobility skill\n                     development for low-vision users through visual-to-auditory sensory substitution.\n                     The system combines the vOICe algorithm with walk-in-place locomotion to enable navigation\n                     in immersive environments while minimizing physical risks and spatial requirements.\n                     Training with the system follows a two-phase structure: an initial learning phase\n                     to build familiarity with visual-to-audio substitution, followed by a navigation phase\n                     in which users apply auditory cues to explore and reach destinations in VR. The system\n                     provides a safe, controlled environment for developing non-visual spatial awareness\n                     and serves as an early exploration of a platform for evaluating sensory substitution\n                     techniques. Through this work, we aim to contribute to solutions that promote greater\n                     independence in mobility for visually impaired users.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Cybersickness, Health, and Digital Twins<\/h2>\n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770546\">Meltdown: Bridging the Perception Gap in Sustainable Food Behaviors Through Immersive\n                  VR<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Melissa Anastasia Harijanto<\/li>\n               <li class=\"nameList\">Florentiana Yuwono<\/li>\n               <li class=\"nameList\">Yi Xu<\/li>\n               <li class=\"nameList\">Xiao Xuan Chong<\/li>\n               <li class=\"nameList\">Peisen Xu<\/li>\n               <li class=\"nameList Last\">Anand Bhojan<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Climate change education often struggles to connect personal actions with environmental\n                     consequences. Meltdown is an immersive VR escape room that teaches sustainable food\n                     consumption and waste practices through scenario-based tasks and consequence-driven\n                     feedback. A user study (<em>N<\/em> = 36) found significant gains in familiarity, confidence, and behavioral intentions,\n                     with modest knowledge improvements. Exploratory metrics (<em>n<\/em> = 13) showed high accuracy on familiar decisions but lower accuracy on less intuitive\n                     ones. These findings suggest that consequence-driven VR can effectively engage learners,\n                     link everyday choices to visible outcomes, and foster sustainable behavior change.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770552\">Accessible VR for Older Adults: Mounting without Straps<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Martine Bordeleau<\/li>\n               <li class=\"nameList\">Noureddine Lourimi<\/li>\n               <li class=\"nameList\">Guillaume L\u00e9onard<\/li>\n               <li class=\"nameList\">Maxime Robert<\/li>\n               <li class=\"nameList\">S\u00e9bastien Gaboury<\/li>\n               <li class=\"nameList Last\">Pascal E. Fortin<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Older adults face ergonomic barriers in virtual reality (VR), particularly with head-mounted\n                     displays (HMDs), which are heavy and poorly suited for fragile musculoskeletal systems.\n                     This poster introduces a 3D-printed support system that alleviates the weight of the\n                     headset by redistributing it to an external frame. This solution promotes accessibility\n                     by allowing passive, seated VR experiences, enabling better comfort and extending\n                     the use of VR for seniors in health, leisure, and telepresence contexts.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770541\">Background Sound Tempo Modulation Can Influence Scene-Specific Memory in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Hyuma Auchi<\/li>\n               <li class=\"nameList\">Shogo Fukushima<\/li>\n               <li class=\"nameList\">yuki fujita<\/li>\n               <li class=\"nameList Last\">Keiichi Zempo<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Sustaining user memory in digital environments such as virtual reality (VR) is a significant\n                     challenge. We show that temporary tempo modulations in background music (BGM) can\n                     selectively and naturally enhance memory in VR. In a user study (N&nbsp;=&nbsp;20), decreasing\n                     the BGM tempo by approximately 21% significantly improved recall of events. These\n                     findings point to a new acoustic design approach that adapts scene by scene to narrative\n                     pacing and importance while maintaining a natural user experience.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768400\">Varying Ecological Validity of the Virtual Environment Influences Soccer Pass Reaction\n                  Times<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Wendy E. Huddleston<\/li>\n               <li class=\"nameList\">Alexander J. Hofer<\/li>\n               <li class=\"nameList\">Sawyer S. Ladd<\/li>\n               <li class=\"nameList\">Caleb C. Krage<\/li>\n               <li class=\"nameList Last\">Jerald Thomas<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\"><\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768409\">Keeping It Clean: A VR Simulation for Dental Sterilization<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">John W Peters<\/li>\n               <li class=\"nameList\">Raymond J Perry<\/li>\n               <li class=\"nameList\">Katherine Kireburns<\/li>\n               <li class=\"nameList Last\">Paul D. S. Fink<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This paper explores the applicability of virtual reality (VR) to dental education,\n                     specifically focused on learning the routine but critical process of sterilizing dental\n                     equipment. We present a novel sterilization training prototype and evaluate the extent\n                     to which educators and practicing hygienists would implement VR training tools in\n                     the future. A survey was conducted with (N=28) dental providers and a subsequent in-person\n                     evaluation of the prototype was conducted with (N=7) educators in a dental hygienist\n                     training program. Results suggest overall positive opinions of VR for dentistry, while\n                     also offering suggestions for improving the prototype.<\/p>\n                  <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Interaction Design and Input Techniques II<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768407\">Interior design method in AR based on AI with a gesture modification<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList Last\">Ma\u0142gorzata Telesi\u0144ska<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Computer-aided design has a long history. Such techniques are founded on the construction\n                     of a three-dimensional model using Building Information Modelling software or dedicated\n                     software for generating visualisations. These processes can be supported by modern\n                     inventory techniques and the visual effects can be processed graphically and exported\n                     to various environments, including virtual reality environments. The integration of\n                     AI-powered model generation into the design process can facilitate the advancement\n                     of spatial modeling techniques. Furthermore, the incorporation of technology based\n                     on interaction will enhance existing architectural design methods. We propose a solution\n                     that enables the generation of 3D models through voice interaction, with the possibility\n                     of modification through gestures.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            <h2>SESSION: Human Factors and Perception<\/h2>\n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770538\">Depth-Shifting Aerial Image Display Using Angle Changes Between the Display and Optical\n                  Elements<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ayami Hoshi<\/li>\n               <li class=\"nameList\">Motohiro Makiguchi<\/li>\n               <li class=\"nameList\">Ayaka Sano<\/li>\n               <li class=\"nameList Last\">Naoto Abe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We propose a new optical system that creates three-dimensionally movable aerial images\n                     with a minimal optical configuration. The proposed optical system achieves depth shift\n                     of the aerial image by adjusting the angle of each element. Experimental results showed\n                     that the proposed optical system can reproduce depth and can improve and uniform the\n                     luminance of the aerial image compared to the conventional linear movement method.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768406\">An Experimental Study of Tilt Sensation Displayed with a Combination of Visual and\n                  Physical Tilt: A Case of an Interactive VR Work &#8220;Gravity Paradox&#8221;<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Shunsuke Otani<\/li>\n               <li class=\"nameList\">Ryosuke Ogura<\/li>\n               <li class=\"nameList\">Koichi Takeshima<\/li>\n               <li class=\"nameList\">Daichi Kitagawa<\/li>\n               <li class=\"nameList Last\">Keita Ushida<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>The authors developed an interactive VR work. In this work, the player&#8217;s goal is to\n                     escape from a mansion by changing the direction of gravity. This gravity change is\n                     represented by visual and physical tilt. The visual tilt is displayed through an HMD,\n                     and the physical tilt is presented through a motion platform. This presentation was\n                     effective, and players enjoyed the work at exhibitions. In this paper, the authors\n                     experimented to study the effect of adding physical tilt to visual tilt, especially\n                     its role in sensation. Participants were presented with different tilt stimuli (images\n                     with varying tilt angles and the presence or absence of physical tilt) and asked to\n                     report their perceived tilt angles. The participants felt more tilted when physical\n                     tilt was presented. Moreover, though the perceived tilt angle seemed to depend on\n                     the history of the visual tilt angle, this phenomenon need not be considered when\n                     simple tilt emphasis is expected.<\/p>\n                  <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770537\">An Experiment on a High-Speed Image Projection Perceived Only During Smooth Pursuit\n                  using Striped Patterns<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Ryusuke Miyazaki<\/li>\n               <li class=\"nameList\">Shio Miyafuji<\/li>\n               <li class=\"nameList Last\">Hideki Koike<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We have proposed a novel image display technique that uses high-speed projection to\n                     present different images depending on the user\u2019s gaze movement direction. This method\n                     projects stripe-based frames at high refresh rates; when a user moves their gaze in\n                     a specific direction, previously hidden visual content becomes perceptible due to\n                     temporal interference patterns. In this method, the optimal gaze velocity is determined\n                     from parameters such as the decomposition parameters of the striped patterns and the\n                     projection speed of these images. However, it is known that in the human visual system,\n                     gaze velocity lags behind the target object\u2019s velocity during pursuit eye movements.\n                     Therefore, we conducted an experiment with human subjects to compare the theoretically\n                     optimal gaze velocity with the actual measured velocity. The results suggest that\n                     the velocity of the guide point used to induce gaze movement should be set to approximately\n                     1.5 times the theoretical optimal gaze velocity.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Multimodal Experiences<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770540\">SketchTo3DGen : GenAI Powered Articulation Ready 3D Asset Ideation using 3D Sketches\n                  and Audio Descriptions<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Shivam Ashok Shukla<\/li>\n               <li class=\"nameList\">Raghav Mittal<\/li>\n               <li class=\"nameList\">Lokender Tiwari<\/li>\n               <li class=\"nameList Last\">Brojeshwar Bhowmick<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We present <em>SketchTo3DGen<\/em>, a novel system for rapid 3D content ideation on a VR headset. SketchTo3DGen combines\n                     freehand 3D sketching and audio descriptions to generate photo-realistic 3D assets\n                     on-the-fly. Running on a Meta Quest headset with a Unity application, our system leverages\n                     remote GPU-accelerated services for AI-driven content creation using intuitive inputs.\n                     The user can draw a 3D sketch in mid-air and describe the intended asset verbally;\n                     our pipeline transcribes and normalizes the speech into a text prompt, selects informative\n                     viewpoints of the 3D VR sketch, generates corresponding images via a state-of-the-art\n                     text-to-image model, and finally reconstructs a 3D mesh using an image-to-3D generator.\n                     The entire workflow is experienced in VR with minimal interface elements. We describe\n                     the design motivations, technical pipeline, and user interaction details of SketchTo3DGen.\n                     This VR in-headset pipeline, using intuitive inputs in the form of hand-drawn 3D VR\n                     sketch and speech, streamlines 3D modeling, accelerating the generation of articulation\n                     ready 3D assets.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            <h2>SESSION: Immersive visualization and Interaction<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3772055\">Visualizing Simulated Airflow and Thermal Comfort in Extended Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">HuiSeong Lee<\/li>\n               <li class=\"nameList\">Myoung Gon Kim<\/li>\n               <li class=\"nameList\">Hyemi Kim<\/li>\n               <li class=\"nameList Last\">JungHyun Han<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We propose an XR (eXtended Reality) system that simulates and visualizes airflow and\n                     thermal comfort controlled by an air conditioner. This system reconstructs an indoor\n                     space to make its 3D map, tracks the air conditioner\u2019s pose without markers, runs\n                     fluid and thermal comfort simulators, and visualizes the simulation results on a mobile\n                     XR device. It enables users to instantly find comfortable\/uncomfortable spots in the\n                     indoor space and control the air conditioner more effectively. The modules for 3D\n                     mapping, object pose tracking, simulation and visualization are integrated on a single\n                     XR device, making the system portable and widely usable.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768396\">HoloViz Office: Location-Independent Mixed Reality Workspace for 3D Medical Data Visualization<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Shiqi Yu<\/li>\n               <li class=\"nameList Last\">Casper Harteveld<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>The rapid shift to remote work has exposed limitations in traditional tools for 3D\n                     data visualization, particularly in medical training. We present HoloViz Office, a\n                     portable Mixed Reality (MR) workspace that enables immersive 3D medical data visualization\n                     independent of physical location. Unlike Virtual Reality (VR) solutions that isolate\n                     users, our MR approach using HoloLens preserves situational awareness while providing\n                     intuitive interaction with complex medical datasets. We demonstrate the system through\n                     comparative brain analysis for Cerebral Small Vessel Disease (CSVD) and dynamic human\n                     anatomy exploration. Evaluation with 10 participants confirms that HoloViz Office\n                     provides location-independent, convenient, and immersive visualization capabilities,\n                     contributing to effective remote collaboration tools in medical education.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770535\">ChemersiveLLM: Prompt-to-VR Simulation of Chemistry Experiments Using Generative AI<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Thanh Ngoc-Dat Tran<\/li>\n               <li class=\"nameList\">Viet-Tham Huynh<\/li>\n               <li class=\"nameList\">G. Michael Poor<\/li>\n               <li class=\"nameList\">Minh-Triet Tran<\/li>\n               <li class=\"nameList Last\">Tam V. Nguyen<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Large Language Models (LLMs) offer significant potential for integration with Virtual\n                     Reality (VR), but current AI systems struggle to generate accurate 3D environments\n                     and support semantic interaction. We present ChemersiveLLM, a VR-based chemistry learning\n                     platform that leverages LLMs for instruction sequencing, natural language grounding,\n                     and real-time guidance. Using a semantic action-mapping framework, the system translates\n                     AI-generated content into structured lab actions, enabling multimodal interaction,\n                     embodied experimentation, and intelligent feedback. Comparative evaluation across\n                     textbook, chatbot-based, and VR learning shows that our system improves engagement,\n                     comprehension, and satisfaction, underscoring its promise as a next-generation tool\n                     for science education.<\/p>\n                  <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Reproducibility<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3771087\">VRCare &#8211; Improving Diagnostic Eyecare Experience &#8211; An investigative study<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Yuvaraj Kumaresan<\/li>\n               <li class=\"nameList\">Aryan Arora<\/li>\n               <li class=\"nameList Last\">Anand Bhojan<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We present VRCare, a VR-based eye screening tool enabling remote, self-guided vision\n                     assessments for colour blindness, visual field, myopia, and contrast sensitivity.\n                     A user study with 33 participants evaluated usability, comfort, and perceived effectiveness.\n                     Participants rated the system highly for intuitiveness (4.45\/5), ease of use (4.48\/5),\n                     and comfort (4.27\/5), with 91% willing to reuse the tool. Lower confidence in diagnostic\n                     accuracy (3.76\/5) and reports of mild discomfort highlight the need for clinical validation\n                     and ergonomic refinement. Overall, findings demonstrate VR\u2019s potential for accessible\n                     vision screening outside clinical settings.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768402\">Virtual Reality in the Treatment of Male Sexual Disorders: Protocol of a Replication\n                  Study<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Zuzanna Paw\u0142owska<\/li>\n               <li class=\"nameList\">Martyna Gaj<\/li>\n               <li class=\"nameList\">Joanna Wojnowska<\/li>\n               <li class=\"nameList Last\">Pawe\u0142 Jemio\u0142o<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Sexual disorders negatively affect mental health and quality of life. Current therapies\n                     are often limited by adaptability and the challenges of safe exposure. Virtual reality\n                     (VR) offers immersive, controllable environments that can overcome these barriers.\n                     Early studies suggest benefits of VR-based therapies, but the small samples, unclear\n                     protocols, and outdated devices used limit their implementation. In this paper, we\n                     propose to replicate and evaluate VR-assisted therapy through a two-phase design:\n                     a pilot assessing immersion and cybersickness of participants, followed by a&nbsp;randomized\n                     trial comparing psychotherapy with and without VR.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Interaction Design and Input Techniques III<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770551\">StereoVisPoseNet: Stereo-based Visibility-aware Egocentric 3D Pose Estimation Network<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Sungjin Hong<\/li>\n               <li class=\"nameList\">Hye-sun Kim<\/li>\n               <li class=\"nameList\">Cho-rong Yu<\/li>\n               <li class=\"nameList Last\">Youn-Hee Gil<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Egocentric 3D pose estimation is challenging due to occlusions and errors at articulated\n                     joints. We propose StereoVisPoseNet, a stereo-based visibility-aware network that\n                     integrates depth and explicit joint visibility prediction to guide Transformer-based\n                     regression and refinement. Our method reduces MPJPE from 76.04 mm to 31.91 mm and\n                     PA-MPJPE from 63.43 mm to 28.73 mm compared to UnrealEgo, with substantial improvements\n                     for arms and legs. These results demonstrate the importance of combining stereo depth\n                     with visibility-aware modeling for robust egocentric pose estimation.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770544\">FlowZone: Real-Time Pose-Tracking Virtual Reality(VR) Yoga and Mindfulness Meditation<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Joey Aw<\/li>\n               <li class=\"nameList\">Xiaoyang Chen<\/li>\n               <li class=\"nameList\">Yuehan Ma<\/li>\n               <li class=\"nameList\">Mengfei Song<\/li>\n               <li class=\"nameList\">Javier Ng<\/li>\n               <li class=\"nameList Last\">Bhojan Anand<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This work presents <em>FlowZone<\/em>, an exploratory VR prototype that integrates yoga and mindfulness meditation. The\n                     system uses headset and controller data for lightweight pose tracking and provides\n                     real-time feedback in guided sessions. A preliminary user study (N=6) revealed common\n                     challenges in yoga practice, including uncertainty about pose correctness, difficulty\n                     maintaining consistency, and the importance of calming environments. These insights\n                     informed the design of FlowZone, which combines accessible pose guidance with immersive,\n                     meditation-oriented settings. While not a full evaluation, our early findings suggest\n                     that VR yoga and meditation can lower barriers to practice and support stress reduction,\n                     pointing toward promising directions for future research.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            <h2>SESSION: Affective, Collaborative, and Social Interaction<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770539\">MultiSphere: Latency Optimized Multi-User 360\u00b0 VR Telepresence with Edge-Assisted\n                  Viewport Adaptive IPv6 Multicast<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Dieter Frehlich<\/li>\n               <li class=\"nameList\">Xincheng Huang<\/li>\n               <li class=\"nameList Last\">Robert Xiao<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>360\u00b0 video telepresence with VR enables immersive remote collaboration, but scaling\n                     to multiple users is subject to bandwidth and latency constraints. We present <em>MultiSphere<\/em>, a multi-user edge-assited 360\u00b0 VR telepresence system, that combines viewport-adaptive\n                     IPv6 multicast tiling with a novel dual keyframe interval (KeyInt) streaming technique.\n                     Our approach addresses the latency bottleneck inherent in joining live streams of\n                     video using standard video codecs while maintaining visual quality through strategic\n                     use of low and high KeyInt streams. Our system achieves 75-94% bandwidth savings and\n                     an average request-to-decode latency of 56&nbsp;ms, a 79% reduction compared to using a\n                     regular single-KeyInt stream.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h2>SESSION: Avatars, Agents, and Embodiedment<\/h2>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770547\">MoPriC : Two Stage Approach for Text Guided Motion-Primitives Composition<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jeong Yeon Lee<\/li>\n               <li class=\"nameList\">Soungsill Park<\/li>\n               <li class=\"nameList Last\">youngho chai<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Text-to-motion generative models suffer from the long-term dependency problem, where\n                     it becomes difficult to maintain the context of text instructions as the motion length\n                     increases. Also, current MoCap datasets include only predefined actions and fail to\n                     reflect diverse individual styles. To address these limitations, we introduced MoPriC,\n                     a two-stage motion composition framework that produces sequential motions from elementary\n                     motion primitives guided by text descriptions. We also presented DancePrimitives,\n                     a new dataset of collected motion primitives to capture the semantics of each unit\n                     motion.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768403\">AR and LLM-Based Virtual Agent for ABA-Oriented Social Training in Autistic Children<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">wenfang cui<\/li>\n               <li class=\"nameList Last\">Zichun Guo<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This paper presents an interactive training system that combines augmented reality\n                     (AR) with large language models (LLMs) to support children with autism spectrum disorders\n                     (ASD) in practicing social skills. The virtual AR agent equipped with multimodal sensing,\n                     including eye tracking, speech recognition, and facial emotion detection, monitors\n                     the status of children in real time, and provides adaptive training in four key skills:\n                     emotional expression, eye contact, initiating interaction, and understanding social\n                     etiquette, following the principles of Applied Behavior Analysis (ABA). Driven by\n                     LLMs, the agent delivers personalized verbal instructions and animated feedback. Expert\n                     reviews suggest that the proposed system offers an expandable, context-aware intervention\n                     framework, serving as a valuable supplement to traditional behavioral therapies for\n                     children with ASD.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3772054\">The Effect of Avatar Transparency on Collaboration in Shared Virtual Spaces<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Bruno Caby<\/li>\n               <li class=\"nameList\">Guillaume Bataille<\/li>\n               <li class=\"nameList\">Florence Danglade<\/li>\n               <li class=\"nameList Last\">Jean-R\u00e9my Chardonnet<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>In shared virtual spaces, users tend to mimic real-world social behaviors, such as\n                     maintaining interpersonal distances and avoiding collisions. During remote collaboration,\n                     these behaviors can limit movement and positioning, despite users are not co-located.\n                     Prior work found that avatar transparency reduces users\u2019 positioning constraints induced\n                     by social behaviors, as long as social presence between collaborators. Our goal was\n                     to develop a system that enhances navigation freedom without significantly increasing\n                     social presence. We designed a transparency management system based on interpersonal\n                     distances and collision avoidance. A user study involving groups of three remote collaborators\n                     indicated that our system reduced the distances between users compared to fully opaque\n                     avatars without any significant difference in social presence.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Security and Systems<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768408\">Head Movement Biometrics for Continuous Authentication in Virtual Reality<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Paul Sansah Gyreyiri<\/li>\n               <li class=\"nameList Last\">Diksha Shukla<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>This paper presents an approach for continuous user authentication in VR using head\n                     movement biometrics, utilizing bilateral head position data from the stereoscopic\n                     rendering systems of VR headsets. Our method employs a 1D Convolutional Neural Network\n                     (CNN) with a specialized feature extractor designed to capture the temporal head movement\n                     patterns, head impulse movements, pose stabilization behaviors, and frequency-domain\n                     characteristics from bilateral head velocities. We evaluated the system using 30 participants\n                     who performed door-opening and walking tasks across two sessions, separated by 17\n                     days. The system achieved an average Equal Error Rate (EER) of 2.9% for door-opening\n                     tasks, 8.3% for walking tasks, and 5.67% in activity-invariant scenarios, when an\n                     authentication decision was made \u2248 every 0.3 seconds after an initial 14-second calibration\n                     period.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: VRST 2025 Demos: Affective, Collaborative, and Social Interaction Demos I<\/h2>\n            \n            \n            \n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768399\">Virtual Reality for Urban Soundscape Design: Exploring knowledge sharing, creation,\n                  and workplace integration<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Richard Yanaky<\/li>\n               <li class=\"nameList Last\">Catherine Guastavino<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>Urban sound is treated as noise, a nuisance to be mitigated, often in reaction to\n                     complaints. However, well planned urban sounds can also contribute to the quality\n                     of urban spaces. Yet, Professionals of the Built Environment are not equipped to work\n                     with sound proactively as a resource for sustainable city making. City Ditty, an interactive\n                     soundscape simulator, was developed to utilize a multisensory approach for professionals\n                     who are not accustomed to working with sound. City Ditty acts as conduit for knowledge\n                     discovery and sharing for engaging professionals in proactive urban sound planning.\n                     It provides a platform for education, rapid development, and evaluation of urban soundscapes.\n                     An overview of City Ditty, its use in knowledge creation and sharing, and early adopter\n                     use cases are provided.<\/p>\n                  <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Immersive Visualization and Interaction Demos<\/h2>\n            \n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3768401\">An Inner-Wrist Trackball Interaction Technique for Pointing and Gesture Input through\n                  Body-Rubbing<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Jin Okii<\/li>\n               <li class=\"nameList Last\">Hiroyuki Manabe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>With the increasing adoption of smart glasses, there is a growing need for efficient\n                     and precise input techniques in augmented reality (AR) environments. Current input\n                     options, like smartphones and hand-tracking, have limitations, such as requiring external\n                     devices to be picked up or needing sufficient space in front of the user. We propose\n                     an inner wrist-worn device with a trackball that allows users to operate it by sliding\n                     their wrist against their body. This technique provides precise, long-range pointing\n                     without extensive motion space and enables continuous interaction. Experiments show\n                     that this technique achieves reliable pointing performance regardless of user posture\n                     or sliding area. Potential applications we developed include shooting games, gesture\n                     control, and smartwatch input, demonstrating its effectiveness for compact and precise\n                     interactions in AR\/VR environment and daily use.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Cybersickness, Health, and Digital Twins Demos<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770548\">Optical-Flow-Compensated Virtual Screens: Mitigating Visually Induced Motion Sickness\n                  in Mixed Reality Video Viewing<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Kota Hayakawa<\/li>\n               <li class=\"nameList Last\">Tatsuo Nakajima<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We propose a novel technique for mitigating Visually Induced Motion Sickness (VIMS)\n                     during video viewing on virtual screens in a mixed-reality (MR) environment. The key\n                     idea is to counteract the motion of objects in a video (optical flow) by moving the\n                     virtual screen in an opposing direction and speed. We implemented this approach in\n                     a head mounted display (HMD) application and conducted the evaluation experiment to\n                     examine its effectiveness. The results indicate that the proposed method reduces VIMS\n                     in complex clips containing irregular rotations and translational motion. Moreover,\n                     it appears to be more effective than dynamic field of view restriction techniques.\n                     The proposed method also seems to enhance spatial presence relative to both conventional\n                     viewing and dynamic field of view restriction, therefore achieving VIMS mitigation\n                     and heightened immersion.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770545\">Interactive Depth-Shifting Aerial Image Display Using Angle Changes Between the Display\n                  and Optical Elements<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Motohiro Makiguchi<\/li>\n               <li class=\"nameList\">Ayami Hoshi<\/li>\n               <li class=\"nameList Last\">Naoto Abe<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We implemented an interactive system that enables a three-dimensional movement of\n                     an aerial image. The proposed system employs two servo motors to adjust the angles\n                     of the retroreflective element and the light source display, thereby controlling the\n                     depth position of the aerial image. The developed prototype can reproduce the position\n                     of a virtual character in three dimensions within a cubic range of 100 mm per side.\n                     Furthermore, by sensing the user\u2019s fingertips, the prototype enables spatial interactions\n                     such as the character following the fingertip, landing on it, and being flicked away\n                     by the user.<\/p>\n               <\/div>\n            <\/div>\n            \n            \n            <h2>SESSION: Multimodal Experiences Demos I<\/h2>\n            \n            \n            \n            <h2>SESSION: Affective, Collaborative, and Social Interaction Demos II<\/h2>\n            \n            \n            \n            \n            \n            \n            \n            \n            \n            <h2>SESSION: Multimodal Experiences Demos II<\/h2>\n            \n            <h3><a class=\"DLtitleLink\" title=\"Full Citation in the ACM Digital Library\" referrerpolicy=\"no-referrer-when-downgrade\" href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3756884.3770553\">A Multimodal Haptic System for Pulling Virtual Objects<\/a><\/h3>\n            <ul class=\"DLauthors\">\n               <li class=\"nameList\">Kaito Shibata<\/li>\n               <li class=\"nameList Last\">Akihiro Matsuura<\/li>\n            <\/ul>\n            <div class=\"DLabstract\">\n               <div style=\"display:inline\">\n                  <p>We propose an interactive haptic system that lets users pull virtual objects with\n                     various loads and multimodal feedback. The system is based on a syringe-plunger mechanism\n                     with a proportional solenoid valve at the syringe tip, while the plunger serves as\n                     the extractable object. The plunger integrates load cells for vertical and bending\n                     forces, an accelerometer, and a vibro transducer for audio and tactile feedback. Experiments\n                     examined how vertical loads vary with valve opening and timing, and how bending loads\n                     change with deflection. A virtual plant application demonstrates pulling and bending\n                     with diverse load profiles and combined tactile-auditory sensations.<\/p>\n               <\/div>\n            <\/div>\n            \n            <\/div>\n      <\/div>\n   <\/body>\n<\/html>\n","protected":false},"excerpt":{"rendered":"<p>VRST &#8217;25: Proceedings of the 2025 31st ACM Symposium on Virtual Reality Software and Technology VRST &#8217;25: Proceedings of the 2025 31st ACM Symposium on Virtual Reality Software and Technology Full Citation in the ACM Digital Library SESSION 1: Interaction Design and Input Techniques I SpatialMouse: A Hybrid Pointing Device for Seamless Interaction Across 2D and 3D Spaces Sebastian Hubenschmid&#8230;<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_eb_attr":"","footnotes":""},"class_list":["post-290","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/pages\/290","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/comments?post=290"}],"version-history":[{"count":7,"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/pages\/290\/revisions"}],"predecessor-version":[{"id":899,"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/pages\/290\/revisions\/899"}],"wp:attachment":[{"href":"https:\/\/vrst.acm.org\/vrst2025\/index.php\/wp-json\/wp\/v2\/media?parent=290"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}