VR remote collaboration system lets share user experience on the move without causing VR sickness

NewsGuard 100/100 Score

Researchers from Tokyo Metropolitan University have engineered a virtual reality (VR) remote collaboration system which lets users on Segways share not only what they see but also the feeling of acceleration as they move. Riders equipped with cameras and accelerometers can feedback their sensations to a remote user on a modified wheelchair wearing a VR headset. User surveys showed significant reduction in VR sickness, promising a better experience for remote collaboration activities.

Virtual reality (VR) technology is making rapid headway, letting users experience and share an immersive, 3D environment. In the field of remote work, one of the major advances it offers is a chance for workers in different locations to share what they see and hear in real-time. An example is users on personal mobility devices in large warehouse facilities, factories, and construction sites. Riders can cover large areas with ease while highlighting issues in real-time to a remote co-worker. However, one major drawback can ruin the whole experience: VR sickness. VR sickness is a type of motion sickness which comes from users seeing "motion" through their headsets without actually moving. Symptoms include headaches, nausea, and sometimes vomiting. The problem is particularly acute for the example above, when the person sharing the experience is moving about.

To get around this issue, researchers from Tokyo Metropolitan University led by Assistant Professor Vibol Yem have created a system which lets users share not only what they see, but the sensation of movement as well. They focused on Segways as a common, widely available personal mobility vehicle, mounting two 3D cameras and a set of accelerometers to measure not only visual cues but detailed information on the acceleration of the vehicle. This was fed back via the internet to a remote user wearing a VR headset on a modified wheelchair, with separate motors attached to the wheels. As the user on the Segway accelerated, so did the wheelchair, allowing remote users to not only see the same scenery, but feel the same acceleration. Of course, the wheelchair wasn't allowed to move the same distances as the Segway; it was gently returned to its original position when the Segway was not accelerating.

The team put their device to the test by asking volunteers to become remote users and rate their experience. There was a reduction in VR sickness of 54% when the sensations of movement were added, with excellent ratings for the user experience. They also noticed subtleties in how the information should be fed back. For example, users found it best when around 60% of the acceleration suggested by the visual cues was fed back to the wheels, largely due to the sensitivity of the vestibular system (how we sense balance, orientation and motion) compared to our vision.

Though improvements are still needed, the team's system promises exciting new possibilities for remote collaboration, freeing remote users from a major drawback of VR technology.

This work was supported by the Local-5GProject at Tokyo Metropolitan University, MIC/SCOPE # 191603003 and JSPS KAKENHI Grant Number 18H04118.

Source:
Journal reference:

Yem, V., et al. (2022) Vehicle-ride sensation sharing system with stereoscopic 3D visual perception and vibro-vestibular feedback for immersive remote collaboration. Advanced Robotics. doi.org/10.1080/01691864.2022.2129033.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
New sustainable diagnostic approach offers precision cancer testing with minimal environmental impact