Body-Based Technology Alleviates VR Motion Sickness

Hopefully more effective than Antivert (motion-sickness) pills, Alka Seltzer, and Pepto Bismol, MONKEYmedia offers its patented body-based navigation solution (BodyNav) for hands-free virtual reality interactions. BodyNav leverages the existing on-board sensors of smartphones and advanced 3D headsets in novel and unanticipated ways to engage the body’s innate center of gravity. This human-centered interaction approach reduces motion sickness artifacts and enhances navigation abilities in virtual and augmented realities (VR/AR), as well as first-person view (FPV) drone aviation contexts.

 

Motion Sickness

Free Monthly Newsletter

Compelling read? Subscribe to FierceEmbeddedTech!

The embedded tech sector runs the market’s trends. FierceEmbeddedTech subscribers rely on our suite of newsletters as their must-read source for the latest news, developments and analysis impacting their world. Sign up today to get news and updates delivered to your inbox and read on the go.

 

Motion sickness has long been a complaint amongst virtual reality gamers and drone pilots. Traditional stereoscopic headset interfaces use multiple sensor axes (e.g. rotate left/right, pivot up/down, tip left/right) to establish viewer orientation, while requiring handheld controllers (e.g., joysticks, gamepads, keyboards, etc.) for locomotion. Visually “moving” through space while in a sedentary posture creates sensory imbalances that can cause dizziness and nausea in the viewer. Oculus’ former Chief Scientist goes so far as to call hand controllers "sickness generators." Addressing this problem, MONKEYmedia’s patented, hands-free BodyNav technology creates more intuitive virtual interactions by remapping control axes to accomplish both orientation and locomotion with natural body movement. This provides the organic equilibrium needed to circumvent sensory imbalances.

 

How BodyNav Works

 

Without any custom hardware, BodyNav uses distinct sensor axes for independent functions to maintain equilibrium in the body’s proprioceptive system. Viewers simply lean, using either their head or torso, to move themselves through virtual spaces, or to move their drones through remote physical spaces. This allows the sensory receptors, which receive stimuli internally and relate to the body’s position and movement, to properly engage with virtual or remote content, synchronizing visual and vestibular senses and reducing motion sickness-inducing factors.

 

Use Cases

 

BodyNav can be readily adapted to modernize a user experience with just a few lines of code. For developers like Garriott, it enhances first-person gaming by freeing the hands from managing avatar movement to focus on other tasks. It will also amplify multi-camera performances and sporting events, control of remote vehicles, video conferencing and telepresence applications, street-view maps, augmented reality, architectural simulations, and 3D user interfaces for browsing data models, documents and images – all while keeping users comfortable, engaged and entertained.

 

For more information, visit MONKEYmedia and to learn how to incorporate BodyNav technology into a VR or drone piloting user experience, contact [email protected].

Suggested Articles

The technology and business issues of hybrid and electric vehicles will be open for debate at the Electric and Hybrid Vehicle Technology Expo and Battery Show…

HCC Embedded (HCC) has unveiled SafeexFAT, a failsafe RTOS-independent exFAT file system implementation for deeply embedded applications. The system is…

General Motors has decided to team with Google to integrate more of that platform’s features into future vehicle infotainment systems.