Mobile Game Doctor Logo

VR uses existing UX & UI models

Paul Rybicki • September 7, 2023
VR uses existing UX & UI models
Paul Rybicki • September 7, 2023

Start with existing UX & UI models in XR


For designers new to creating novel User Interfaces and novel User Interactions in a spatial environment like the current round of battery-powered VR headsets, you can leverage two areas of your prior experience as a conceptual bridge to this design space. We will get into those in detail in this article and you will leave with a conceptual foundation to operate from when designing interactions and interfaces in spatial environments.


We need to start with two aspects of human centered design: Memory and Attention. 


Your attention please…

Attention is an aspect of information processing that impacts how our users gather information from a sequence of screens over time (spatio-temporal environments) and process that information using Working Memory into larger constructs which they then attach to the goals they have formed while playing your game. This applies to a phone, a laptop or a headset.


There are three forms which impact a user’s information processing and task completion. Selective attention, sustained attention and divided attention.

Selective Attention is the act of choosing which stimuli to focus on and to process. Usually our emphasis is on visual stimuli and information input from a screen, but auditory stimuli (sound, especially spatial sound), tactile stimuli (vibration and haptic feedback) also use selective attention. These stimuli interact and can influence each other.


Sustained or focused attention is the effort it takes to maintain processing of stimuli or information elements while ignoring other stimuli, or ignoring other information from the same stimuli.


Divided attention is the ability to process or focus on more than one event or stimulus at the same time. 


When designing interfaces and interactions we often combine stimuli in a single experience, or substitute one for another, as in the case where a user taps on an on-screen button and the handset vibrates to indicate input received, but the button does not change color or size.

Types of memory

Recall dips into our Long-Term memory, our built-up knowledge over time. To recall a piece of information from long-term memory we need to see/hear/feel enough pattern fragments to connect those fragments with the memory and its associated information. 


Users should not need to recall information from screen to screen. When using an interface, Recognition is the primary system at play and uses Working, or Short-Term memory. Short-term memory is highly influenced by our attention resources, so we often use pictures or visual representations (icons) in our interfaces as a shorthand to aid in recognition and lower the burden on our attention.


Knowing our place within the screens or the flow of a game we play regularly is very similar to our ability to recall the layout and arrangement of physical spaces, known as Spatial Memory and is drawn from our long-term memory. Combining our spatial memory and our present “location” within a flow or virtual space results in Situational Awareness and is how we know which way to go, or which UI element will take us to which feature or system within the game.


Using an interface is a form of spatial navigation in short-term memory based on our processing of stimuli or an event intentionally constructed to carry information.


How we use the space around us to work (the Promise)

Recalling ideas or concepts stored in long-term memory requires enough memory traces to be reactivated through recognition (pattern completion) of any one of the stored patterns from a fragment (Smith & Kosslyn 2013). Therefore our ability to recall what functionality or feature sits behind the click of a button or which information is in what file relies on cues and on context of the user’s immediate spatio-temporal environment. It’s highly associative. 

Looking around your office, people are constantly using the walls for work spaces. Using the wall for a brainstorming session with Post-Its? Perhaps there’s a mood board of images on another wall of your workspace. That rolling dry-erase board makes a fantastic post-meeting reminder of discussions recently or even long past. Why is that? And how do we design digital interfaces and products now that have this recall capability and support multiple users just like people used to do in an office?


An author once described a shared, networked visual workspace as “Unthinkable complexity … ranged in the nonspace of the mind, clusters and constellations of data.”


The “shared” portion of that quote has been realized by wi-fi, fiber and cellular networks. The non-space of the mind can be thought of as our cognition, and our situational awareness. 


Any interface is a reference point for the user to their long-term and short-term memory. A reminder of the cluster of data and the constellation (connection to other information or decisions) in which it resides. Existing 2D and 3D GUI’s use spatial layout and information hierarchy as well as temporal sequencing and progressive refinement to give users a way to segment and compartmentalize ideas, concepts & memories. For example: you need a flashlight. Where is your backpack? Tap an icon or reach with an arm & finger gesture towards your shoulder. A visual representation of the backpack opens and items are displayed, you open a side pocket to reveal more items and the view of the backpack visually shifts to focus on the pocket and it’s contents, following your attention and focus.This is the same thing as clicking your way through folders to get to a file. In terms of UI, the Treeview model can be employed in spatial and volume-centric environments as progressive refinement, but it’s still all progressive refinement & disclosure in the end.


2D UX to a 3D space

Current head mounted display products (battery powered headsets using Quallcomm’s chips) already allow for gaze-based interactions as well as hand and finger gesture recognition. 


Think of a user's gaze and head movement like a mouse pointer. This gives you mouseOver / mouseOut events where the scene can react to your gaze. The mouse interaction model that items scale up or expand in size when you look at them, is incredibly powerful in VR, where it appears to move slightly towards you. 


When using pointing or arm gestures for interaction, thinking of your interfaces and the user’s hand as magnetic objects which can attract or repel each other will yield very natural feeling results as an interface moves slightly toward or away from a use’s hand. While this interaction is simply using the 2d interaction model of scaling up a UI element onMouseOver / onMouseOut it has a fundamentally different feeling when employed in 3d space.


Finally, immersion. Polygon count and textures of the uncanny valley are not the only key to User immersion. They are one way to get there but there is a far better method which works well on both the current and previous generation of user-purchasable hardware: Interaction & reaction. Imagine you’re in a cartoony outdoor nature scene. Everywhere you look, birds flutter away when you move your arm suddenly or make a noise. A squirrel wanders towards you then runs away when you shift your weight form one leg to the other, moving your whole upper body. Environmental sounds come from all around you, and you hear a cartoon dog sniffing at your shoes from behind you. There are literally dozens of small, medium and even large ways you can interact with the scene around you. Spend 30 seconds in that environment? Focused attention. Situational awareness. Immersion due to the sheer amount of interactivity of the digital environment you are in. This approach to immersion is buildable now and when displays improve, adding more detailed textures or more 3d audio events to this immersive foundation will pay enormous dividends.



Mobile-centric decision chunking FTW

Despite the seemingly infinite canvas of an immersive headset experience, displaying limited information to your user in a sequence of chunks, as employed by mobile games, actually works much better.Keeping in mind that Attention is generally both selective and focused, let’s consider the following scenario. A player is refilling their backpack from a storage locker. The backpack stores items in a “tool roll” style interface allowing for 2d UI projection in world space. 


They run out of room in the pack - how do you allow them to sort or discard items? 


You can magically ‘unfold’ the tool roll to expand it larger and larger than is possible on a laptop screen, sure. But do you also allow them to see how many of each item they have in storage elsewhere? And how much they could get for selling each item later? Think: what would we do on Mobile? We would show the UI needed for the sorting task the user’s attention is focused on. We might give them a filter view icon to see the same items with different detail information. Mobile design keeps user-actions chucked by context and is far more effective at task competition. Countless design patterns exist for small-screen devices and work well when applied to sequenced tasks in immersive environments.



Because letters. Text is still hard to read on the displays in current gen hardware so we use larger fonts. The next-gen hardware with their improved displays will change this however hardware limitations in reading text or UI up close to the user still exist. For the screen door effect to be resolved, a requirement for the long-use sessions of work-related VR, a combination of varifocal and high density displays will likely be needed. Our eyes naturally rotate to bring items in view into correct stereoscopic focus at different distances, but since headsets have a fixed view plane, our natural focusing muscles work against the visual stimuli when focusing ‘up close’ in the headset leading to eye strain and headaches. So for now, experiences designed using distant (2 meter plus) Interfaces with large fonts to make them readable at said distance are best.


Foveated rendering is a rendering technique which concentrates resources in the area of the display where the user is looking. This allows for battery-friendly rendering with a noticeable visual improvement. Fixed versions of this concept have existed for some time and render the periphery of a user's vision at lower resolution. Designers can rely on auditory stimuli and use 3d sound to compensate for a lower fidelity visual system. 


And don’t forget the other senses! Designers can rely on auditory stimuli and use 3d sound to compensate for a lower fidelity visual system. This can be especially useful in providing immersion and directional cues outside the User’s FOV.

How to plan for the future (the potential)

Shared VR spaces have the potential to engage groups in a shared immersive experience. There can be a temptation to use the technique we know from console or desktop and show high poly and visual detail, and it’s totally beautiful! But visual detail, while it provides an initial reaction in our users, only gets us so far. Interaction is the true key to deeper immersion, where things react to your user’s focused attention or casual action. Building from a foundation of deep and subtle interaction means as the hardware generations get more powerful, you can add layers of additional visual detail to your game that will be both beautiful and performant. 


All the steps Mobile games use to get performance and battery life from a handset should be employed. The current IDE/engines will do many things for you out of the box: combined draw calls, backface & occlusion culling, dynamic level of detail (billboards) baked lighting, dynamic resizing of UI for consistent size at varying distances, and more!


It’s up to us as designers to avoid surfaces with high color variances which create visual hierarchy problems, avoid translucency on current devices due to material or layer compositing overhead & subsequent hit to battery life, avoid exhausting our user’s attention with too much simultaneous information or choice; and to create spaces with low line of sight and multiple object occlusions for a perception of depth and distance for users new to headsets. 


In the past people have sought to use the promise of VR to provide virtual spaces in a virtual office that embrace real world workspaces and extend them. But in practice this likely results in information display that overwhelms the user’s attention & recognition. Use VR for what it’s good for right now: targeted immersive experiences of limited session time with minimal use of text that employs known interaction patterns and recognizable gestures for input. Don’t try to make it a general purpose laptop solution (yet).

A doctor is holding a cell phone in front of a blue wall.
January 8, 2025
As 2024 comes to a close, we take a moment to reflect on the milestones, challenges, and achievements that shaped MGD this year. 🚀 From groundbreaking projects to strengthened partnerships, this retrospective highlights the journey, the lessons learned, and our vision for the future. Join us as we celebrate our successes and gear up for an even more impactful 2025!
A man is typing on a laptop computer with icons coming out of it.
By Jeff Gurian October 3, 2024
Explore the transformative shift in digital transactions in "The (re)Emergence of Web, Part I." Discover how evolving technologies are reshaping online interactions and redefining the future of commerce. Dive into insights on emerging trends and their implications for businesses and consumers alike.
A poster for a mobile game doctor with flowers and butterflies.
By Patricia Pizer July 9, 2024
Discover the enchanting world of 'Wylde Flowers' with Mobile Game Doctor. Learn about prosocial gaming and how this game promotes cooperative play.
The logo for aptoide your android app store
June 26, 2024
Explore expert reactions to Aptoide, offering insights and analysis on its impact and significance in the app distribution landscape.
A person is holding a cell phone in their hand.
By Eiso Kawamoto June 12, 2024
Deciding when to change your app icon? Discover expert tips on timing and frequency to ensure your app stays fresh and relevant in a competitive market.
A yellow circle with the word appmagic on it
By Athena Peters May 28, 2024
The LiveOps & Updates Calendar shows the events occurring within a set timeframe for the chosen game, alongside a graph of the metric you want to track. It’s a one-stop view that can show the frequency and density of the events in perspective. Above the event grid, there is a performance chart that shows how this or that update influenced the targeted metric.
A doctor is holding a cell phone in front of a blue wall.
By Dave Rohrl May 8, 2024
Reflect on GDC 2024 with insights from a veteran attendee and lecturer. Discover the highlights, industry trends, and personal connections made during this year's conference. Join Mobile Game Doctor for an in-depth look at the event and its impact on the gaming community.
A blue and white poster for the game developers conference.
By Tadhg Kelly April 17, 2024
Unlock valuable insights on the mobile gaming industry with our latest blog post: 'Appmagic Response to GDC'. Dive into in-depth research and expert analysis from the Mobile Game Doctor team as they navigate through prevailing industry shifts. Don't miss out on staying ahead of the curve – read now!
A green owl is sitting on top of a yellow coin on a purple background
By Patricia Pizer March 21, 2024
Discover how Duolingo's success impacts the mobile gaming landscape in our latest blog post! Explore the intersection of gamification and language learning apps, and learn valuable insights for competing in today's fierce market. With over 500 million users, Duolingo's freemium model sets a new standard for engagement and monetization. Dive into the analysis now and stay ahead of the competition!
More Posts
Share by: