VR Gestures
[Work in Progress]

Design & Code

Explorations of gesture-based UI. Made in Unity on an Oculus Quest.

I am not fond of large menu panels or over-relying on the 'menu' button in VR. Gestures offer many opportunities to activate small specific UI tools when you need them at arm reach.

I made these prototypes with very simple colliders on the fingers, which is very quick for sketching ideas, and can be refined for some light playtesting.

Close Fist

Stretch Palm


Gestures should be tied to the orientation of the hand.

The same action has a very different feeling when your palm is facing you, and also affects the positioning of the UI.

Very specific actions can be created, especially when using both hands.
We can also use comfortable hand positions for quick repeated actions, similar to a button press.

Picture Frame. Both hands make an "L", at a distance apart.

Steering Wheel. Both hands make fist, at a distance apart.

Thumb tap on knuckle with closed fist. More accurate than expected. Could replace a button press for browsing actions.

Gestures can be combined. Here I'm using Stretch Palm for the 'open' action, and Close Fist for the 'close' action.

Palm-sized UI elements are great for things that benefit from the slight rotational tilt of the hand.

A gesture can also open a large UI that doesn't fit directly on the hand, like this world map.

We can choose our gestures based on how easily we want to access a feature.
A more specific gesture has higher friction to activate. It can also include more detailed secondary features like scaling.

Single hand gesture makes this panel easy to access.

Two hands, bottom corners. Allows for scaling the panel.

Two hands, diagonal corners. Allows for width + height scaling of the panel.

When using two hands, the second hand can be mapped to a related secondary feature.

Here I mapped the finger pinch to the zoom (FOV) of the camera.

Since there is not much distance between the fingers, smoothing the values using Lerp was essential.