Research
Gesture Photo Stack
A research prototype for browsing photographs like a
physical stack.
Gesture-controlled photo stack study.
Move with a cursor, hand, or voice to release images
into view.
Overview
Most image previews turn photographs into flat tiles. This project explores a moving stack of image cards that feels slightly physical, where images are released into view rather than simply clicked open.
The same interaction can be performed with a cursor, a tracked hand, or an optional voice trigger. The aim is to make image browsing feel more natural, more inclusive, and less dependent on constant mouse and keyboard use.
Inclusive Interaction
This began as an interaction experiment, but it quickly became a broader interface question: can image browsing ask less from hardware and more from movement, attention, and intent?
Repeated clicking, dragging, and holding can be tiring for some people and a barrier for others. This prototype is not trying to replace every input. It is opening more than one way to interact with the same visual space.
The Core Question
What feels intuitive in a tactile image interface, and what does not?
This project studies how gesture, card physics, pacing, and light multimodal input affect the way people browse photographs.
It is less about novelty and more about making image preview feel embodied, readable, and easy to use.
How It Works
- Photographs appear as a physics-based card stack inside a single viewport
- Cursor movement can release new cards into the stack
- A tracked index finger can act like a live pointer, so the same motion works through hand movement
- Pinch-based resizing stays secondary and only activates when thumb-and-index intent is clear
- Left-right pinch changes image size through relative motion rather than fixed screen position
- An optional voice trigger can release another card without taking over the interface
The resize interaction is designed to feel eased and slightly inertial, so it reads as a deliberate mode rather than an always-on gesture.
What Testing Changed
Testing the prototype with people became one of the most important parts of the project. Watching how others approached the system made it clear which interactions felt immediate and which ones created hesitation.
- Card release is the main action people understand fastest
- Hand movement works best when it behaves like direct pointer movement
- Resize becomes more trustworthy when pinch is a clear secondary mode
- Voice input works better as lightweight optional support than as a primary controller
- The screen needs to stay visually quiet so motion remains legible
The interface became simpler because testing showed where confidence dropped.
What Is Being Studied
- How gesture changes image preview
- What kinds of motion feel intuitive
- Whether light friction improves attention
- How sequencing affects the meaning of photographs
- How multiple input methods can coexist on one screen
- How a browsing interface can reduce reliance on repetitive hardware actions
Current Build
The current build is intentionally compact: one screen, one image stack, and one clear interaction space. That keeps the prototype focused on motion, release, resizing, and input behavior.
The visual system stays minimal with a dark background, a subtle grid, and quiet typography. The interface supports the experience rather than competing with it.
Why It Matters
This project is grounded in a simple belief: interacting with images does not always need to begin with clicks. Browsing can be guided by motion, and control can come from the hand, the body, or a small signal.
This prototype is one step toward a browsing interface that asks for less strain, offers more flexibility, and makes inclusive interaction part of the default experience.
