Foggy with a Chance of Vibration: Exploring Light, Sound, and Spatial Perception in Multisensory Experience Design

Additional Authors

Lydia Hicks, Department of Digital Media, College of Business and Technology, East Tennessee State University, Johnson City, TN

Abstract

The perception of a space that we feel, hear, and see emerges from the interplay of our senses rather than solely from its physical structure. This study investigates how artificial fog, sound, vibrations, and visuals shape depth perception, emotional engagement, and movement, offering insights into how sensory manipulation can transform spatial experiences. Understanding these interactions may inform the design of immersive environments, influencing how people interpret and navigate spaces both digitally and physically. By analyzing installations such as Karolina Halatek’s fog-based light sculptures in Metz, France, and teamLab’s interactive projection environments in Tokyo, Japan, this research explores how 3D soundscapes, fog density, and low-frequency vibrations impact spatial awareness and user perception. To test these effects, computer simulations in Unreal Engine 5 were conducted to examine how controlled sensory conditions influence perception. Results indicate that artificial fog alters depth perception, making spaces appear, larger, smaller, or stretched depending on density and lighting. Low-frequency vibrations generate tactile sensations from a distance, allowing sound to be felt rather than just heard, influencing interaction and movement. The synchronization of sound and visuals enhances emotional engagement, memory retention, and spatial awareness, reinforcing immersion and user presence. These findings have broad applications in Augmented Reality (AR), educational settings, interactive art, museum exhibitions, therapeutic environments, urban design, Virtual Reality (VR), and more. As these technologies evolve, the ability to fine-tune environmental stimuli will redefine immersive storytelling, interactive media, and spatial perception. Therefore, if we can manipulate perception, we can redefine reality—transforming how we experience art, space, and technology in ways once thought impossible.

Start Time

16-4-2025 9:00 AM

End Time

16-4-2025 11:30 AM

Presentation Type

Poster

Presentation Category

Science, Technology and Engineering

Student Type

Graduate Student - Masters

Faculty Mentor

Jacy Richardson

Faculty Department

Digital Media

This document is currently not available here.

Share

COinS
 
Apr 16th, 9:00 AM Apr 16th, 11:30 AM

Foggy with a Chance of Vibration: Exploring Light, Sound, and Spatial Perception in Multisensory Experience Design

The perception of a space that we feel, hear, and see emerges from the interplay of our senses rather than solely from its physical structure. This study investigates how artificial fog, sound, vibrations, and visuals shape depth perception, emotional engagement, and movement, offering insights into how sensory manipulation can transform spatial experiences. Understanding these interactions may inform the design of immersive environments, influencing how people interpret and navigate spaces both digitally and physically. By analyzing installations such as Karolina Halatek’s fog-based light sculptures in Metz, France, and teamLab’s interactive projection environments in Tokyo, Japan, this research explores how 3D soundscapes, fog density, and low-frequency vibrations impact spatial awareness and user perception. To test these effects, computer simulations in Unreal Engine 5 were conducted to examine how controlled sensory conditions influence perception. Results indicate that artificial fog alters depth perception, making spaces appear, larger, smaller, or stretched depending on density and lighting. Low-frequency vibrations generate tactile sensations from a distance, allowing sound to be felt rather than just heard, influencing interaction and movement. The synchronization of sound and visuals enhances emotional engagement, memory retention, and spatial awareness, reinforcing immersion and user presence. These findings have broad applications in Augmented Reality (AR), educational settings, interactive art, museum exhibitions, therapeutic environments, urban design, Virtual Reality (VR), and more. As these technologies evolve, the ability to fine-tune environmental stimuli will redefine immersive storytelling, interactive media, and spatial perception. Therefore, if we can manipulate perception, we can redefine reality—transforming how we experience art, space, and technology in ways once thought impossible.