Simulating Synesthesia In Spatially-Based Real-Time Audio-Visual Performance

Research output: Contribution to journalArticlepeer-review

13 Downloads (Pure)

Abstract

In this paper I will describe and present examples of my live audio-visual work for 3D spatial environments. These projects use motion-tracking technology to enable users to interact with sound, light and video using their body movements in 3D space. Specific video examples of one past project (Virtual DJ) and one current project (Virtual VJ) will be shown to illustrate how flexible user interaction is enabled through a complex and precise mapping of 3D space to media control. In these projects audience members can interact with sound, light and video in real-time by simply moving around in space with a tracker in hand. Changes in sound can be synchronized with changes in light and/or real-time visual effects (i.e. music volume = light brightness = video opacity). These changes can be dynamically mapped in real-time to allow the user to consolidate the roles of DJ, VJ and light designer in one interface. This interaction model attempts to reproduce the effect of synesthesia, in which certain people experience light or color in response to music.
Original languageEnglish
Pages (from-to)214-229
JournalLive Visuals, Leonardo Electronic Almanac
Volume19
Issue number3
Publication statusPublished - Jul 2013

Fingerprint

Dive into the research topics of 'Simulating Synesthesia In Spatially-Based Real-Time Audio-Visual Performance'. Together they form a unique fingerprint.

Cite this