Title :
Perceptual coherence as an analytic for procedural music and audio mappings in virtual space
Author_Institution :
Center for Computer Research in Music and Acoustics Stanford University
fDate :
3/1/2015 12:00:00 AM
Abstract :
Real-time data generated by virtual actors and their mediated interactions in simulated space can be repurposed to dynamically generate sound and music. Procedural audio and music systems afford interaction designers, composers and sound artists the opportunity to create tight couplings between the visual and auditory modalities. Designing procedural mapping schemata can become problematic when players or observers are presented with audio-visual events within novel environments wherein the validity of their own prior knowledge and learned expectations about sound, image and interactivity are put into question. This paper presents the results of a user-study measuring users´ perceptions of audio-visual crossmodal correspondences between low-level attributes of motion and sound. Study results were analyzed using the Bradley-Terry statistical model, effectively calculating the relative contribution of each crossmodal attribute within each attribute pairing to the perceived coherence or ´fit´ between audio and visual data.
Keywords :
"Videos","Avatars","Visualization","Instruments","Music","Coherence","Observers"
Conference_Titel :
Sonic Interactions for Virtual Environments (SIVE), 2015 IEEE 2nd VR Workshop on
DOI :
10.1109/SIVE.2015.7361288