Sound Design Strategies for Latent Audio Space Explorations Using Deep Learning Architectures

05/24/2023
by   Kıvanç Tatar, et al.
0

The research in Deep Learning applications in sound and music computing have gathered an interest in the recent years; however, there is still a missing link between these new technologies and on how they can be incorporated into real-world artistic practices. In this work, we explore a well-known Deep Learning architecture called Variational Autoencoders (VAEs). These architectures have been used in many areas for generating latent spaces where data points are organized so that similar data points locate closer to each other. Previously, VAEs have been used for generating latent timbre spaces or latent spaces of symbolic music excepts. Applying VAE to audio features of timbre requires a vocoder to transform the timbre generated by the network to an audio signal, which is computationally expensive. In this work, we apply VAEs to raw audio data directly while bypassing audio feature extraction. This approach allows the practitioners to use any audio recording while giving flexibility and control over the aesthetics through dataset curation. The lower computation time in audio signal generation allows the raw audio approach to be incorporated into real-time applications. In this work, we propose three strategies to explore latent spaces of audio and timbre for sound design applications. By doing so, our aim is to initiate a conversation on artistic approaches and strategies to utilize latent audio spaces in sound and music practices.

READ FULL TEXT

page 4

page 5

page 6

research
12/05/2022

Audio Latent Space Cartography

We explore the generation of visualisations of audio latent spaces using...
research
07/01/2019

Universal audio synthesizer control with normalizing flows

The ubiquity of sound synthesizers has reshaped music production and eve...
research
05/31/2020

Introducing Latent Timbre Synthesis

We present the Latent Timbre Synthesis (LTS), a new audio synthesis meth...
research
06/30/2021

A Generative Model for Raw Audio Using Transformer Architectures

This paper proposes a novel way of doing audio synthesis at the waveform...
research
02/10/2020

Unsupervised Learning of Audio Perception for Robotics Applications: Learning to Project Data to T-SNE/UMAP space

Audio perception is a key to solving a variety of problems ranging from ...
research
04/15/2019

Are Nearby Neighbors Relatives?: Diagnosing Deep Music Embedding Spaces

Deep neural networks have frequently been used to directly learn represe...
research
07/23/2018

Auto-adaptive Resonance Equalization using Dilated Residual Networks

In music and audio production, attenuation of spectral resonances is an ...

Please sign up or login with your details

Forgot password? Click here to reset