Example-Based Framework for Perceptually Guided Audio Texture Generation

08/23/2023
by   Purnima Kamath, et al.
0

Generative models for synthesizing audio textures explicitly encode controllability by conditioning the model with labelled data. While datasets for audio textures can be easily recorded in-the-wild, semantically labeling them is expensive, time-consuming, and prone to errors due to human annotator subjectivity. Thus, to control generation, there is a need to automatically infer user-defined perceptual factors of variation in the latent space of a generative model while modelling unlabeled textures. In this paper, we propose an example-based framework to determine vectors to guide texture generation based on user-defined semantic attributes. By synthesizing a few synthetic examples to indicate the presence or absence of a semantic attribute, we can infer the guidance vectors in the latent space of a generative model to control that attribute during generation. Our results show that our method is capable of finding perceptually relevant and deterministic guidance vectors for controllable generation for both discrete as well as continuous textures. Furthermore, we demonstrate the application of this method to other tasks such as selective semantic attribute transfer.

READ FULL TEXT

page 3

page 5

page 7

page 8

research
04/23/2023

Towards Controllable Audio Texture Morphing

In this paper, we propose a data-driven approach to train a Generative A...
research
03/24/2017

Perception Driven Texture Generation

This paper investigates a novel task of generating texture images from p...
research
11/18/2019

Learning to Synthesize Fashion Textures

Existing unconditional generative models mainly focus on modeling genera...
research
02/25/2020

Unsupervised Semantic Attribute Discovery and Control in Generative Models

This work focuses on the ability to control via latent space factors sem...
research
02/27/2023

Continuous descriptor-based control for deep audio synthesis

Despite significant advances in deep models for music generation, the us...
research
06/27/2022

Sound Model Factory: An Integrated System Architecture for Generative Audio Modelling

We introduce a new system for data-driven audio sound model design built...
research
11/25/2020

MTCRNN: A multi-scale RNN for directed audio texture synthesis

Audio textures are a subset of environmental sounds, often defined as ha...

Please sign up or login with your details

Forgot password? Click here to reset