RAVE Model Challenge - Award Ceremony

video

information

Type
Ensemble de conférences, symposium, congrès
performance location
Ircam, Salle Igor-Stravinsky (Paris)
date
March 28, 2025

RAVE (Realtime Audio Variational autoEncoder) is an algorithm designed for real-time, high-quality audio waveform synthesis using neural networks. It leverages a variational autoencoder (VAE) architecture, which compresses audio data into a compact latent representation, allowing efficient reconstruction of audio signals.

Key features of RAVE include:

Fast, high-quality audio generation: It excels at producing accurate audio in real-time, making it ideal for interactive applications (20x real-time at 48 kHz sampling rate on standard CPU)
Real-time use: Integrated with tools like Max and Pure Data (Pd), RAVE can be used with the nn~ decoder for real-time sound generation and transformation. A VST plugin makes it easy to use in any DAW.
Applications: Common uses include audio synthesis, timbre transformation, and style transfer.
In short, RAVE is a powerful tool for real-time audio generation, offering both speed and quality.

In just a few months, RAVE popularized the creation of models based on audio recordings, thanks in particular to the publication of a series of tutorials and open-source code. A growing and ebullient community of users took hold of the algorithm, and numerous models emerged. Although these models can be quite costly to produce (around twenty GPU hours), very few have so far been published, often due to copyright issues. This challenge concerns models trained on personal recordings for which the authors own all rights.

The aim of this challenge is to support the authors of the best models and to collectively establish a repertoire of RAVE models, enabling everyone to benefit from the richness and variety of approaches in the field of timbre/music transfer.

The challenge is hosted by the DAFNE+ platform, which promotes content using NFTs.

A public vote awards three prizes to participants.

speakers

From the same archive

Mettre en son The Powder Toy - Kieran McAuliffe

The “falling sand” genre of games provide a unique “sandbox” experience to players, encouraging curiosity and creativity. Players experiment with a variety of powdered elements which are subjected to a detailed physics system and may react

March 28, 2025

Video

Sinusoidal run rhythm

sinusoidal run rhythm is created by the addition of in-phase cosine functions in integer ratios. Their maxima are temporally and dynamically shifted in relation to corresponding notated rhythms and exhibit a physicality that is not present

March 28, 2025

Video

ART MUSIC DENMARK presents : Presentation of “vssl” (new hardware electronic instrument) - Xavier Bonfill

March 28, 2025

Video

Performance télématique immersive - Randall Packer, Théophile Clet, Federico Foderaro

At its core, Telematic Theater features the Audio-Visual Panner, a spatialization tool that synchronizes 3D performer positioning in spherical environments with 3D ambisonic and binaural sound. This correlation between sound, image, movemen

March 28, 2025

Video

Session C-LAB : Application de Dicy2 dans la production - The Day in Gad-Avia - Chia Hui Chen, Jing-shiuan Tsang

Process - Sound Collection and Training Using sound data from 2023 to 2024 registered by NanFormosa, it is placed in the Memory Creator for analysis, with Nana performing improvisational interactions. Additionally, a MIDI controller is use

March 28, 2025

Video

Installation vidéo interactive « Here's the Information We Collect » - Tansy Xiao

« Here's the Information We Collect » is a multi-channel interactive video installation tailored to respond to selected privacy policy on major social media platforms. The audience members are invited to engage with the work by speaking int

March 28, 2025

Video

Overton - Synthèse spatiale décorrélée - Martin Antiphon

Decorrelated Spatial Synthesis involves the addition of classical synthesizer parts to spatial coordinates, and establishes a correlation between synthesis parameters and spatial positions. For each polyphony voice, each section of the synt

March 28, 2025

Video

« Latent Terrain » : Dissecting the Latent Space of Neural Audio Autoencoders - Jasper Zheng Shuoyang

We present Latent Terrain, an algorithmic approach to dissecting the latent space of a neural audio autoencoder into a two-dimensional plane. Latent Terrain questions the conventional paradigms of dimensionality reduction in creative intera

March 28, 2025

Video

share


Do you notice a mistake?

IRCAM

1, place Igor-Stravinsky
75004 Paris
+33 1 44 78 48 43

opening times

Monday through Friday 9:30am-7pm
Closed Saturday and Sunday

subway access

Hôtel de Ville, Rambuteau, Châtelet, Les Halles

Institut de Recherche et de Coordination Acoustique/Musique

Copyright © 2022 Ircam. All rights reserved.