Do you notice a mistake?
NaN:NaN
00:00
RAVE (Realtime Audio Variational autoEncoder) is an algorithm designed for real-time, high-quality audio waveform synthesis using neural networks. It leverages a variational autoencoder (VAE) architecture, which compresses audio data into a compact latent representation, allowing efficient reconstruction of audio signals.
Key features of RAVE include:
Fast, high-quality audio generation: It excels at producing accurate audio in real-time, making it ideal for interactive applications (20x real-time at 48 kHz sampling rate on standard CPU)
Real-time use: Integrated with tools like Max and Pure Data (Pd), RAVE can be used with the nn~ decoder for real-time sound generation and transformation. A VST plugin makes it easy to use in any DAW.
Applications: Common uses include audio synthesis, timbre transformation, and style transfer.
In short, RAVE is a powerful tool for real-time audio generation, offering both speed and quality.
In just a few months, RAVE popularized the creation of models based on audio recordings, thanks in particular to the publication of a series of tutorials and open-source code. A growing and ebullient community of users took hold of the algorithm, and numerous models emerged. Although these models can be quite costly to produce (around twenty GPU hours), very few have so far been published, often due to copyright issues. This challenge concerns models trained on personal recordings for which the authors own all rights.
The aim of this challenge is to support the authors of the best models and to collectively establish a repertoire of RAVE models, enabling everyone to benefit from the richness and variety of approaches in the field of timbre/music transfer.
The challenge is hosted by the DAFNE+ platform, which promotes content using NFTs.
A public vote awards three prizes to participants.
The “falling sand” genre of games provide a unique “sandbox” experience to players, encouraging curiosity and creativity. Players experiment with a variety of powdered elements which are subjected to a detailed physics system and may react
March 28, 2025
sinusoidal run rhythm is created by the addition of in-phase cosine functions in integer ratios. Their maxima are temporally and dynamically shifted in relation to corresponding notated rhythms and exhibit a physicality that is not present
March 28, 2025
March 28, 2025
At its core, Telematic Theater features the Audio-Visual Panner, a spatialization tool that synchronizes 3D performer positioning in spherical environments with 3D ambisonic and binaural sound. This correlation between sound, image, movemen
March 28, 2025
Process - Sound Collection and Training Using sound data from 2023 to 2024 registered by NanFormosa, it is placed in the Memory Creator for analysis, with Nana performing improvisational interactions. Additionally, a MIDI controller is use
March 28, 2025
« Here's the Information We Collect » is a multi-channel interactive video installation tailored to respond to selected privacy policy on major social media platforms. The audience members are invited to engage with the work by speaking int
March 28, 2025
Decorrelated Spatial Synthesis involves the addition of classical synthesizer parts to spatial coordinates, and establishes a correlation between synthesis parameters and spatial positions. For each polyphony voice, each section of the synt
March 28, 2025
We present Latent Terrain, an algorithmic approach to dissecting the latent space of a neural audio autoencoder into a two-dimensional plane. Latent Terrain questions the conventional paradigms of dimensionality reduction in creative intera
March 28, 2025
Do you notice a mistake?