Riffusion leverages stable-diffusion image-to-sound models to generate infinite, evolving musical “riffs” as spectrogram images that are converted back into audio.
Blends AI art and music, encouraging cross-disciplinary creativity in visual-to-audio synthesis. Reasons why this application is interesting for young people: • Visual interface generates sonic textures • Instant audio–visual feedback loop • Great for live coding and electronic music
Basic Usage Description
1. Visit Riffusion.com. 2. Paint or select a spectrogram image. 3. Hit “Generate Audio.” 4. Tweak diffusion parameters for variation.
Usage Characteristics
• Free in-browser demo • Open-source; can self-host for unlimited use • Compute-intensive; slower on low-spec machines
Possible Problems, Restrictions, Risks
• Output can be noisy or dissonant • Requires some audio-editing skill to polish
Ethical and Safety Issues
• Generated riffs may inadvertently mimic copyrighted melodies • Encourage attribution of AI-assisted work
Scalability
• Classroom demos in audio-visual art • Scoring snippets for indie games or films • Research in generative audio
-Scalability to use in other education levels: • Primary: Simple paint-sound play • Secondary: Teach spectrogram basics • University: Advanced AI-audio research
Additional Information
• Hosted notebooks and GitHub repo • Integrations with Ableton via Max for Live