logo
ResearchBunny Logo
Exploring how a Generative AI interprets music

The Arts

Exploring how a Generative AI interprets music

G. Barenboim, L. D. Debbio, et al.

Discover how Google's MusicVAE interprets music, revealing fascinating insights into 'music neurons' and how they distinguish elements like pitch, rhythm, and melody. This groundbreaking research was conducted by Gabriela Barenboim, Luigi Del Debbio, Johannes Hirn, and Verónica Sanz.... show more
Abstract
We use Google's MusicVAE, a Variational Auto-Encoder with a 512-dimensional latent space to represent a few bars of music, and organize the latent dimensions according to their relevance in describing music. We find that, on average, most latent neurons remain silent when fed real music tracks: we call these "noise" neurons. The remaining few dozens of latent neurons that do fire are called "music neurons". We ask which neurons carry the musical information and what kind of musical information they encode, namely something that can be identified as pitch, rhythm or melody. We find that most of the information about pitch and rhythm is encoded in the first few music neurons: the neural network has thus constructed a couple of variables that non-linearly encode many human-defined variables used to describe pitch and rhythm. The concept of melody only seems to show up in independent neurons for longer sequences of music.
Publisher
Springer Nature
Published On
Jan 01, 2023
Authors
Gabriela Barenboim, Luigi Del Debbio, Johannes Hirn, Verónica Sanz
Tags
MusicVAE
Variational Auto-Encoder
Latent Space
Music Neurons
Pitch
Rhythm
Melody
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny