Estimation of genre, mood and instrumentation classes
Sound as Data
The transformation of listening habits and the rapid growth of audio markets have led to new requirements for broadcasters, as well as increased expectations from listeners. How can these be met to remain one step ahead?
With Ircam Amplify
You enjoy access to state-of-the-art technology for the indexation and deduplication of your sound and music databases
You reap the benefits of cutting-edge tools to automatically analyze the quality of your music files
You capitalize on solutions based on artificial intelligence technology to generate playlists and other personalized audio content
These offers are based on IRCAM’s audio descriptors:
- “low-level” descriptors, involving roughly 4,000 coefficients that are calculated using complex scientific methods.
- “high level” descriptors, which can result from algorithms or machine learning, that describe the music being listened to using simple parameters (meta data).
These audio descriptors can be used to automatically extract information from music recordings using different modules:
Estimation of tempo, meter, rhythmic complexity, percussiveness, temporal positioning of beats and downbeats
Estimation of tonality
Estimation of the succession of music chords
Estimation of the temporal structure of an audio file
Sound summary of an audio track