a musical data mining primer

18
A Musical Data Mining Primer CS235 – Spring ’03 Dan Berger [email protected]

Upload: alexandra-brennan

Post on 30-Dec-2015

25 views

Category:

Documents


2 download

DESCRIPTION

A Musical Data Mining Primer. CS235 – Spring ’03 Dan Berger [email protected]. Outline. Motivation/Problem Overview Background Types of Music Digital Representations Psychoacoustics Query (Content vs. Meta-Data) Categorization & Clustering Finding More Conclusion. Motivation. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: A Musical Data Mining Primer

A Musical Data Mining Primer

CS235 – Spring ’03Dan Berger

[email protected]

Page 2: A Musical Data Mining Primer

Outline

Motivation/Problem Overview Background

Types of Music Digital Representations Psychoacoustics

Query (Content vs. Meta-Data) Categorization & Clustering Finding More Conclusion

Page 3: A Musical Data Mining Primer

Motivation

More music is being stored digitally: PressPlay offers 300,000 tracks for download

As collections grow – organizing and searching manually become hard; How to find the “right” music in a sea of

possibilities? How to find new artists given current

preferences? How to find a song you heard on the radio?

Page 4: A Musical Data Mining Primer

Problem Overview

Music is a highly dimension time series: 5 minutes @ CD quality > 13M samples!

It seems logical to apply data mining and IR techniques to this form of information. Query, Clustering, Prediction, etc.

Application isn’t straightforward for reasons we’ll discuss shortly.

Page 5: A Musical Data Mining Primer

Background: Types of Music Monophonic: one note sounds at a

time. Homophonic: multiple note sound –

all starting (and ending) at the same instant.

Polyphonic: no constraints on concurrency. Most general – and difficult to handle.

Page 6: A Musical Data Mining Primer

Background: Digital Representations Structured (Symbolic):

MIDI – stores note duration & intensity, instructions for a synthesizer

Unstructured (Sampled): PCM – stores quantized periodic samples

Leverages Nyquist/Shannon’s sampling thm. to faithfully capture the signal.

MP3/Vorbis/AAC – discards “useless” information – reduces storage and fidelity

Use psychoacoustics Some work at rediscovering musical structure.

Page 7: A Musical Data Mining Primer

Background: Psychoacoustics Two main relevant results:

Limited, freq. dependant resolution Auditory masking

We hear different frequencies differently: sound spectrum broken into “critical bands”

We “miss” signals due to spectral &/or temporal “collision.” Loud sounds mask softer ones, Two sounds of similar frequency get blended

Page 8: A Musical Data Mining Primer

Query – Content is King

Current systems use textual meta-data to facilitate query: Song/Album Title, Artist, Genre*

The goal is to query by the musical content: Similarity

‘find songs “like” the current one’ ‘find songs “with” this musical phrase’

Page 9: A Musical Data Mining Primer

Result: Query By Humming A handful of research systems

have been built that locate songs in a collection based on the user humming or singing a melodic portion of the song. Typically search over a collection of

monophonic MIDI files.

Page 10: A Musical Data Mining Primer

Content Based Query

Recall: music is a time series with high dimensionality. Need robust dimensionality

reduction. Not all parts of music are equally

important. Feature extraction – remember the

important features. Which features are important?

Page 11: A Musical Data Mining Primer

Similarity/Feature Extraction The current “hard problem” – there are

ad-hoc solutions, but little supporting theory. Tempo (bpm), volume, spectral qualities,

transitions, etc. Sound source: is it a piano? a trumpet? Singer recognition: who’s the vocalist?

Collectively: “Machine Listening” These are hard problems with some positive

results.

Page 12: A Musical Data Mining Primer

Compression Complexity

Different compression schemes (MP3/Vorbis/AAC) use psychoacoustics differently. Different implementations of a

scheme may also! Feature extraction needs to be

robust to these variations. Seems to be an open problem.

Page 13: A Musical Data Mining Primer

Categorization/Clustering

Genre (rock/r&B/pop/jazz/blues/etc.) is manually assigned – and subjective. Work is being done on automatic

classification and clustering. Relies on (and sometimes reinvents) the

similarity metric work described previously.

Page 14: A Musical Data Mining Primer

Browsing & Visualization:

LOUD: physical exploration Islands of Music: uses self

organizing maps to visualize clusters of similar songs.

Page 15: A Musical Data Mining Primer

Current Efforts

Amazon/iTunes/etc. use collaborative filtering. If the population is myopic and predictable, it

works well, otherwise not. Hit Song Science – clusters a provided set

of songs against a database of top 30 hits to predict success. Claims to have predicted the success of Nora

Jones. Relatable – musical “fingerprint”

technology – involved with “Napster 2”

Page 16: A Musical Data Mining Primer

Finding More

Conferences: Int. Symposium on Music IR (ISMIR) Int. Conference on Music and AI (ICMAI) Joint Conference on Digital Libraries

Journals: ACM/IEEE Multimedia

Groups: MIT Media Lab: Machine Listening Group

Page 17: A Musical Data Mining Primer

Conclusion

Slow steady progress is being made.

“Music Appreciation” is fuzzy we can’t define it but we know it

when we hear it. References, and more detail, are in

my survey paper, available shortly on the web. http://www.cs.ucr.edu/~dberger

Page 18: A Musical Data Mining Primer

Fini

Questions?