40
MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Embed Size (px)

Citation preview

Page 1: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MUSIC INFORMATION RETRIEVAL SYSTEMS

Author: Amanda Cohen

Page 2: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Music Information Retrieval Systems Based on content from the following

webpage: http://mirsystems.info/index.php?id=mirsystems

Other good sources on MIR and MIR systems http://www.music-ir.org - Virtual home of

music information retrieval research http://www.ismir.net - The International

Symposium on Music Information Retrieval

Page 3: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Audentify!

Developers: F. Kurth, A. Ribbrock, M. Clausen Relevant Publications:

Kurth, F., Ribbrock, A., Clausen, M. Identification of Highly Distorted Audio Material for Querying Large Scale Data Bases. 112th Convention of the Audio Engineering Society, May 2002, Munich, Convention Paper

Kurth, F., Ribbrock, A., Clausen, M. Efficient Fault Tolerant Search Techniques for Full-Text Audio Retrieval. 112th Convention of the Audio Engineering Society, May 2002, Munich, Convention Paper

Ribbrock, A. Kurth, F. A Full-Text Retrieval Approach to Content-Based Audio Identification. International Workshop on Multimedia Signal Processing. St. Thomas, US Virgin Islands, December 9-11, 2002

Kurth, F. A Ranking Technique for fast Audio Identification. International Workshop on Multimedia Signal Processing. St. Thomas, US Virgin Islands, December 9-11, 2002

Clausen, M., Kurth, F. A Unified Approach to Content-Based and Fault Tolerant Music Recognition, IEEE Transactions on Multimedia. Accepted for publication

Page 4: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Audentify

System Description Takes signal queries (1-5 seconds, 96-128

kbps) Searches by audio fingerprint Returns a file ID that corresponds with a

song in the database Currently a part of the SyncPlayer system

Page 5: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

SyncPlayer

Developers: F. Furth, M. Muller, D. Damm, C. Fremerey, A. Ribbrock, M. Clausen

Relevant Publications: Kurth, F., Müller, M., Damm, D., Fremerey, Ch. Ribbrock, A.,

Clausen, M. SyncPlayer - An Advanced System for Multimodal Music Access, Proceedings of the 6th International Conference on Music Information Retrieval (ISMIR 2005), London, GB

Kurth, F., Müller, M., Ribbrock, A., Röder, T., Damm, D., Fremerey, Ch. A Prototypical Service for Real-Time Access to Local Context-Based Music Information. Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR 2004), Barcelona, Spain. http://www-mmdb.iai.uni-bonn.de/download/publications/kurth-service-ismir04.pdf

Fremerey, Ch., SyncPlayer - a Framework for Content-Based Music Navigation, Diplomarbeit at the Multimedia Signal Processing Group Prof. Dr. Michael Clausen, University of Bonn, 2006, Bonn, Germany

URL: http://audentify.iai.uni-bonn.de/synchome/index.php?pid=01

Page 6: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

SyncPlayer

System Description Query type(s): audio files (mp3, wav, MIDI), lyrics,

MusicXML, Score scans (primary data) Generates “derived data” from query

extracts features generates annotations compiles synchronization data

Submitted to SyncPlayer Server, which can perform three services (at present) audio identification (through audentify) provide annotations for a given song retrieval in lyrics annotation

SyncPlayer Client: audio-visual user interface, allow user to playback, navigate, and search in the primary data

Page 7: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

ChoirFish

Developers: A. Van Den Berg, S. Groot Relevant Publications:

Groot, S., Van Den Berg, A., The Singing Choirfish: An application for Tune Recognition, Proceedings of the 2003 Speech Recognition Seminar, LIACS 2003

URL: http://www.ookii.org/university/speech/default.aspx

Page 8: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

ChoirFish

System Description: Query by humming Contour features used for matching Used Parson’s Code to determine contour

Code is based on the direction of note transitions 3 characters for each possible direction:

R: The note is the same frequency as the previous note D: The note is lower in frequency than the previous note U: The note is higher in frequency than the previous note

Generated by changing the audio to the frequency domain via Fast Fourier Transform and using the highest frequency peak to determine pitch and pitch change

Page 9: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

CubyHum

Developers: S. Pauws Relevant Publications:

Pauws, S., CubyHum: A Fully Operational Query by Humming System, ISMIR 2002 Conference Proceedings (2002): 187--196, doi:10.1.1.108.8515

PDF of paper: http://ismir2002.ismir.net/proceedings/02-FP06-2.pdf

Page 10: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

CubyHum

System Description: Query by Humming: user queries system by humming

the desired song Pitch is estimated by computing the sum of harmonically

compressed spectra (sub-harmonic summation, or SHS). Musical events (note onsets, gliding tones, inter-onset-

intervals) are detected Query is transformed via quantization into musical

score, which is used to create a MIDI melody for auditory feedback

Approximate pattern matching used to find matching song

Distance between melodies defined based on interval sizes and duration ratios to compensate for imperfect query (people don’t always hum the correct melody in the correct key)

Page 11: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Fanimae

Developers: Iman S.H. Suyoto, Alexandra L. Uitdenbogerd and Justin Zobel

Relevant Publications Suyoto, I.S.H., Uitdenbogerd, A.L., Simple

efficient n-gram indexing for effective melody retrieval, Proceedings of the Annual Music Information Retrieval Evaluation eXchange, 2005

URL: http://mirt.cs.rmit.edu.au/fanimae/

Page 12: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Fanimae

System Description: Desktop Music Information Retrieval System Search by symbolic melodic similarity Query: a melody sequence that contains both pitch and

duration information Melody sequence is standardized

Intervals are encoded as a number of semitones, with large intervals being reduced

Coordinate matching used to detect melodic similarity Query is split into n-grams of length 5, as are any possible

answers count the common distinct terms between query and

possible answer return results ranked by similarity

Page 13: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Foafing the Music

Developers: Music Technology Group of the Universitat Pompeu Fabra

Relevant Publications: Celma, O. Ramírez, M. Herrera, P., Foafing the

music: A music recommendation system based on RSS feeds and user preferences Proceedings of 6th International Conference on Music Information Retrieval; London, UK, 2005, http://ismir2005.ismir.net/proceedings/3119.pdf

URL: http://foafing-the-music.iua.upf.edu

Page 14: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Foafing the Music

System Description: Returns personalized music recommendations

based on a user’s profile (listening habits, location) Bases recommendation information on info

gathered across the web Similarity between artists determined by their

relationships between one another (ex: influences, followers)

Creates RSS feed for news related to favorite artists Computes musical similarity between specific songs

Page 15: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Meledex/Greenstone

Developers: McNab, Smith, Bainbridge, Witten

Relevant Publications: McNab, Smith, Bainbridge, Witten, The New

Zealand digital library MELody inDEX, D-Lib Magazine, May 1997

URL: http://www.nzdl.org/fast-cgi-bin/music/musiclibrary

Page 16: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Meldex/Greenstone

System Description: Receives audio queries (hummed, sung, or

played audio) Filters audio to get fundamental frequency Input sent to pitch tracker, which returns

average pitch estimate for each 20ms Note duration can optionally be taken into

account, as well as user defined tuning Results found using approximate string

matching based on melodic contour

Page 17: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Musipedia/Melodyhound/Tuneserver Developer: Rainer Typke Relevant Publications:

Prechelt, L., Typke, R., An Interface for Melody Input. ACM Transactions on Computer-Human Interaction, June 2001

URL: http://www.musipedia.org

Page 18: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Musipedia/Melodyhound/Tuneserver System Description

Query by humming system Record sound, system converts into sound

wave Converts query sound wave into Parson’s Code Match by melodic contour

Determine distance between query and possible results via editing distance (calculate the number of modifications necessary to turn one string into the other)

Return results with smallest distance

Page 19: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MIDIZ

Developers: Maria Cláudia Reis Cavalcanti, Marcelo Trannin Machado , Alessandro de Almeida Castro Cerqueira, Nelson Sampaio Araujo Júnior and Geraldo Xexéo

Relevant Publications: Cavalcanti, Maria Cláudia Reis et al. MIDIZ:

content based indexing and retrieving MIDI files. J. Braz. Comp. Soc. [online]. 1999, vol. 6, no. 22008-11-02]. http://www.scielo.br/scielo.php?script=sci_arttext&pid=S0104-65001999000300002&lng=&nrm=iso ISSN 0104-6500. doi: 10.1590/S0104-65001999000300002

Page 20: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MIDIZ

System description Database that stores, indexes, searches for and recovers MIDI files

based on the description of a short musical passage Allows for non-exact queries Musical sequence is based on intervals between notes Uses wavelet transform and a sliding window in the melody

Window defines a note sequence of a given size (2^k) and moves through the song note by note

Each sequence in the window is converted into a vector storing the interval distances

First note in a sequence is assigned the value 1 Values of the following notes are determined by their chromatic

distance in relation to the first note Those values are added together in pairs, and the result is converted

into coordinates in the final vector Songs in database are stored in a BD Tree, determined by

Discriminator Zone Expression Completed vector of query is submitted to tree, similar results are

returned

Page 21: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Mu-seek

Developers: Darfforst LLP URL: http://www.mu-seek.com/ System Description:

Search by title, lyrics, tune fragment, or MIDI

Uses pitch, contour, and rhythm to find matches

Page 22: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MusicSurfer

Developers: Music Technology Group of the Universitat Pompeu Fabra

Relevant Publications: Cano et al. An Industrial-Strength Content-

based Music Recommendation System, Proceedings of 28th Annual International ACM SIGIR Conference; Salvador, Brazil 2005. http://mtg.upf.edu/files/publications/3ac0d3-SIGIR05-pcano.pdf

URL: http://musicsurfer.iua.upf.edu/

Page 23: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

MusicSurfer

System Description: Automatically extracts features from songs

in database based on rhythm, instrumentation, and harmony Uses spectral analysis to determine timbre

Uses those features to search for similar songs

Page 24: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

NameMyTune

Developers: Strongtooth, Inc URL: http://www.namemytune.com/ System Description:

User hums query into microphone Results are found by other users

determining what the song is

Page 25: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Orpheus

Developers: Rainer Typke Relevant Publications:

Typke, Giannopoulos, Veltkamp, Wiering, van Oostrum, Using Transportation Distances for Measuring Melodic Similarity, ISMIR 2003

URL: http://teuge.labs.cs.uu.nl/Ruu/?id=5

Page 26: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Orpheus

System Description: Query can be example from database,

hummed or whistled melody, or a MIDI file All queries are converted into internal

database format before submission Similarity between query and results based

on Earth Mover’s Distance Two distributions are represented by signatures Distance represents the amount of “work”

required to change one signature to the other Work = user defined distance between two

signatures

Page 27: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Probabilistic “Name That Song” Developers: Eric Brochu and Nando de

Freitas Publications:

Brochu, E., Freitas, N.D., "Name That Song!": A Probabilistic Approach to Querying on Music and Text. NIPS. Neural Information Processing Systems: Natural and Synthetic 2002 (2003)

Page 28: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Probabilistic “Name That Song” System Description:

Query is composed of note transitions (Qm) and words (Qt). A match is found when a corresponding song has all elements of Qm and Qt with a frequency of 1 or greater.

Database songs are clustered. Query is performed on each song in each cluster until a match is found

Page 29: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Query by Humming (Ghias et al.) Developers: Asif Ghias, Jonathan Logan,

David Chamberlin, Brian C. Smith Relevant Publications:

Ghias, A., Logan, J., Chamberlin, D., Smith, B.C., Query by Humming - Musical Information Retrieval in an Audio Database, ACM Multimedia (1995)

URL: http://www.cs.cornell.edu/Info/Faculty/bsmith/query-by-humming.html

Page 30: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Query by Humming (Ghias et al.) System Description:

Hummed queries are recorded in Matlab Pitch tracking is performed

Converted into a string of intervals similar to Parson’s Code (U/D/S used as characters instead of R/D/U)

Baesa-Yates/Perleberg pattern matching algorithm used to find pattern matches Find all instances of the query string in the result

string with at most k mismatches Results returned in order of how they best fit the

query

Page 31: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Search by Humming

Developers: Steven Blackburn Relevant Publications:

Blackburn, S. G., Content Based Retrieval and Navigation of Music Using Melodic Pitch Contours. PhD Thesis, 2000

Blackburn, S. G., Content Based Retrieval and Navigation of Music. Masters, 1999

DeRoure, D., El-Beltagy, S., Blackburn, S. and Hall, W., A Multiagent System for Content Based Navigation of Music. ACM Multimedia 1999 Proceedings Part 2, pages 63-6.

Blackburn, S. G. and DeRoure, D. C., A tool for content based navigation of music. Proceedings of ACM Multimedia 1998, pages 361—368

DeRoure, D. C. and Blackburn, S. G., Amphion: Open Hypermedia Applied to Temporal Media, Wiil, U. K., Eds. Proceedings of the 4th Open Hypermedia Workshop, 1998, pages 27--32.

DeRoure, D. C., Blackburn, S. G., Oades, L. R., Read, J. N. and Ridgway, N., Applying Open Hypermedia to Audio, Proceedings of ACM Hypertext 1998, pages 285--286.

URL: http://www.beeka.org/research.html

Page 32: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Search by Humming

System Description: Takes query by humming, example, or MIDI Queries and database contents represented by

gross melodic pitch contour Within database, each track is stored as a set of

overlapping sub-contours of a constant length Distance between songs is determined by the

minimum cost of transforming one contour into another (similar to EMD)

Query is expanded into a set of all possible contours of the same length as the database’s sub-contours

A score is calculated for each file based on the number of times a contour in the expanded query set occurs in the file. Results are sorted in order of score

Page 33: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

SOMeJB (The SOM enhanced JukeBox) Developers: Andreas Rauber, Markus Frühwirth, E. Pampalk, D. Merkl Relevant Publications:

A. Rauber, E. Pampalk, D. Merkl, The SOM-enhanced JukeBox: Organization and Visualization of Music Collections based on Perceptual Models, Journal of New Music Research (JNMR), Swets and Zeitlinger, 2003

E. Pampalk, A. Rauber, D. Merkl, Content-based Organization and Visualization of Music Archives In: Proceedings of ACM Multimedia 2002, pp. 570-579, December 1-6, 2002, Juan-les-Pins, France

A. Rauber, E. Pampalk, D. Merkl, Using Psycho-Acoustic Models and Self-Organizing Maps to Create a Hierarchical Structuring of Music by Musical Styles, Proceedings of the 3rd International Symposium on Music Information Retrieval (ISMIR 2002), pp. 71-80, October 13-17, 2002, Paris, France.

A. Rauber, E. Pampalk, D. Merkl, Content-based Music Indexing and Organization, Proceedings of the 25. Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 02), pp. 409-410, August 11-15, 2002, in Tampere, Finland

A. Rauber, and M. Frühwirth, Automatically Analyzing and Organizing Music Archives, Proceedings of the 5. European Conference on Research and Advanced Technology for Digital Libraries (ECDL 2001), Sept. 4-8 2001, Darmstadt

URL: http://www.ifs.tuwien.ac.at/~andi/somejb

Page 34: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

SOMeJB (The SOM enhanced JukeBox) System Description:

Interface is a static, web-based map where similar pieces of music are clustered together

Music organized by a novel set of features based on rhythm patterns in a set of frequency bands and psycho-acoustically motivated transformations Extracts features that apply to loudness sensation

(intensity), and rhythm Self-organizing map algorithm is applied to

organize the pieces on a map (trained neural network)

Page 35: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

SoundCompass

Developers: Naoko Kosugi, Yuichi Nishihara, Tetsuo Sakata, Masashi Yamamuro and Kazuhiko Kushima, NTT Laboratories

System Description: User sets a metronome and hums melody in time with

clicks Database songs have three feature vectors

Tone Transition Feature Vector: contains the dominant pitch for each 16-beat window

Partial Tone Transition Feature Vector: Covers a time window different from the Tone Transition Feature Vector

Tone Distribution Feature Vector: histogram containing note distribution

Query is matched against each of the vectors, results are combined by determining the minimum

Page 36: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Tararira

Developers: Ernesto López, Martin Rocamora, Gonzalo Sosa

Relevant Publications: E. Lopez y M. Rocamora. Tararira: Sistema

de búsqueda de música por melodía cantada. X Brazilian Symposium on Computer Music. October, 2005.

URL: http://iie.fing.edu.uy/investigacion/grupos/gmm/proyectos/tararira/

Page 37: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

Tararira

System Description: User submits a hummed query Pitch tracking applied to query Audio segmentation determines note

boundaries Melodic analysis adjusts pitches to tempered

scale Results found by coding query note sequence,

find occurrences using flexible similarity rules (string matching), and refining the selection using pitch time series

Page 38: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

TreeQ

Developer: Jonathan Foote Publications:

Foote, J.T., Content-Based Retrieval of Music and Audio, C.-C. J. Kuo et al., editor, Multimedia Storage and Archiving Systems II, Proc. of SPIE, Vol. 3229, pp. 138-147, 1997

URL: http://sourceforge.net/projects/treeq/

Page 39: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

TreeQ

System Description: Primarily query by example, can also search by

classification Tree based supervised vector quantizer is built from

labeled training data Database audio is parameterized via conversion into MFCC

and energy vectors Each resulting vector is quantized into the tree

Vector space divided into “bins”, any MFCC vector will fall into one bin

A histogram based on the distribution of MFCC vectors into each bin is created for query and database audio

Songs matched based on histograms of feature counts at tree leaves Distance is determined using Euclidian distance between

corresponding templates of each audio clip Results sorted by magnitude and returned as a ranked list

Page 40: MUSIC INFORMATION RETRIEVAL SYSTEMS Author: Amanda Cohen

VisiTunes

Developers: Scott McCaulay  Further Information:

www.slis.indiana.edu/research/phd_forum/2006/mccaulay.doc

URL: http://www.naivesoft.com/  System Description:

Uses audio content of songs to calculate similarity between music and creates playlists based on the results Converts sample values of each frame to frequency

data Extracts sum total of sound energy by frequency band Uses results to simplify audio data into 256 integer

values for fast comparison