6
Music Identification via Vocabulary Tree with MFCC Peaks Tianjing Xu, Adams Wei Yu, Xianglong Liu, Bo Lang State Key Laboratory of Software Development Environment Beihang University Beijing, P.R.China {xutj, yu_wei, xlong_liu}@nlsde.buaa.edu.cn, [email protected] ABSTRACT In this paper, a Vocabulary Tree based framework is pro- posed for music identification whose target is to recognize a fragment from a song database. The key to a high recogni- tion precision within this framework is a novel feature, name- ly MFCC Peaks, which is a combination of MFCC and Spec- tral Peaks features. Our approach consists of three stages. We first build the Vocabulary Tree with 2 million MFCC Peaks features extracted from hundreds of music. Then each song in the database is quantified into some words by trav- eling from root down to a certain leaf. Given a query input, we apply the same quantization procedure to this fragmen- t, score the archive according to the TF-IDF scheme and return the best matches. The experimental results demon- strate that our proposed feature has strong identifying and generalization ability. Other trials show that our approach scales well with the size of database. Further comparison also demonstrates that while our algorithm achieves approx- imately the same retrieval precision as other state-of-the-art methods, it cost less time and memory. Categories and Subject Descriptors H.5.5 [Information Interfaces and Presentation]: Sound and Music Computing—Methodologies and techniques General Terms Algorithm,Design,Experimentation Keywords music identification, MFCC Peaks, Vocabulary Tree 1. INTRODUCTION Music identification, the most typical application of audio fingerprinting, has been a popular subject both in research [2][9][10] and industry [7][16][17]. Audio fingerprinting is defined as a condensed digital summary to establish the e- quality or similarity of two audio objects. By comparing Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. MIRUM’11, November 30, 2011, Scottsdale, Arizona, USA. Copyright 2011 ACM 978-1-4503-0986-8/11/11 ...$10.00. the fingerprinting of a query music fragment of a few sec- onds with those stored in a fingerprinting database, music identification system returns matching results and related metadata. Besides this query-by-example application, this technology can also be used to monitor broadcast and detect copyrighted material. A number of approaches have been attempted in the past for audio fingerprinting systems. One of the most famous systems is proposed by J. Haitsma [7],which selects 33 fre- quency bands covering a range from 300Hz to 2000Hz for spectral representation, generates 32-bit sub-fingerprinting for every interval of 0.37s and indicates the difference be- tween consecutive frames. Cano et al.[4] provide an overview of existing audio fingerprinting systems. Most audio finger- printing systems derive their fingerprinting from some kind of time-frequency representation. But the features they use to construct the fingerprinting are various, such as spectral peaks [18], Fourier coefficients [19] and Mel-Frequency Cep- strum Coefficients (MFCC) [5]. A recent extension work is presented in [9]. Ke et al. in- troduce a computer vision approach for music identification. The spectrogram of each music clip is treated as a 2-D im- age and music identification is transformed into a sub-image retrieval problem. Baluja et al.[2] extend [9] by proposing a wavelet-based approach combining computer-vision tech- niques and large-scale-data-stream processing. These latest methods all refer to computer vision tech- nique, so does our approach. Our work is largely inspired by D. Nister et al.[14]. They apply the Vocabulary Tree ap- proach for object recognition. We extend this work to music retrieval for the reason that it allows much more efficient lookup of visual words and shows a significant improvement of retrieval quality. In this work we propose a vocabulary-tree-based frame- work to music identification. To achieve a high retrieval precision, we also introduce a novel feature named MFCC Peaks to build the tree. We apply an indexing mechanism to enable efficient retrieval in the Vocabulary Tree. Various experiments are implemented to test the identifying abili- ty, generalization ability, scalability and time and memory cost of our framework and feature. The remainder of this paper is organized as follows: Section 2 describes our music identification approach including the MFCC Peaks feature selection, Vocabulary Tree construction, indexing and scor- ing scheme. Section 3 presents our experiment results and corresponding analysis. Section 4 draws our conclusion. 21

Music Identification via Vocabulary Tree with MFCC Peaksweiyu/Adams_Wei_Yu_Homepage_files/...with_MF… · Music Identification via Vocabulary Tree with MFCC Peaks Tianjing Xu,

Embed Size (px)

Citation preview

Music Identification via Vocabulary Tree with MFCC Peaks

Tianjing Xu, Adams Wei Yu, Xianglong Liu, Bo LangState Key Laboratory of Software Development Environment

Beihang UniversityBeijing, P.R.China

{xutj, yu_wei, xlong_liu}@nlsde.buaa.edu.cn, [email protected]

ABSTRACTIn this paper, a Vocabulary Tree based framework is pro-posed for music identification whose target is to recognize afragment from a song database. The key to a high recogni-tion precision within this framework is a novel feature, name-ly MFCC Peaks, which is a combination of MFCC and Spec-tral Peaks features. Our approach consists of three stages.We first build the Vocabulary Tree with 2 million MFCCPeaks features extracted from hundreds of music. Then eachsong in the database is quantified into some words by trav-eling from root down to a certain leaf. Given a query input,we apply the same quantization procedure to this fragmen-t, score the archive according to the TF-IDF scheme andreturn the best matches. The experimental results demon-strate that our proposed feature has strong identifying andgeneralization ability. Other trials show that our approachscales well with the size of database. Further comparisonalso demonstrates that while our algorithm achieves approx-imately the same retrieval precision as other state-of-the-artmethods, it cost less time and memory.

Categories and Subject DescriptorsH.5.5 [Information Interfaces and Presentation]: Soundand Music Computing—Methodologies and techniques

General TermsAlgorithm,Design,Experimentation

Keywordsmusic identification, MFCC Peaks, Vocabulary Tree

1. INTRODUCTIONMusic identification, the most typical application of audio

fingerprinting, has been a popular subject both in research[2][9][10] and industry [7][16][17]. Audio fingerprinting isdefined as a condensed digital summary to establish the e-quality or similarity of two audio objects. By comparing

Permission to make digital or hard copies of all or part of this work forpersonal or classroom use is granted without fee provided that copies arenot made or distributed for profit or commercial advantage and that copiesbear this notice and the full citation on the first page. To copy otherwise, torepublish, to post on servers or to redistribute to lists, requires prior specificpermission and/or a fee.MIRUM’11, November 30, 2011, Scottsdale, Arizona, USA.Copyright 2011 ACM 978-1-4503-0986-8/11/11 ...$10.00.

the fingerprinting of a query music fragment of a few sec-onds with those stored in a fingerprinting database, musicidentification system returns matching results and relatedmetadata. Besides this query-by-example application, thistechnology can also be used to monitor broadcast and detectcopyrighted material.

A number of approaches have been attempted in the pastfor audio fingerprinting systems. One of the most famoussystems is proposed by J. Haitsma [7],which selects 33 fre-quency bands covering a range from 300Hz to 2000Hz forspectral representation, generates 32-bit sub-fingerprintingfor every interval of 0.37s and indicates the difference be-tween consecutive frames. Cano et al.[4] provide an overviewof existing audio fingerprinting systems. Most audio finger-printing systems derive their fingerprinting from some kindof time-frequency representation. But the features they useto construct the fingerprinting are various, such as spectralpeaks [18], Fourier coefficients [19] and Mel-Frequency Cep-strum Coefficients (MFCC) [5].

A recent extension work is presented in [9]. Ke et al. in-troduce a computer vision approach for music identification.The spectrogram of each music clip is treated as a 2-D im-age and music identification is transformed into a sub-imageretrieval problem. Baluja et al.[2] extend [9] by proposinga wavelet-based approach combining computer-vision tech-niques and large-scale-data-stream processing.

These latest methods all refer to computer vision tech-nique, so does our approach. Our work is largely inspiredby D. Nister et al.[14]. They apply the Vocabulary Tree ap-proach for object recognition. We extend this work to musicretrieval for the reason that it allows much more efficientlookup of visual words and shows a significant improvementof retrieval quality.

In this work we propose a vocabulary-tree-based frame-work to music identification. To achieve a high retrievalprecision, we also introduce a novel feature named MFCCPeaks to build the tree. We apply an indexing mechanismto enable efficient retrieval in the Vocabulary Tree. Variousexperiments are implemented to test the identifying abili-ty, generalization ability, scalability and time and memorycost of our framework and feature. The remainder of thispaper is organized as follows: Section 2 describes our musicidentification approach including the MFCC Peaks featureselection, Vocabulary Tree construction, indexing and scor-ing scheme. Section 3 presents our experiment results andcorresponding analysis. Section 4 draws our conclusion.

21

2. MUSIC IDENTIFICATION METHODThe flow chart of our retrieval approach is shown in Figure

1. It consists of three steps. The first step aims at buildingthe Vocabulary Tree with features extracted from trainingmusic. In the second step, raw music data is processed andfeatures are quantified into words and then stored in wordsdatabase. These two steps are both off-line module while thequery step is an on-line module. Incoming music snippetis converted into words by traveling down the VocabularyTree built in the second step. After scored, these songs withhighest scores are returned to users.

Figure 1: The flow chart of proposed approach.

In our retrieval approach, there are three major compo-nents which are marked in different patterns in Figure1. Thefirst and the most important one is feature extraction, whichobtains features from raw music data and is used in all threesteps. What feature to extract is the primary difference be-tween the original Vocabulary Tree method in image pro-cessing and our proposed method. In the following sectionswe describe the three components in more detail.

2.1 Feature SelectionThe Vocabulary Tree framework is first proposed for ob-

ject recognition in image processing and the features theyused such as SIFT [11] have three characteristics:

1. The features extracted from an image are independent.2. The feature should be unique and can represent small

granularity.3. The feature should be somehow robust with distortion.While in audio processing, there are no features meeting

all these characteristics. For one thing, most features con-tain the time information. A popular method dealing withfragment retrieval is slotting. For example, in [16], Spevaket al. use 20ms windows above MFCC features. However,in this way, the feature size is quite large, because one songis typically 4 minutes long. And it is really time consumingto align query’s feature vector to the whole song’s featurevectors using methods like DTW [16]. For another, somemethods compress the large size of spectral features by onlyrecording the differences between consecutive features, suchas [7] and [9], which eliminate the dependence on time infor-mation, but lose the original spectral information of audio.When turning to robustness, the feature called spectrogrampeaks is turned to be robust in the presence of noise andapproximate linear superposability by[17] and [18].

Considering aspects described above, we propose MFC-C Peaks which combines MFCC features and spectrogrampeaks. We firstly transform audio to spectrogram by FFT(FastFourier Transform) with 0.125ms hamming windows, andthen MFCCs are computed. An illustration of spectrogramis in the left of Figure 2 while the right of Figure 2 shows thecorresponding MFCC Map which we only select the first 13

MFCCs. The peaks in counted as spectral peaks describedin [17], which is defined as a point which has a higher energythan all its neighbors in a region centered around the pointin spectrogram. We find the peaks in MFCC Map and notethe MFCC vector which contains at least one peak as ourfeature - MFCC Peaks. Figure 3 shows one MFCC Peaks inthe above Figure 2 MFCC Map. In these figures, the darkerthe color, the bigger the double value of the correspondingcell. The final feature is a 13-dimension double-value vector.

Our feature meets all the three characteristics above. Weadopt peaks, so this feature is less dependent on time infor-mation and somehow robust with distortion. Since we retainthe MFCC features, the spectral characteristic remains. Be-sides, it has low dimensionality and thus ensures high com-putational efficiency.

Time(s)F

requ

ency

0 2 4 6 8 10 12 14 16 18 200

500

1000

1500

2000

2500

3000

3500

4000

Time(s)

MF

CC

Val

ues

0 2 4 6 8 10 12 14 16 18 201

2

3

4

5

6

7

8

9

10

11

12

13

Figure 2: The spectrogram and MFCC Map of a 20seconds fragment from a pop song.

Time(s)

MF

CC

Val

ues

0 2 4 6 8 10 12 14 16 18 201

2

3

4

5

6

7

8

9

10

11

12

13

Peak Point

Our Features

Figure 3: An illustration of MFCC Peaks.

For testing our feature, we compare with the other twofeatures with the same Vocabulary Tree based framework,which are:MFCC Slot: This feature mentioned in [16], is defined as

slotting audio with 20ms windows. We also use the first 13MFCCs.Spectral Peaks: This feature is the first part of features

in [17] before hashing. We adopt the same parameters aswe adopt at MFCC Peaks. This features we used are 256-dimension vectors with double value while in [17] they haveother implementation.

22

2.2 Vocabulary Tree Construction and UsageThe Vocabulary Tree is defined as a hierarchical quan-

tization built by hierarchical k-means clustering [14]. Thealgorithm of building is shown in Algorithm 1.

Algorithm 1: Build(D,level)The Building Vocabulary Tree Algorithm

Require: A predefined branch K and level LThe initial value of level is 0

Input : Feature set DThe level going to build

Output : Vocabulary Tree, K forks and L levels

1 if level > L then2 return

3 {ci}Ki=1 ←Cluster(D)4 for i← 1 to K do5 {Di} ←Partition(D, ci)6 end7 for i← 1 to K do8 Build(Di, level + 1)9 end

First, a k-means cluster is applied on the training data,defining K cluster centers as ci. The training data is thenpartitioned into K groups, where each group consists of thedescriptors closest to a particular cluster center. The sameprocess is then recursively applied to each group, recursivelyquantified by splitting each group into K new parts. Thetree is constructed level by level, up to the maximum levelL.

Tree quantization methods are previous used by other re-searchers, but note that they are different from VocabularyTree. Instead of defining K as the total number of clusterswhich is always used in quantization tree indexing, K standsfor the number of children. The total quantization cells areKL. The quantization and the indexing are the same inVocabulary Tree, which guarantees the efficiency.

Figure 4: An illustration of the Vocabulary Treewith five branches and two levels.

Figure 4 is an illustration of the Vocabulary Tree with

K = 5, L = 2 and totally 25 words. The thicker the line,the more features belong to this group. MFCC Peaks firstlyextract from original music, then each one is travel throughthe Vocabulary Tree from the root to the leaf and get aword. We only store the words in database.

2.3 Indexing and Scoring schemeOnce the quantization is obtained, the remaining chal-

lenge is how to efficiently search in a large scale system.Based on experimental results in [14], we treat all the leavesas independent and use text retrieval method. Every leafnode in the Vocabulary Tree is associated with a word, andone song can be transformed into one document with somewords.

We use an inverted file index [15] for large-scale indexingand retrieval, whose efficiency has been proved in practice.Figure 5 shows the structure of our index. Each word hasan list in the index containing the music where the wordappears. In our current implementation, there are 5 MFCCPeaks features in each second. Hence a typical song of 200slong can be extracted out 1,000 MFCC Peaks features. TheVocabulary Tree has about 99,966 leaf nodes. Though westore the words as string, for 3 characters on average, ourfeature has a size of 0.12kbit/s, which is much smaller than2.6kbit/s in [7].

Figure 5: An illustration of stored file and invertedfile index structure.

After index has been built, the database music are rankedby their scores. We use the standard TF-IDF score. Accord-ing to the words the query contains, each song in databasewill be given a score s =

∑i wimi,where wi is the weight

of leaf-node i (leaf-node i is called wordi) and mi is the s-core wordi get in this song.wi is consider as IDF while mi

is according to the TF.When a snippet is given, some features are firstly extract-

ed from the query music, then these features are quantifiedinto some words using the Vocabulary Tree. The words areused in retrieval process to score all the songs in database.And the top x highest scoring songs are returned. The wholeretrieval method is described in Algorithm 2.

23

Algorithm 2: Retrieval(F)The Matching Algorithm

Require: Vocabulary Tree TMusic in DataBaseM

Input : A feature set F extracted from fragmentOutput : A subset ofM sorted by score

1 {wordi}ni=1 ←Quantify(F , T )2 for musicj ⊂M do3 for i← 1 to n do

4 sji ←Score(wordi,musicj)5 end

6∑

i(sji )

7 end8 Sort(sj)

3. EXPERIMENTSIn this section, we present experimental results to demon-

strate the effectiveness of our proposed feature MFCC Peaksand the efficiency of the vocabulary-tree-based retrieval al-gorithm in terms of accuracy and time cost. Specifically, wecompare the retrieval accuracy of MFCC Peaks with thoseof other features, evaluate the accuracy under distortions,test the generalization ability as well as the scalability ofthe algorithm, and finally compare our approach with thestate-of-the-art method.

The query data are built by randomly cutting segmentsfrom music in database. The database is queried with everysegment and our measurements are based on Top 1 and Top5 accuracy which means the correct song is listed withinthe first one and the first five position respectively. Theexperiments are run in a Java environment, on a PC withIntel Core2 CPU at 2.66GHz and 3GB RAM.

3.1 Experiment Setup

3.1.1 Parameters SettingOne key procedure of the experiment is to build the vo-

cabulary tree, whose dominated factors are the branch pa-rameter K and level parameter L in Algorithm 1. We testvarious combination cases of these two on a small databasewith 1,000 music, as shown in Figure 6, and find that whenK = 10, L = 5, the vocabulary tree trained by about 2million features achieves best performance in the retrievalprocedure afterwards. Therefore, we will adopt this settingto build the tree in the following experiment.

3 4 5 6

0.7

0.75

0.8

0.85

0.9

0.95

1

0.65

Levels of Vocabulary Tree

Acc

urac

y

Top1Top5

0 200 400 600 800 10000.95

0.96

0.97

0.98

0.99

1

Branch Factor k

Acc

urac

y

Top1Top5

Figure 6: Comparison on performance among vari-ous of levels and branches of Vocabulary Tree.

3.1.2 Feature SelectionIn order to provide concrete evidence to advocate the M-

FCC Peaks, we build vocabulary trees with different fea-tures, i.e. MFCC Slot, Spectral Peaks and MFCC Peaks,and then compare their identifying ability in the retrievalprocedure. Figure 7 and Table 1 present the result of thecomparison on a 1,000 music database. As shown by thefigure, the accuracy obtained by MFCC Peaks is far beyondthose of the others in terms of both Top 1 and Top 5 crite-rion, which indicates its superior identifying ability. Hencewe would use MFCC Peaks in the rest part of experimentsin this section.

MFCC Slot Spectral Peaks MFCC Peaks0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

FeaturesA

ccur

acy

Top 1Top 5

Figure 7: Comparison on the performance amongfour features. The MFCC Peaks significantly out-performs other features with 97.3% in Top 1 and99.6% in Top 5 retrieval accuracy.

Feature Top 1 Top 5MFCC Slot 77.8% 91.8%

Spectrum Peaks 40.8% 57.5%MFCC Peaks 97.3% 99.6%

Table 1: Detailed numerical result of Figure 7.

3.1.3 Melody-Line SelectionTo attain a higher precision in retrieval, we would first

apply the vocabulary tree method to the whole databaseand then a Melody-Line [8] method to the top 5 music re-turned in the above phrase, which will further distill theresult. And we will adopt this approach throughout thissection. Melody-Line tracking is first proposed in query byhumming for its toleration of deviation [6]. The sum of toneenergy is evaluated. We simply use the three level melodycontour with ”U”, ”D”, ”O” respectively noted current toneenergy which is larger, smaller than, and almost the same asits preceding one. Thus, the audio is converted to a stringconsisting of three letters {U,O,D}. The score is measuredas the edit distance between query string and music string.Edit distance of two string is defined as the number of editingoperations needed to match one with the other and dynamicprogramming algorithm is used to compute it [13]. Figure 8shows the process of tone length selection. We observe that

24

when the length of Melody-Line is 8, the best performanceof Top1 accuracy is 97.1%. The retrieval time is decreasingas the length of Melody-Line is increasing. When the lengthof Melody-Line is 8, the retrieval time is about 0.22s. Sowe set the length of Melody-Line to be 8 in all the followingexperiments.

1 4 7 10 13 16 19 200.9

0.92

0.94

0.96

0.98

1

Length Of Melody−Line

Acc

urac

y

Top 1

1 4 7 10 13 16 19 200

0.1

0.2

0.3

0.4

0.5

0.6

0.7

Length Of Melody−Line

Ret

rieva

l Tim

e (s

)

Time

Figure 8: Performance among various lengths ofMelody-Line.

3.2 Generalization Ability in Multi-type Mu-sic Retrieval

In this subsection, we will test the generalization abilityof the vocabulary tree. That is, we first build a vocabularytree using the MFCC Peaks feature extracted from the popsongs, and then retrieve other types such as Jazz, Rock,Country Folk and Classical Music via the same tree. Theresult in Table 3 reveals the fact that the tree built by onetype of music could also be applied to identify other typeswith a pretty high accuracy. Therefore, the vocabulary treerepresentation constructed by the MFCC Peaks feature ofa single type is general enough to identify multi-types ofmusic without reconstructing a tree for each type. In realscenarios especially when the training data is limited, thismerit would be beneficial.

Types top1 top5Pop 97.7% 99.6%Jazz 99.2% 99.9%

Classic 97.0% 99.8%Country folk 98.5% 99.7%Rock & Roll 96.9% 99.5%

Table 2: The retrieval accuracy of multi-types ofmusic via the vocabulary tree built by the pop songMFCC Peaks feature. The overall high accuracy in-dicates the generalization ability of the presentation.

3.3 ScalabilityWe would also evaluate the scalability of our approach in

terms of retrieval time cost in different sizes of database.From the left of Figure 9 we can see that Top 1 accuracydeclines when the size of database increases, while Top 5 ac-curacy maintains beyond 98% even in a large size database of20,000 songs. The right of Figure 9 reveals that our methodhas a stable query time cost of less than 0.3s. Hence our ap-proach can be scaled to large size of data without sacrificingmuch of the accuracy or increasing the time cost.

2000 5000 8000 11000 14000 17000 200000.8

0.82

0.84

0.86

0.88

0.9

0.92

0.94

0.96

0.98

1

The Size of Database

Acc

urac

y

Top1Top5

2000 5000 8000 11000 14000 17000 200000

0.1

0.2

0.3

0.4

0.5

The Size Of Database

Ret

rival

Tim

e (s

)

Time

Figure 9: Accuracy and query time for different sizesof database. From the pictures, we can see that ourapproach scales well with the size of data withoutsacrificing much of the performance.

Database Size Memory Database Size Memory2,000 40.7MB 12,000 219MB4,000 80.8MB 14,000 163MB6,000 120MB 16,000 306MB8,000 160MB 18,000 349MB10,000 203MB 20,000 391MB

Table 3: Memory usage for different sizes ofdatabase. From the table, we can see that the mem-ory cost grows linearly with the size of database.

3.4 Comparison with Other MethodsIn the final experiment, we will compare our method with

others, including the computer vision method proposed in[9] which will be referred to as CVMI in the following andthe state-of-the-art algorithm Shazam [17] which has beenadopted in commercial use and proven to be an effective andefficient music retrieval approach. We implement Shazam al-gorithm ourselves and carefully examined the parameters toassure it adheres strictly of that in [17], however, we shouldpoint out that the system of [17] is proposed in 2003 andShazam’s commercial implementation may have changed alot now. The implementation of CVMI is from the author’sweb site1. We will use queries of 10 seconds in a database of2000 songs. We carry out the comparison in terms of Top 1accuracy, memory and time cost. As is shown in Table 4.

Top 1 Memory Time CostCVMI 99.6% 458MB 42.6sShazam 79.1% 453MB 39.9s

Our Approach 97.1% 40.7MB 0.21s

Table 4: Comparison between CVMI, Shazam andour approach. While our method achieves a satisfy-ingly high accuracy, it costs much less memory andtime than the other two methods.

Our approach achieves a satisfyingly high accuracy, thoughnot the highest. More importantly, our method consumesmuch less memory and costs significantly less time than theother two. Therefore, our method would be suitable for most

1The publicly code is from http://www.cs.cmu.edu/~yke/musicretrieval/

25

of the real application scenarios where a high response speedis badly needed.

Baluja et al [2] also introduces an effective method formusic identification. Even though it achieves a high retrievalaccuracy, it requires about 420MB memory to store 10,000songs and takes approximately 0.73 seconds to find in thedatabase, as is reported in [2]. In this sense, our methodoutperforms it.

Moreover, since our approach hits a 99.9% Top 5 accura-cy, which is not listed above, one might further apply othertechniques instead of Melody-Line to re-rank the five re-turned songs to improve the Top 1 accuracy. This indicatesthe extension potential of our algorithm.

4. CONCLUSIONIn this paper, we propose a Vocabulary Tree based frame-

work for music identification whose target is to recognize afragment from a song database. For high recognition pre-cision and efficiency within this framework , we propose anovel feature, namely MFCC Peaks, which is a combinationof MFCC and Spectral Peaks features. The experimentalresults demonstrate that our proposed feature has strong i-dentifying and generalization ability. Other trials show thatour approach scales well with the size of database. Furthercomparison with state-of-the-art methods also demonstratesthat our method costs less time and memory.

In our experiment, we also point out one way to improveour approach. The fact that the Top 5 accuracy is quitehigh indicates that we could adopt some technique such asMelody-Line to re-rank the first five or ten returned songsand then yield a better ranking list.

5. ACKNOWLEDGMENTSThis work is supported by the National Major Research

Plan of Infrastructure Software under Grant No.2010ZX01042-002-001-00, the Foundation of the State Key Laboratory ofSoftware Development Environment under Grant No. SKLSDE-2011ZX-01.

6. REFERENCES[1] E. Allamanche, J. Herre, and M. Cremer. Audioid:

Towards content-based identification of audiomaterial. In Web Delivering of Music, 2001.

[2] S. Baluja and M. Covell. Audio fingerprinting:Combining computer vision & data stream processing.In International Conference on Acoustics, Speech, andSignal Processing.

[3] E. Batlle, J. Masip, and E. Guaus. Automatic songidentification in noisy broadcast audio. In Signal andImage Processing, 2002.

[4] P. Cano, E. Batlle, T. Kalker, and J. Haitsma. Areview of audio fingerprinting. Journal of SignalProcessing Systems, 41:271–284, 2005.

[5] P. J. O. Doets and R. L. Lagendijk. Stochastic modelof a robust audio fingerprinting system. InInternational Symposium/Conference on MusicInformation Retrieval, 2004.

[6] A. Ghias, J. Logan, D. Chamberlin, and B. C. Smith.Query by humming: musical information retrieval inan audio database. In ACM Multimedia Conference,pages 231–236, 1995.

[7] J. Haitsma, T. Kalker, and J. Oostveen. Robust audiohashing for content identification. 2001.

[8] T. han Tsai, Y. tsung Wang, J. H. Hung, and C. longWey. Compressed domain content-based retrieval ofmp3 audio example using quantization tree indexingand melody-line tracking method. In IEEEInternational Symposium on Circuits and Systems,2006.

[9] Y. Ke, D. Hoiem, and R. Sukthankar. Computervision for music identification. In Computer Visionand Pattern Recognition, pages 597–604, 2005.

[10] W. Li, Y. Liu, and X. Xue. Robust audio identificationfor mp3 popular music. In Research and Developmentin Information Retrieval, pages 627–634, 2010.

[11] D. G. Lowe. Object recognition from localscale-invariant features. In International Conferenceon Computer Vision, volume 2, pages 1150–1157,1999.

[12] M. Mohri, P. Moreno, and E. Weinstein. Efficient androbust music identification with weighted finite-statetransducers. IEEE Transactions on Audio, Speech &Language Processing, 18:197–207, 2010.

[13] G. Navarro and M. Raffinot. Flexible pattern matchingin strings - practical on-line search algorithms fortexts and biological sequences. 2002.

[14] D. Nistl ↪er and H. Stewl ↪enius. Scalable recognitionwith a vocabulary tree. In Computer Vision andPattern Recognition, pages 2161–2168, 2006.

[15] G. Salton and C. Buckley. Term-weighting approachesin automatic text retrieval. Information Processingand Management, 24:513–523, 1988.

[16] C. Spevak and E. Favreau. Soundspotter - a prototypesystem for content-based audio retrieval. 2002.

[17] A. Wang. An industrial strength audio searchalgorithm. In International Symposium/Conference onMusic Information Retrieval, 2003.

[18] C. Yang. Macs: Music audio characteristic sequenceindexing for similarity retrieval. 2001.

[19] C. Yang. Music database retrieval based on spectralsimilarity. In International Symposium/Conference onMusic Information Retrieval, 2001.

26