Automatic song-type classification and speaker identification of norwegian ortolan bunting (Emberiza Hortulana) vocalizations

Marek B. Trawicki, Michael T. Johnson, Tomasz S. Osiejuk

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

27 Scopus citations

Abstract

This paper presents an approach to song-type classification and speaker identification of Norwegian Ortolan Bunting (Emberiza Hortulana) vocalizations using traditional human speech processing methods. Hidden Markov Models (HMMs) are used for both tasks, with features including Mel-Frequency Cepstral Coefficients (MFCCs), log energy, and delta (velocity) and delta-delta (acceleration) coefficients. Vocalizations were tested using leave-one-out cross-validation. Classification accuracy for 5 song-types is 92.4%, dropping to 63.6% as the number and similarity of the songs increases. Song-type dependent speaker identification rates peak at 98.7%, with typical accuracies of 80-95% and a low end at 76.2% as the number of speakers increases. These experiments fit into a larger framework of research working towards methods for acoustic censusing of endangered species populations and more automated bioacoustic analysis methods.

Original languageEnglish
Title of host publication2005 IEEE Workshop on Machine Learning for Signal Processing
Pages277-282
Number of pages6
DOIs
StatePublished - 2005
Event2005 IEEE Workshop on Machine Learning for Signal Processing - Mystic, CT, United States
Duration: Sep 28 2005Sep 30 2005

Publication series

Name2005 IEEE Workshop on Machine Learning for Signal Processing

Conference

Conference2005 IEEE Workshop on Machine Learning for Signal Processing
Country/TerritoryUnited States
CityMystic, CT
Period9/28/059/30/05

ASJC Scopus subject areas

  • General Engineering

Fingerprint

Dive into the research topics of 'Automatic song-type classification and speaker identification of norwegian ortolan bunting (Emberiza Hortulana) vocalizations'. Together they form a unique fingerprint.

Cite this