Show simple item record

Automatic singer identification in polyphonic music.

dc.contributor.authorBartsch, Mark A.
dc.contributor.advisorWakefield, Gregory H.
dc.date.accessioned2016-08-30T15:36:12Z
dc.date.available2016-08-30T15:36:12Z
dc.date.issued2004
dc.identifier.urihttp://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqm&rft_dat=xri:pqdiss:3138110
dc.identifier.urihttps://hdl.handle.net/2027.42/124358
dc.description.abstractTrained human listeners show a remarkable ability to identify singers from their voices alone even in new contexts. The identity of a singer can provide a good deal of information about a particular song, and thus systems that can identify a singer's voice can be important for browsing and searching through large databases of multimedia content. Music, however, comprises a very rich signal class that is often difficult to analyze due to the prevalence of complex mixtures of sound sources. In some areas of audio content analysis, this difficulty is addressed without performing source separation by using monophonic (single-source) recordings. Practical systems for musical content analysis, however, must be able to handle polyphonic (multi-source) recordings, which are by far the norm. In this work, we develop a set of methods for singer identification in polyphonic music and present a detailed performance evaluation of these methods. In particular, we seek to determine whether an approach that separates the voice signal from the polyphonic mixture yields superior performance over one that does not. We begin by developing and evaluating a set of basic but extensible methods for singer identification in monophonic music. These methods are then extended through the introduction of PESCE, a system for voice detection and separation in musical mixtures. With PESCE used for pre-processing, we evaluate the performance of our singer identification methods under a large variety of conditions. In particular, we examine performance for features that employ voice separation, temporal voice location only, and neither separation nor location. These evaluations include a basic database of piano-accompanied songs, a database of vocal music with orchestra, and a database containing a variety of popular music. Contrary to our initial hypothesis, we find that there is no uniform performance improvement when using voice separation. Notably, however, knowing when the singing voice appears within the musical mixture improves performance substantially. In general, identification performance is found to degrade for types of music with greater interference from accompanying instruments. Additionally, performance decreases when the system attempts to generalize to novel recordings by a given singer.
dc.format.extent213 p.
dc.languageEnglish
dc.language.isoEN
dc.subjectAutomatic Singer Identification
dc.subjectPolyphonic Music
dc.subjectSinging
dc.titleAutomatic singer identification in polyphonic music.
dc.typeThesis
dc.description.thesisdegreenamePhDen_US
dc.description.thesisdegreedisciplineApplied Sciences
dc.description.thesisdegreedisciplineElectrical engineering
dc.description.thesisdegreegrantorUniversity of Michigan, Horace H. Rackham School of Graduate Studies
dc.description.bitstreamurlhttp://deepblue.lib.umich.edu/bitstream/2027.42/124358/2/3138110.pdf
dc.owningcollnameDissertations and Theses (Ph.D. and Master's)


Files in this item

Show simple item record

Remediation of Harmful Language

The University of Michigan Library aims to describe library materials in a way that respects the people and communities who create, use, and are represented in our collections. Report harmful or offensive language in catalog records, finding aids, or elsewhere in our collections anonymously through our metadata feedback form. More information at Remediation of Harmful Language.

Accessibility

If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.