Please use this identifier to cite or link to this item: http://hdl.handle.net/11452/32501
Full metadata record
DC FieldValueLanguage
dc.contributor.authorKinnunen, Tomi-
dc.contributor.authorSaeidi, Rahim-
dc.contributor.authorPohjalainen, Jouni-
dc.contributor.authorAlku, Paavo-
dc.date.accessioned2023-05-03T10:43:45Z-
dc.date.available2023-05-03T10:43:45Z-
dc.date.issued2013-
dc.identifier.citationHanilçi, C. vd. (2013). “Speaker identification from shouted speech: Analysis and compensation”. International Conference on Acoustics Speech and Signal Processing ICASSP, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 8027-8031.en_US
dc.identifier.issn1520-6149-
dc.identifier.urihttps://doi.org/10.1109/ICASSP.2013.6639228-
dc.identifier.urihttp://hdl.handle.net/11452/32501-
dc.descriptionBu çalışma, 26-31 Mayıs 2013 tarihleri arasında Vancouver[Kanada]’da düzenlenen IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)’da bildiri olarak sunulmuştur.tr_TR
dc.description.abstractText-independent speaker identification is studied using neutral and shouted speech in Finnish to analyze the effect of vocal mode mismatch between training and test utterances. Standard mel-frequency cepstral coefficient (MFCC) features with Gaussian mixture model (GMM) recognizer are used for speaker identification. The results indicate that speaker identification accuracy reduces from perfect (100 %) to 8.71 % under vocal mode mismatch. Because of this dramatic degradation in recognition accuracy, we propose to use a joint density GMM mapping technique for compensating the MFCC features. This mapping is trained on a disjoint emotional speech corpus to create a completely speaker- and speech mode independent emotion-neutralizing mapping. As a result of the compensation, the 8.71 % identification accuracy increases to 32.00 % without degrading the non-mismatched train-test conditions much.en_US
dc.description.sponsorshipInst Elect & Elect Engineersen_US
dc.description.sponsorshipInst Elect & Elect Engineers Signal Proc Socen_US
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.rightsAtıf Gayri Ticari Türetilemez 4.0 Uluslararasıtr_TR
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectAcousticsen_US
dc.subjectEngineeringen_US
dc.subjectSpeaker identificationen_US
dc.subjectShouted speechen_US
dc.subjectLoudspeakersen_US
dc.subjectMappingen_US
dc.subjectSignal processingen_US
dc.subjectSpeechen_US
dc.subjectEmotional speechen_US
dc.subjectGaussian mixture modelen_US
dc.subjectIdentification accuracyen_US
dc.subjectMapping techniquesen_US
dc.subjectMel-frequency cepstral coefficientsen_US
dc.subjectRecognition accuracyen_US
dc.subjectSpeaker identificationen_US
dc.subjectText-independent speaker identificationen_US
dc.subjectSpeech recognitionen_US
dc.titleSpeaker identification from shouted speech: Analysis and compensationen_US
dc.typeProceedings Paperen_US
dc.identifier.wos000329611508038tr_TR
dc.identifier.scopus2-s2.0-84890452416tr_TR
dc.relation.publicationcategoryKonferans Öğesi - Uluslararasıtr_TR
dc.contributor.departmentUludağ Üniversitesi/Mühendislik Fakültesi/Elektrik-Elektronik Mühendisliği Bölümü.tr_TR
dc.identifier.startpage8027tr_TR
dc.identifier.endpage8031tr_TR
dc.relation.journalInternational Conference on Acoustics Speech and Signal Processing ICASSP, 2013 IEEE International Conference on Acoustics, Speech and Signal Processingen_US
dc.contributor.buuauthorHanilçi, Cemal-
dc.contributor.buuauthorErtaş, Figen-
dc.contributor.researcheridAAH-4188-2021tr_TR
dc.contributor.researcheridS-4967-2016tr_TR
dc.relation.collaborationYurt dışıtr_TR
dc.subject.wosAcousticsen_US
dc.subject.wosEngineering, electrical & electronicen_US
dc.indexed.wosCPCISen_US
dc.indexed.scopusScopusen_US
dc.contributor.scopusid35781455400tr_TR
dc.contributor.scopusid24724154500tr_TR
dc.subject.scopusWhispers; Speech Recognition; Public Speakingen_US
Appears in Collections:Scopus
Web of Science

Files in This Item:
File Description SizeFormat 
Hanilçi_vd_2013.pdf563.22 kBAdobe PDFThumbnail
View/Open


This item is licensed under a Creative Commons License Creative Commons