Audio, video and multimodal person identification in a smart room

dc.contributor.authorLuque, Jordica
dc.contributor.authorMorros, R.ca
dc.contributor.authorGarde, I.ca
dc.contributor.authorAnguita, Janca
dc.contributor.authorFarrús, Mireiaca
dc.contributor.authorMacho, D.ca
dc.contributor.authorMarqués López, Fernandoca
dc.contributor.authorMartínez, C.ca
dc.contributor.authorVilaplana, Verónicaca
dc.contributor.authorHernando, Javierca
dc.date.accessioned2017-09-04T16:02:39Z
dc.date.available2017-09-04T16:02:39Z
dc.date.issued2007
dc.descriptionComunicació presentada a: The First International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006, celebrat a Southampton, Regne Unit, el 6 i 7 d'abril de 2006.ca
dc.description.abstractIn this paper, we address the modality integration issue on the example of a smart room environment aiming at enabling person identification by combining speech and 2D face images. First we introduce the monomodal audio and video identification techniques and then we present the use of combined input speech and face images for person identification. The various sensory modalities, speech and faces, are processed both individually and jointly. It’s shown that the multimodal approach results in improved performance in the identification of the participants.en
dc.description.sponsorshipThis work has been partially sponsored by the EC-funded project CHIL (IST-2002-506909) and by the Spanish Government-funded project ACESCA (TIN2005-08852).en
dc.format.mimetypeapplication/pdf
dc.identifier.citationLuque J, Morros R, Garde I, Anguita J, Farrús M, Macho D, Marqués F, Martínez C, Vilaplana V, Hernando J. Audio, video and multimodal person identification in a smart room. In: Stiefelhagen R, Garofolo J, editors. Multimodal technologies for perception of humans: first International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006; 2006 Apr. 6-7; Southampton (UK). Germany: Springer; 2007. p. 258-69. (LNCS; no. 4122). DOI: 10.1007/978-3-540-69568-4_23
dc.identifier.doihttp://dx.doi.org/10.1007/978-3-540-69568-4_23
dc.identifier.issn0302-9743
dc.identifier.urihttp://hdl.handle.net/10230/32740
dc.language.isoeng
dc.publisherSpringerca
dc.relation.ispartofStiefelhagen R, Garofolo J, editors. Multimodal technologies for perception of humans: first International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006; 2006 Apr. 6-7; Southampton (UK). Germany: Springer; 2007. p. 258-69.
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/FP6/506909
dc.relation.projectIDinfo:eu-repo/grantAgreement/ES/2PN/TIN2005-08852
dc.rights© Springer The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-540-69568-4_23
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess
dc.subject.keywordMultimodalityen
dc.subject.keywordSpeaker recognitionen
dc.titleAudio, video and multimodal person identification in a smart roomca
dc.typeinfo:eu-repo/semantics/conferenceObject
dc.type.versioninfo:eu-repo/semantics/acceptedVersion

Fitxers

Paquet original

Mostrant 1 - 1 de 1
Carregant...
Miniatura
Nom:
farrus_CLEAR2006_audi.pdf
Mida:
335.56 KB
Format:
Adobe Portable Document Format