Welcome to the UPF Digital Repository

Audio, video and multimodal person identification in a smart room

Show simple item record

dc.contributor.author Luque, Jordi
dc.contributor.author Morros, R.
dc.contributor.author Garde, I.
dc.contributor.author Anguita, Jan
dc.contributor.author Farrús, Mireia
dc.contributor.author Macho, D.
dc.contributor.author Marqués López, Fernando
dc.contributor.author Martínez, C.
dc.contributor.author Vilaplana, V.
dc.contributor.author Hernando, Javier
dc.date.accessioned 2017-09-04T16:02:39Z
dc.date.available 2017-09-04T16:02:39Z
dc.date.issued 2007
dc.identifier.citation Luque J, Morros R, Garde I, Anguita J, Farrús M, Macho D, Marqués F, Martínez C, Vilaplana V, Hernando J. Audio, video and multimodal person identification in a smart room. In: Stiefelhagen R, Garofolo J, editors. Multimodal technologies for perception of humans: first International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006; 2006 Apr. 6-7; Southampton (UK). Germany: Springer; 2007. p. 258-69. (LNCS; no. 4122). DOI: 10.1007/978-3-540-69568-4_23
dc.identifier.issn 0302-9743
dc.identifier.uri http://hdl.handle.net/10230/32740
dc.description Comunicació presentada a: The First International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006, celebrat a Southampton, Regne Unit, el 6 i 7 d'abril de 2006.
dc.description.abstract In this paper, we address the modality integration issue on the example of a smart room environment aiming at enabling person identification by combining speech and 2D face images. First we introduce the monomodal audio and video identification techniques and then we present the use of combined input speech and face images for person identification. The various sensory modalities, speech and faces, are processed both individually and jointly. It’s shown that the multimodal approach results in improved performance in the identification of the participants.
dc.description.sponsorship This work has been partially sponsored by the EC-funded project CHIL (IST-2002-506909) and by the Spanish Government-funded project ACESCA (TIN2005-08852).
dc.format.mimetype application/pdf
dc.language.iso eng
dc.publisher Springer
dc.relation.ispartof Stiefelhagen R, Garofolo J, editors. Multimodal technologies for perception of humans: first International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006; 2006 Apr. 6-7; Southampton (UK). Germany: Springer; 2007. p. 258-69.
dc.rights © Springer The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-540-69568-4_23
dc.title Audio, video and multimodal person identification in a smart room
dc.type info:eu-repo/semantics/conferenceObject
dc.identifier.doi http://dx.doi.org/10.1007/978-3-540-69568-4_23
dc.subject.keyword Multimodality
dc.subject.keyword Speaker recognition
dc.relation.projectID info:eu-repo/grantAgreement/EC/FP6/506909
dc.relation.projectID info:eu-repo/grantAgreement/ES/2PN/TIN2005-08852
dc.rights.accessRights info:eu-repo/semantics/openAccess
dc.type.version info:eu-repo/semantics/acceptedVersion


This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account

Statistics

Compliant to Partaking