Multimodal unconstrained people recognition with face and ear images using deep learning

Descripción del Articulo

Multibiometric systems rely on the idea of combining multiple biometric methods into one single process that leads to a more reliable and accurate system. The combination of two different biometric traits such as face and ear results in an advantageous and complementary process when using 2D images...

Descripción completa

Detalles Bibliográficos
Autor: Ramos Cooper, Solange Griselly
Formato: tesis de maestría
Fecha de Publicación:2023
Institución:Universidad Católica San Pablo
Repositorio:UCSP-Institucional
Lenguaje:inglés
OAI Identifier:oai:repositorio.ucsp.edu.pe:20.500.12590/17819
Enlace del recurso:https://hdl.handle.net/20.500.12590/17819
Nivel de acceso:acceso abierto
Materia:Multibiometric system
Multimodal recognition
Face recognition
Ear recognition
Feature-level fusión
Score-level fusión
Two-stream CNN.
https://purl.org/pe-repo/ocde/ford#1.02.01
Descripción
Sumario:Multibiometric systems rely on the idea of combining multiple biometric methods into one single process that leads to a more reliable and accurate system. The combination of two different biometric traits such as face and ear results in an advantageous and complementary process when using 2D images taken under uncontrolled conditions. In this work, we investigate several approaches to fuse information from the face and ear images to recognize people in a more accurate manner than using each method separately. We leverage the research maturity level of the face recognition field to build, first a truly multimodal database of ear and face images called VGGFace-Ear dataset, second a model that can describe ear images with high generalization called VGGEar model, and finally explore fusion strategies at two different levels in a common recognition pipeline, feature and score levels. Experiments on the UERC dataset have shown, first of all, an improvement of around 7% compared to the state-of-the-art methods in the ear recognition field. Second, fusing information from the face and ear images increases recognition rates from 79% and 82%, in the unimodal face and ear recognition respectively, to 94% recognition rate using the Rank-1 metric.
Nota importante:
La información contenida en este registro es de entera responsabilidad de la institución que gestiona el repositorio institucional donde esta contenido este documento o set de datos. El CONCYTEC no se hace responsable por los contenidos (publicaciones y/o datos) accesibles a través del Repositorio Nacional Digital de Ciencia, Tecnología e Innovación de Acceso Abierto (ALICIA).