Update cookies preferences
 logo Idiap Research Institute        
 [BibTeX] [Marc21]
Benchmarking Multimodal Large Language Models for Face Recognition
Type of publication: Conference paper
Citation: OtroshiShahreza_ICASSP_2026
Publication status: Accepted
Booktitle: Proceedings of the 2025 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)
Year: 2026
URL: https://www.idiap.ch/paper/fac...
Abstract: Multimodal large language models (MLLMs) have achieved remarkable performance across diverse vision-and-language tasks. However, their potential in face recognition remains underexplored. In particular, the performance of open-source MLLMs needs to be evaluated and compared with existing face recognition models on standard benchmarks with similar protocol. In this work, we present a systematic benchmark of state-of-the-art MLLMs for face recognition on several face recognition datasets, including LFW, CALFW, CPLFW, CFP, AgeDB and RFW. Experimental results reveal that while MLLMs capture rich semantic cues useful for face-related tasks, they lag behind specialized models in high-precision recognition scenarios in zero-shot applications. This benchmark provides a foundation for advancing MLLM-based face recognition, offering insights for the design of next-generation models with higher accuracy and generalization. The source code of our benchmark is publicly available in the project page.
Main Research Program: Sustainable & Resilient Societies
Additional Research Programs: AI for Everyone
Keywords: Benchmark, Face Recognition, Foundation Models, Multimodal Large Language Models (MLLMs)
Projects: SAFER
CARMEN
Authors: Otroshi Shahreza, Hatef
Marcel, Sébastien
Added by: [UNK]
Total mark: 0
Attachments
  • OtroshiShahreza_ICASSP_2026.pdf
Notes