Skip to yearly menu bar Skip to main content


Poster
in
Workshop: 5th Workshop on practical ML for limited/low resource settings (PML4LRS) @ ICLR 2024

FacePhi: Lightweight Multimodal Large Language Model for Facial Landmark Emotion Recognition

Hongjin Zhao · Zheyuan Liu · Yang Liu · Zhenyue Qin · Jiaxu Liu · Tom Gedeon


Abstract:

We introduce FacePhi, a multimodal large language model (LLM) for emotion recognition through facial landmarks. By focusing on facial landmarks, FacePhi ensures privacy preservation in emotion detection tasks. FacePhi is optimized for computational efficiency by incorporating Phi-2, a LLM with a small number of parameters, as well as utilizing lightweight facial landmark data. This design choice makes FacePhi suitable for deployment in resource-constrained settings. Our investigation highlights the importance of feature alignment during the training phase, indicating its pivotal role in enhancing the model's performance for the challenging task of facial landmark emotion recognition.

Chat is not available.