Facial emotion recognition (FER) has gained increasing attention in human–computer interaction and affective computing; however, existing methods often suffer from high computational cost and limited generalization, especially in real-world scenarios with subtle expressions and noisy inputs. To cope with these issues, this study proposes a knowledge distillation-based framework for FER. The teacher network utilizes a parallel architecture of NasNetMobile and MobileNet as dual backbones for comprehensive feature extraction, further, these features are enhanced by a Deformable Attention (DA) module, which primarily focuses spatial feature representations. To transfer this rich knowledge effectively, we introduce a lightweight student model, TinyNasNet, inspired by the internal architecture of NasNetMobile. In this framework, the student model is trained to mimic the behavior of the teacher network, aiming to achieve higher performance while maintaining computational complexity. Moreover, extensive experiments were conducted over two benchmarks, such as FER and KDEF. In contrast, the proposed network offers higher performance compared to various competitive networks, demonstrating a highly efficient yet robust solution for real-time FER.
목차
Abstract 1. Introduction 2. Proposed Method 3. Results and Discussion 3.1. Comparative analysis of various techniques 5. Conclusion Acknowledgement Reference
저자
Taimoor Khan [ Department of Computer Engineering, IT Convergence, Gachon University ]
Abdul Hai Karimi [ Department of Computer Engineering, IT Convergence, Gachon University ]
Chang Choi [ Department of Computer Engineering, IT Convergence, Gachon University ]
교신저자