Facial Emotion Recognition (FER) has become an area of research in human-computer interaction and affective computing. This study presents a deep learning-based FER framework that can recognize facial emotions in both controlled and real-world environments. The proposed method uses an architecture that brings together VGG19 and ResNet152V2 to make use of their different feature learning strengths. VGG19 helps to pick up spatial and texture information while ResNet152V2 focuses on pulling out deeper semantic representations through residual learning. A full preprocessing pipeline, including normalization, resizing and techniques like rotation and flipping is used to improve robustness against changes in lighting pose variations and class imbalance. The effective framework is tested using two datasets, CK+ and FER2013. The experimental results show that the combined model consistently outperforms baseline architectures. On CK+ it achieves a classification accuracy of 98.0%. On FER2013 it achieves 92.0% along with good precision, recall and F1-score. The results show that combining features at the feature level greatly improves the framework ability to generalize. This makes the proposed framework suitable for FER applications and can be used in various applications.
Keywords
Deep Learning, Facial Emotion Recognition, Transfer Learning, Computer Vision, VGG-19, ResNet152V2.