FACIAL EXPRESSION MULTI-VIEW RECOGNITION USING A NEURAL NETWORK-BASED ENSEMBLE MODEL
Abstract
Facial expression recognition (FER) plays a critical role in human- computer interaction, enabling systems to interpret and respond to human emotions effectively. Traditional FER methods struggle to handle variations in viewpoint, illumination, and facial occlusions, leading to reduced accuracy and robustness. A neural network-based ensemble approach is proposed to address these challenges by combining the strengths of multiple deep learning models to enhance multi-view facial expression recognition. The proposed method integrates Convolutional Neural Networks (CNNs) with a stacking ensemble mechanism, where individual CNN models are trained on different facial angles and lighting conditions. The outputs of these models are combined using a meta-classifier to improve recognition accuracy. The ensemble network is trained and tested on the Multi-PIE and BU-3DFE datasets, achieving an accuracy of 96.4%, outperforming existing single-model approaches. The proposed method demonstrates robustness across varying facial poses and occlusions, highlighting its potential for real-world applications in emotion-aware systems and interactive technologies.

Authors
Karthikeyan Thangavel, P. Sowmya Saraswathi, Saravanan Velusamy
University of Technology and Applied Sciences, The Sultanate of Oman

Keywords
Facial Expression Recognition, Multi-View, Ensemble Learning, CNN, Meta-Classifier
Yearly Full Views
JanuaryFebruaryMarchAprilMayJuneJulyAugustSeptemberOctoberNovemberDecember
000300000000
Published By :
ICTACT
Published In :
ICTACT Journal on Data Science and Machine Learning
( Volume: 6 , Issue: 2 , Pages: 747 - 750 )
Date of Publication :
March 2025
Page Views :
17
Full Text Views :
3

Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.