Ventral-Dorsal Attention Capsule Network for facial expression recognition
Recent deep learning-based approaches on facial expression recognition face challenges on achieving optimal solution due to ineffective extraction of semantic expression features and difficulty in perceiving the positional relationship among facial expression features. In this paper we propose a novel framework, named Ventral-Dorsal Attention Capsule Network (VDACaps) to address the challenges. VDACaps adopts ResNet18 with a ventral channel attention and dilated spatial attention mechanism to strengthen the feature extraction. A novel attention layer containing multiple groups of ventral and dorsal attention is also designed to focus on the key features of facial expressions at the channel and spatial levels, exploring optimal sizes of receptive fields. Experiments on three well-known public expression datasets (CK+, JAFFE, and SFEW) have demonstrated that VDACaps achieves better or competitive performance on facial expression recognition against the state-of-the-art, with the accuracy of 98.98%, 98.44% and 54.07% respectively.
Duke Scholars
Published In
DOI
ISSN
Publication Date
Volume
Related Subject Headings
- Networking & Telecommunications
- 46 Information and computing sciences
- 40 Engineering
- 1005 Communications Technologies
- 0913 Mechanical Engineering
- 0906 Electrical and Electronic Engineering
Citation
Published In
DOI
ISSN
Publication Date
Volume
Related Subject Headings
- Networking & Telecommunications
- 46 Information and computing sciences
- 40 Engineering
- 1005 Communications Technologies
- 0913 Mechanical Engineering
- 0906 Electrical and Electronic Engineering