Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Facial attribute estimation and expression recognition based on contextual channel attention mechanism
Jie XU, Yong ZHONG, Yang WANG, Changfu ZHANG, Guanci YANG
Journal of Computer Applications    2025, 45 (1): 253-260.   DOI: 10.11772/j.issn.1001-9081.2024010098
Abstract266)   HTML6)    PDF (2220KB)(917)       Save

Facial features contain a lot of information and hold significant value in facial attribute and expression analysis tasks, but the diversity and complexity of facial features make facial analysis tasks difficult. Aiming at the above issue, a model of Facial Attribute estimation and Expression Recognition based on contextual channel attention mechanism (FAER) was proposed from the perspective of fine-grained facial features. Firstly, a local feature encoding backbone network based on ConvNext was constructed, and by utilizing the effectiveness of the backbone network in encoding local features, the differences among facial local features were represented adequately. Secondly, a Contextual Channel Attention (CC Attention) mechanism was introduced. By adjusting the weight information on feature channels dynamically and adaptively, both global and local features of deep features were represented, so as to address the limitations of the backbone network ability in encoding global features. Finally, different classification strategies were designed. For Facial Attribute Estimation (FAE) and Facial Expression Recognition (FER) tasks, different combinations of loss functions were employed to encourage the model to learn more fine-grained facial features. Experimental results show that the proposed model achieves an average accuracy of 91.87% on facial attribute dataset CelebA (CelebFaces Attributes), surpassing the suboptimal model SwinFace (Swin transformer for Face) by 0.55 percentage points, and the proposed model achieves accuracies of 91.75% and 66.66% respectively on facial expression datasets RAF-DB and AffectNet, surpassing the suboptimal model TransFER (Transformers for Facial Expression Recognition) by 0.84 and 0.43 percentage points respectively.

Table and Figures | Reference | Related Articles | Metrics
Research progress on motion segmentation of visual localization and mapping in dynamic environment
Dongying ZHU, Yong ZHONG, Guanci YANG, Yang LI
Journal of Computer Applications    2023, 43 (8): 2537-2545.   DOI: 10.11772/j.issn.1001-9081.2022070972
Abstract404)   HTML20)    PDF (2687KB)(256)       Save

Visual localization and mapping system is affected by dynamic objects in a dynamic environment, so that it has increase of localization and mapping errors and decrease of robustness. And motion segmentation of input images can significantly improve the performance of visual localization and mapping system in dynamic environment. Dynamic objects in dynamic environment can be divided into moving objects and potential moving objects. Current dynamic object recognition methods have problems of chaotic moving subjects and poor real-time performance. Therefore, motion segmentation strategies of visual localization and mapping system in dynamic environment were reviewed. Firstly, the strategies were divided into three types of methods according to preset conditions of the scene: methods based on static assumption of image subject, methods based on prior semantic knowledge and multi-sensor fusion methods without assumption. Then, these three types of methods were summarized, and their accuracy and real-time performance were analyzed. Finally, aiming at the difficulty of balancing accuracy and real-time performance of motion segmentation strategy of visual localization and mapping system in dynamic environment, development trends of the motion segmentation methods in dynamic environment were discussed and prospected.

Table and Figures | Reference | Related Articles | Metrics