RETRACTED ARTICLE: Local Directional Maximum Edge Patterns for facial expression recognition
Autor: | G. Varaprasad, V. Uma Maheswari, S. Viswanadha Raju |
---|---|
Rok vydání: | 2020 |
Předmět: |
Facial expression
General Computer Science Pixel Biometrics Computer science business.industry Orientation (computer vision) Emotion classification Feature extraction ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION 020207 software engineering Pattern recognition Computational intelligence 02 engineering and technology Feature (computer vision) Face space 0202 electrical engineering electronic engineering information engineering Feature descriptor 020201 artificial intelligence & image processing Artificial intelligence business |
Zdroj: | Journal of Ambient Intelligence and Humanized Computing. 12:4775-4783 |
ISSN: | 1868-5145 1868-5137 |
Popis: | Cognitive science and neuroscience use human facial expressions of emotion. Every single facial expression can be seen at different passions in a face space. Nowadays, facial expression recognition and analysis is vital due to the demand of introducing advanced biometric applications in every domain space. The imperative task in facial expressions of emotion classification is precise feature extraction, which helps to get detailed description of facial marks. Existing feature descriptors are suffering from various problems such as intensity variations, discrimination, vulnerability etc. In this paper, propose a new feature descriptor method called LDMEP (Local Directional Maximum Edge Patterns) for facial expression analysis to overcome the hindrance. We calculated the gradients in four directions of reference pixel to elicit the more feature for better recognition instead of calculating the local differences among neighboring pixels. We also access the orientations of the pixels then thresholded based on the dynamic threshold to avoid the featureless area calculation. Furthermore, we considered only dominant magnitude and orientation directions instead of all eight directions to generate feature. Thus, imperative and efficient features are covered in dominant positions to detect the strong edges. The paper confers that how the subsequent model can be used for the recognition of facial expression of emotion. |
Databáze: | OpenAIRE |
Externí odkaz: |