Extracting Keywords from Images Using Deep Learning for the Visually Challenged.

Autor: Jaboob, Said, Chauhan, Munes Singh, Dhanasekaran, Balaji, Natarajan, Senthil Kumar
Předmět:
Zdroj: Proceedings of International Conference on Studies in Education & Social Sciences (ICSES); 11/10/2023, Vol. 1, p554-561, 8p
Abstrakt: Assistive technologies can in many ways facilitate the normal day-to-day lives of the disabled. As part of the ongoing research on assistive technologies at UTAS, Oman, that deals with augmenting and finding multimodal aspects of applications for the disabled, this paper aspires to investigate the role of deep learning in the field of image interpretation. Images are one of the most important mediums of conveying information among humans. Visually impaired persons especially with low cognitive abilities face insurmountable difficulties in understanding cues through images. This challenge is met by filtering words from image captions to facilitate understanding of the key notion conveyed by an image. This work utilizes the image captioning technique using deep learning frameworks such as convolution neural networks (CNN) and recurrent neural networks (RNN) to generate captions. These captions are fed to Rake, an NLP library that identifies keywords in the caption. The entire process is automated and uses transfer learning techniques for caption generation from images. This process is then further integrated with our main project, Finger Movement Multimodal Assistive System (FMAS) thereby incorporating text cues for interpreting images for the visually impaired. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index