Katalog Plus
Bibliothek der Frankfurt UAS
Bald neuer Katalog: sichern Sie sich schon vorab Ihre persönlichen Merklisten im Nutzerkonto: Anleitung.
Dieses Ergebnis aus ERIC kann Gästen nicht angezeigt werden.  Login für vollen Zugriff.

Extracting Keywords from Images Using Deep Learning for the Visually Challenged

Title: Extracting Keywords from Images Using Deep Learning for the Visually Challenged
Language: English
Authors: Jaboob, Said; Chauhan, Munes Singh; Dhanasekaran, Balaji; Natarajan, Senthil Kumar
Source: International Society for Technology, Education, and Science. 2022.
Availability: International Society for Technology, Education, and Science. 944 Maysey Drive, San Antonio, TX 78227. Tel: 515-294-1075; Fax: 515-294-1003; email: istesoffice@gmail.com; Web site: http://www.istes.org
Peer Reviewed: Y
Page Count: 8
Publication Date: 2022
Document Type: Speeches/Meeting Papers; Reports - Descriptive
Descriptors: Information Retrieval; Visual Aids; Visual Impairments; Assistive Technology; Foreign Countries; Cognitive Ability; Memory
Geographic Terms: Oman
Abstract: Assistive technologies can in many ways facilitate the normal day-to-day lives of the disabled. As part of the ongoing research on assistive technologies at UTAS, Oman, that deals with augmenting and finding multimodal aspects of applications for the disabled, this paper aspires to investigate the role of deep learning in the field of image interpretation. Images are one of the most important mediums of conveying information among humans. Visually impaired persons especially with low cognitive abilities face insurmountable difficulties in understanding cues through images. This challenge is met by filtering words from image captions to facilitate understanding of the key notion conveyed by an image. This work utilizes the image captioning technique using deep learning frameworks such as convolution neural networks (CNN) and recurrent neural networks (RNN) to generate captions. These captions are fed to Rake, an NLP library that identifies keywords in the caption. The entire process is automated and uses transfer learning techniques for caption generation from images. This process is then further integrated with our main project, Finger Movement Multimodal Assistive System (FMAS) thereby incorporating text cues for interpreting images for the visually impaired. [For the full proceedings, see ED630948.]
Abstractor: As Provided
Entry Date: 2023
Accession Number: ED630963
Database: ERIC