This paper is published in Volume-7, Issue-1, 2021
Area
Information Technology
Author
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade
Org/Univ
Pune Institute of Computer Technology, Pune, Maharashtra, India
Keywords
Object Detection, Convolutional Neural Network, LSTM, Text-to-Speech
Citations
IEEE
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade. Audio detail generation of the environment in a surveillance system for blind people, International Journal of Advance Research, Ideas and Innovations in Technology, www.IJARIIT.com.
APA
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade (2021). Audio detail generation of the environment in a surveillance system for blind people. International Journal of Advance Research, Ideas and Innovations in Technology, 7(1) www.IJARIIT.com.
MLA
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade. "Audio detail generation of the environment in a surveillance system for blind people." International Journal of Advance Research, Ideas and Innovations in Technology 7.1 (2021). www.IJARIIT.com.
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade. Audio detail generation of the environment in a surveillance system for blind people, International Journal of Advance Research, Ideas and Innovations in Technology, www.IJARIIT.com.
APA
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade (2021). Audio detail generation of the environment in a surveillance system for blind people. International Journal of Advance Research, Ideas and Innovations in Technology, 7(1) www.IJARIIT.com.
MLA
Aryan Mishra, Tejas Mutyalwar, Kaustubh Kulkarni, Diwanshu Chouragade. "Audio detail generation of the environment in a surveillance system for blind people." International Journal of Advance Research, Ideas and Innovations in Technology 7.1 (2021). www.IJARIIT.com.
Abstract
Surveillance cameras are widely utilized as a resource to better control home environments such as monitoring children, the elderly, pets, and safety, among other applications. However, as it relates categorically to visual information, people who are visually impaired or blind are omitted due to the barriers that result from accessing information predicated on visual content. The purpose of this work is to propose a cumulation of deep learning techniques for the recognition of objects in video and to utilize an accessibility resource called audio-description to engender a narrative of detected information. The result is a video object narrative-based surveillance system that provides useful contextual information about the environment for blind and visually impaired people.