This paper proposes a wearable system for visually impaired people that can be utilized to obtain an extensive feedback about their surrounding environment. Our system consists of a stereo camera and smart glasses, communicating with a smartphone that is used as an intermediary computational device. Furthermore, the system is connected to a server where all the expensive computations are executed. The whole setup is capable of detecting obstacles in the nearest surrounding, recognizing faces and facial expressions, reading texts, providing a generic description and question answering of a particular input image. In addition, we propose a novel depth question answering system to estimate object size as well as objects relative position in an unconstrained environment in near real-time and in a fully automatic way requiring only stereo image pair and voice request as an input. We have conducted an extensive number of experiments to evaluate the feasibility and practicality of the proposed system which shows promising results to assist visually impaired people.
조회 수 420 댓글 0
|저 자||Oleksandr Bogdan, Oleg Yurchenko, Oleksandr Bailo, Francois Rameau, Donggeun Yoo, In So Kweon|
|학 회||The 8th Pacific Rim Symposium on Image and Video Technology (PSIVT)|
Prev Co-domain Embedding using Deep Quadruplet Network for Unseen ... Co-domain Embedding using Deep Quadruplet Network for Unseen ... 2017.11.08by VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition Next VPGNet: Vanishing Point Guided Network for Lane and Road Marking Detection and Recognition 2017.08.10by 이석주