Subtitles section Play video Print subtitles Horus is a wearable device for blind and vision-impaired people. It's basically a headset with cameras and a powerful pocket-computer that stays in the pocket and processes all the images in real-time. One day, we were in Genova just in front of the train station there. We met a blind person and he begged for help to reach the bus stop. And that event made us wonder all the difficulties that people who have lost sight or are losing it have to face every day. And we discovered that we could move our focus not from computer vision to another topic, but bringing vision from robotics to people who need it, not to robots. We have a bone conduction headset plus a stereo-vision system and a processing unit, which is powered by the NVIDIA Jetson platform. The headset basically acquires images and streams them to the processor where the deep learning and computer vision algorithms run. The current status of our project is of advanced prototype. We've made several iterations and we now have a custom board with a Tegra processor and our own software comprising several functionalities. The way it's being used today is we have a pool of testers and organizations that are helping us testing the device, the packaging, and everything. We also have one blind employee who is testing it every day. We're working on face recognition and there we are working both on face detection and also online learning, to allow the blind person to learn the faces of new people on the go while he is using the device. Then we have object recognition which basically works in a very similar way but the underlying algorithms are a little different. Then we have text reading: [Female voice] "I cannot fully express my gratitude to the exceptional team at Doubleday" and by text reading, I mean that the device can recognize printed text, for example, on books. And the device can also help the blind person in moving around the city or also indoor spaces. In order to do that, we're leveraging the power of stereovision. We're trying to understand what's in front of the person like obstacles, signs, and then creating some sort of audible map of the environment for the blind person so that, when he moves around he can understand what's in front of him and then avoid any obstacles that he might find on the way. We have scene captioning. The person queries the device through the press of a button for a description of what the cameras are seeing and then to generate a complete sentence that tries to describe what's in front of the blind person. All of it is obtained through deep learning. [Female Voice] "A group of people sitting around the table with laptops." We make extensive use of deep learning and computer vision in our product because if you want to be useful you need to gather all the possible information from the surrounding environment, and using only classical computer vision approaches, that's not possible. Our work is also to bring the same algorithms and make them run on our mobile platform. This is made possible by NVIDIA's Jetson platform because it allows for high performance and also power-efficient computing in a small form factor. When we release our product, our goal is to basically radically change their lives, because we want to make the world more accessible to them. We cannot change the accessibility of every city, in every country, in every nation, but we can create something that can bridge a gap, can make what's not accessible, accessible to everybody.
B1 US blind vision device computer vision headset nvidia NVIDIA Jetson Partner Stories: Horus 85 7 alex posted on 2017/04/15 More Share Save Report Video vocabulary