Placeholder Image

字幕表 動画を再生する

  • Horus is a wearable device for blind and vision-impaired people. It's

  • basically a headset with cameras and a powerful pocket-computer that stays in

  • the pocket and processes all the images in real-time. One day, we were in Genova just

  • in front of the train station there. We met a blind person and he begged for help to

  • reach the bus stop. And that event made us wonder all the difficulties that

  • people who have lost sight or are losing it have to face every day. And we discovered

  • that we could move our focus not from computer vision to another topic,

  • but bringing vision from robotics to people who need it, not to robots. We have a bone

  • conduction headset plus a stereo-vision system and a processing unit, which is

  • powered by the NVIDIA Jetson platform. The headset basically acquires images and

  • streams them to the processor where the deep learning and computer vision

  • algorithms run. The current status of our project is of advanced prototype. We've

  • made several iterations and we now have a custom board with a Tegra processor

  • and our own software comprising several functionalities. The way it's being

  • used today is we have a pool of testers and organizations that are helping us

  • testing the device, the packaging, and everything. We also have one blind

  • employee who is testing it every day.

  • We're working on face recognition and there we are working both on face

  • detection and also online learning, to allow the blind person to learn

  • the faces of new people on the go

  • while he is using the device. Then we have object recognition which basically works

  • in a very similar way but the underlying algorithms are a little different.

  • Then we have text reading: [Female voice] "I cannot fully express my gratitude to the exceptional team at

  • Doubleday" and by text reading, I mean that the device can recognize printed text,

  • for example, on books. And the device can also help the blind

  • person in moving around the city or also indoor spaces. In order to do that, we're

  • leveraging the power of stereovision. We're trying to understand what's in

  • front of the person like obstacles, signs, and then creating some sort of

  • audible map of the environment for the blind person so that, when he moves around he can

  • understand what's in front of him and then avoid any obstacles that he might find on the way.

  • We have scene captioning. The person queries the device through the press

  • of a button for a description of what the cameras are seeing and then to

  • generate a complete sentence that tries to describe what's in front of the blind person.

  • All of it is obtained through deep learning. [Female Voice] "A group of people sitting around the

  • table with laptops." We make extensive use of deep learning and computer vision

  • in our product because if you want to be useful you need to gather all the

  • possible information from the surrounding environment, and using only classical

  • computer vision approaches, that's not possible.

  • Our work is also to bring the same algorithms and make them run on our

  • mobile platform. This is made possible by NVIDIA's Jetson platform

  • because it allows for high performance and also power-efficient computing in a

  • small form factor. When we release our product, our goal is to basically

  • radically change their lives, because we want to make the world more accessible to them.

  • We cannot change the accessibility of every city, in every country, in every nation, but

  • we can create something that can bridge a gap, can make what's not accessible,

  • accessible to everybody.

Horus is a wearable device for blind and vision-impaired people. It's

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

B1 中級

NVIDIA Jetsonパートナー体験談。ホルス (NVIDIA Jetson Partner Stories: Horus)

  • 85 7
    alex に公開 2021 年 01 月 14 日
動画の中の単語