B1 中級 147 タグ追加 保存
動画の字幕をクリックしてすぐ単語の意味を調べられます!
単語帳読み込み中…
字幕の修正報告
Hello, I'm Joy, a poet of code,
on a mission to stop an unseen force that's rising,
a force that I called "the coded gaze,"
my term for algorithmic bias.
Algorithmic bias, like human bias, results in unfairness.
However, algorithms, like viruses, can spread bias on a massive scale
at a rapid pace.
Algorithmic bias can also lead to exclusionary experiences
and discriminatory practices.
Let me show you what I mean.
(Video) Joy Buolamwini: Hi, camera. I've got a face.
Can you see my face?
No-glasses face?
You can see her face.
What about my face?
I've got a mask. Can you see my mask?
Joy Buolamwini: So how did this happen?
Why am I sitting in front of a computer
in a white mask,
trying to be detected by a cheap webcam?
Well, when I'm not fighting the coded gaze
as a poet of code,
I'm a graduate student at the MIT Media Lab,
and there I have the opportunity to work on all sorts of whimsical projects,
including the Aspire Mirror,
a project I did so I could project digital masks onto my reflection.
So in the morning, if I wanted to feel powerful,
I could put on a lion.
If I wanted to be uplifted, I might have a quote.
So I used generic facial recognition software
to build the system,
but found it was really hard to test it unless I wore a white mask.
Unfortunately, I've run into this issue before.
When I was an undergraduate at Georgia Tech studying computer science,
I used to work on social robots,
and one of my tasks was to get a robot to play peek-a-boo,
a simple turn-taking game
where partners cover their face and then uncover it saying, "Peek-a-boo!"
The problem is, peek-a-boo doesn't really work if I can't see you,
and my robot couldn't see me.
But I borrowed my roommate's face to get the project done,
submitted the assignment,
and figured, you know what, somebody else will solve this problem.
Not too long after,
I was in Hong Kong for an entrepreneurship competition.
The organizers decided to take participants
on a tour of local start-ups.
One of the start-ups had a social robot,
and they decided to do a demo.
The demo worked on everybody until it got to me,
and you can probably guess it.
It couldn't detect my face.
I asked the developers what was going on,
and it turned out we had used the same generic facial recognition software.
Halfway around the world,
I learned that algorithmic bias can travel as quickly
as it takes to download some files off of the internet.
So what's going on? Why isn't my face being detected?
Well, we have to look at how we give machines sight.
Computer vision uses machine learning techniques
to do facial recognition.
So how this works is, you create a training set with examples of faces.
This is a face. This is a face. This is not a face.
And over time, you can teach a computer how to recognize other faces.
However, if the training sets aren't really that diverse,
any face that deviates too much from the established norm
will be harder to detect,
which is what was happening to me.
But don't worry -- there's some good news.
Training sets don't just materialize out of nowhere.
We actually can create them.
So there's an opportunity to create full-spectrum training sets
that reflect a richer portrait of humanity.
Now you've seen in my examples
how social robots
was how I found out about exclusion with algorithmic bias.
But algorithmic bias can also lead to discriminatory practices.
Across the US,
police departments are starting to use facial recognition software
in their crime-fighting arsenal.
Georgetown Law published a report
showing that one in two adults in the US -- that's 117 million people --
have their faces in facial recognition networks.
Police departments can currently look at these networks unregulated,
using algorithms that have not been audited for accuracy.
Yet we know facial recognition is not fail proof,
and labeling faces consistently remains a challenge.
You might have seen this on Facebook.
My friends and I laugh all the time when we see other people
mislabeled in our photos.
But misidentifying a suspected criminal is no laughing matter,
nor is breaching civil liberties.
Machine learning is being used for facial recognition,
but it's also extending beyond the realm of computer vision.
In her book, "Weapons of Math Destruction,"
data scientist Cathy O'Neil talks about the rising new WMDs --
widespread, mysterious and destructive algorithms
that are increasingly being used to make decisions
that impact more aspects of our lives.
So who gets hired or fired?
Do you get that loan? Do you get insurance?
Are you admitted into the college you wanted to get into?
Do you and I pay the same price for the same product
purchased on the same platform?
Law enforcement is also starting to use machine learning
for predictive policing.
Some judges use machine-generated risk scores to determine
how long an individual is going to spend in prison.
So we really have to think about these decisions.
Are they fair?
And we've seen that algorithmic bias
doesn't necessarily always lead to fair outcomes.
So what can we do about it?
Well, we can start thinking about how we create more inclusive code
and employ inclusive coding practices.
It really starts with people.
So who codes matters.
Are we creating full-spectrum teams with diverse individuals
who can check each other's blind spots?
On the technical side, how we code matters.
Are we factoring in fairness as we're developing systems?
And finally, why we code matters.
We've used tools of computational creation to unlock immense wealth.
We now have the opportunity to unlock even greater equality
if we make social change a priority
and not an afterthought.
And so these are the three tenets that will make up the "incoding" movement.
Who codes matters,
how we code matters
and why we code matters.
So to go towards incoding, we can start thinking about
building platforms that can identify bias
by collecting people's experiences like the ones I shared,
but also auditing existing software.
We can also start to create more inclusive training sets.
Imagine a "Selfies for Inclusion" campaign
where you and I can help developers test and create
more inclusive training sets.
And we can also start thinking more conscientiously
about the social impact of the technology that we're developing.
To get the incoding movement started,
I've launched the Algorithmic Justice League,
where anyone who cares about fairness can help fight the coded gaze.
On codedgaze.com, you can report bias,
request audits, become a tester
and join the ongoing conversation,
#codedgaze.
So I invite you to join me
in creating a world where technology works for all of us,
not just some of us,
a world where we value inclusion and center social change.
Thank you.
(Applause)
But I have one question:
Will you join me in the fight?
(Laughter)
(Applause)
コツ:単語をクリックしてすぐ意味を調べられます!

読み込み中…

【TED】ジョイ・ブォロムウィニ: アルゴリズムに潜む偏見との戦い (How I'm fighting bias in algorithms | Joy Buolamwini)

147 タグ追加 保存
Zenn 2017 年 11 月 12 日 に公開
お勧め動画
  1. 1. クリック一つで単語を検索

    右側のスプリクトの単語をクリックするだけで即座に意味が検索できます。

  2. 2. リピート機能

    クリックするだけで同じフレーズを何回もリピート可能!

  3. 3. ショートカット

    キーボードショートカットを使うことによって勉強の効率を上げることが出来ます。

  4. 4. 字幕の表示/非表示

    日・英のボタンをクリックすることで自由に字幕のオンオフを切り替えられます。

  5. 5. 動画をブログ等でシェア

    コードを貼り付けてVoiceTubeの動画再生プレーヤーをブログ等でシェアすることが出来ます!

  6. 6. 全画面再生

    左側の矢印をクリックすることで全画面で再生できるようになります。

  1. クイズ付き動画

    リスニングクイズに挑戦!

  1. クリックしてメモを表示

  1. UrbanDictionary 俚語字典整合查詢。一般字典查詢不到你滿意的解譯,不妨使用「俚語字典」,或許會讓你有滿意的答案喔