Cathy O'Neil: The era of blind faith in big data must end
凯西·奥尼尔: 盲目信仰大数据的时代必须结束
Data skeptic Cathy O’Neil uncovers the dark secrets of big data, showing how our "objective" algorithms could in fact reinforce human bias. Full bio
Double-click the English transcript below to play the video.
the winners from the losers.
that we don't understand
and often hoping for.
by looking, figuring out.
what is associated with success.
in written code.
为我的家人做饭。
to make a meal for my family.
of ramen noodles as food.
这顿饭就是成功的。
if my kids eat vegetables.
from if my youngest son were in charge.
负责做饭时的情况有所不同。
Nutella巧克力榛子酱就是成功。
he gets to eat lots of Nutella.
most people think of algorithms.
算法的看法是不同的。
and true and scientific.
blind faith in big data.
很多人都可能犯错。
She's a high school principal in Brooklyn.
她是布鲁克林的一名高中校长。
她学校的老师们正在被一个复杂
her teachers were being scored
what the formula is, show it to me.
公式是什么,然后给我看看。
to get the formula,
told me it was math
a Freedom of Information Act request,
and all their scores
发表了这些数据。
as an act of teacher-shaming.
the source code, through the same means,
源代码的时候,
had access to that formula.
got involved, Gary Rubenstein.
加里·鲁宾斯坦。
from that New York Post data
找到了665名教师,
for individual assessment.
with 205 other teachers,
recommendations from her principal
of you guys are thinking,
人工智能专家。
the AI experts here.
这样前后矛盾的算法。”
an algorithm that inconsistent."
也会产生毁灭性的影响。
with good intentions.
that's designed badly
silently wreaking havoc.
并无声地造成破坏。
about sexual harassment.
福克斯新闻有所成就。
to succeed at Fox News.
but we've seen recently
to turn over another leaf?
their hiring process
是一个合理的选择。
21 years of applications to Fox News.
stayed there for four years
学习是什么造就了成功,
to learn what led to success,
historically led to success
to a current pool of applicants.
who were successful in the past.
在过去取得成功的人。
盲目地应用算法。
blindly apply algorithms.
if we had a perfect world,
完美的世界那就太好了,
大多数公司都没有令人尴尬的诉讼,
don't have embarrassing lawsuits,
这意味着他们可以编纂性别歧视
it means they could be codifying sexism
all neighborhoods
所有的社区,
only to the minority neighborhoods
we found the data scientists
下一起犯罪会发生在哪里?
where the next crime would occur?
criminal would be?
about how great and how accurate
会吹嘘他们的模型有多好,
but we do have severe segregations
但我们确实在许多城市里
and justice system data.
the individual criminality,
recently looked into
宣判期间被法官采用。
during sentencing by judges.
was scored a 10 out of 10.
10分中得了满分。
10分中得了3分。
3分代表低风险。
3 out of 10, low risk.
for drug possession.
而被带进了监狱。
the higher score you are,
a longer sentence.
technologists hide ugly truths
把丑陋真相隐藏在
important and destructive,
重要的并具有破坏性的,
and it's not a mistake.
building private algorithms
for teachers and the public police,
与公共警察使用的(算法),
the authority of the inscrutable.
since all this stuff is private
will solve this problem.
有很多钱可以赚。
to be made in unfairness.
in ways that we wish we weren't,
虽然我们希望我们不是,
have consistently demonstrated this
of applications to jobs out,
有些用白人人名,
have white-sounding names
the results -- always.
into the algorithms
about ramen noodles --
在过去的实践中获得的数据
picking up on past practices
to emerge unscathed?
会是毫无瑕疵的呢?
we can check them for fairness.
the truth every time.
We can make them better.
我们可以让他们变得更好。
algorithm I talked about,
we'd have to come to terms with the fact
我们不得不接受这个事实,
smoke pot at the same rate
吸毒的比例是一样的,
to be arrested --
depending on the area.
in other crime categories,
是什么样子的,
the definition of success,
algorithm? We talked about it.
然后被提升了一次?
and is promoted once?
that is supported by their culture.
这可能会有很大的偏差。
the blind orchestra audition
are behind a sheet.
决定了什么是重要的,
have decided what's important
什么是不重要的,
distracted by that.
auditions started,
女性的数量上升了5倍。
went up by a factor of five.
for teachers would fail immediately.
立刻失效的地方。
the errors of every algorithm.
and for whom does this model fail?
让这个模型失败的对象是谁?
如果脸书的工程师们之前考虑过,
had considered that
only things that our friends had posted.
我们朋友所发布的东西。
one for the data scientists out there.
一条是给数据科学家的。
not be the arbiters of truth.
成为真相的仲裁者。
of ethical discussions that happen
所发生的道德讨论的
for our algorithmic overlords.
算法霸主承担问责。
in big data must end.
ABOUT THE SPEAKER
Cathy O'Neil - Mathematician, data scientistData skeptic Cathy O’Neil uncovers the dark secrets of big data, showing how our "objective" algorithms could in fact reinforce human bias.
Why you should listen
In 2008, as a hedge-fund quant, mathematician Cathy O’Neil saw firsthand how really really bad math could lead to financial disaster. Disillusioned, O’Neil became a data scientist and eventually joined Occupy Wall Street’s Alternative Banking Group.
With her popular blog mathbabe.org, O’Neil emerged as an investigative journalist. Her acclaimed book Weapons of Math Destruction details how opaque, black-box algorithms rely on biased historical data to do everything from sentence defendants to hire workers. In 2017, O’Neil founded consulting firm ORCAA to audit algorithms for racial, gender and economic inequality.
Cathy O'Neil | Speaker | TED.com