Placeholder Image

Subtitles section Play video

  • Algorithms are everywhere.

    譯者: Lilian Chiu 審譯者: NAN-KUN WU

  • They sort and separate the winners from the losers.

    演算法無所不在。

  • The winners get the job

    它們能把贏家和輸家區分開來。

  • or a good credit card offer.

    贏家能得到工作,

  • The losers don't even get an interview

    或是好的信用卡方案。

  • or they pay more for insurance.

    輸家連面試的機會都沒有,

  • We're being scored with secret formulas that we don't understand

    或是他們的保險費比較高。

  • that often don't have systems of appeal.

    我們都被我們不了解的 秘密方程式在評分,

  • That begs the question:

    且那些方程式通常 都沒有申訴體制。

  • What if the algorithms are wrong?

    問題就來了:

  • To build an algorithm you need two things:

    如果演算法是錯的怎麼辦?

  • you need data, what happened in the past,

    要建立一個演算法,需要兩樣東西:

  • and a definition of success,

    需要資料,資料是過去發生的事,

  • the thing you're looking for and often hoping for.

    還需要對成功的定義,

  • You train an algorithm by looking, figuring out.

    也就是你在找的東西、 你想要的東西。

  • The algorithm figures out what is associated with success.

    你透過尋找和計算的方式 來訓練一個演算法。

  • What situation leads to success?

    演算法會算出什麼和成功有相關性。

  • Actually, everyone uses algorithms.

    什麼樣的情況會導致成功?

  • They just don't formalize them in written code.

    其實,人人都在用演算法。

  • Let me give you an example.

    他們只是沒把演算法寫為程式。

  • I use an algorithm every day to make a meal for my family.

    讓我舉個例子。

  • The data I use

    我每天都用演算法 來為我的家庭做飯。

  • is the ingredients in my kitchen,

    我用的資料

  • the time I have,

    是我廚房中的原料、

  • the ambition I have,

    我擁有的時間、

  • and I curate that data.

    我的野心、

  • I don't count those little packages of ramen noodles as food.

    我把這些資料拿來做策劃。

  • (Laughter)

    我不把那一小包小包的 拉麵條視為是食物。

  • My definition of success is:

    (笑聲)

  • a meal is successful if my kids eat vegetables.

    我對成功的定義是:

  • It's very different from if my youngest son were in charge.

    如果我的孩子吃了蔬菜, 這頓飯就算成功。

  • He'd say success is if he gets to eat lots of Nutella.

    但如果我的小兒子主導時 一切就不同了。

  • But I get to choose success.

    他會說,如果能吃到很多 能多益(巧克力榛果醬)就算成功。

  • I am in charge. My opinion matters.

    但我能選擇什麼才算成功。

  • That's the first rule of algorithms.

    我是主導的人,我的意見才重要。

  • Algorithms are opinions embedded in code.

    那是演算法的第一條規則。

  • It's really different from what you think most people think of algorithms.

    演算法是被嵌入程式中的意見。

  • They think algorithms are objective and true and scientific.

    這和你認為大部份人 對演算法的看法很不一樣。

  • That's a marketing trick.

    他們認為演算法是 客觀的、真實的、科學的。

  • It's also a marketing trick

    那是種行銷技倆。

  • to intimidate you with algorithms,

    還有一種行銷技倆是

  • to make you trust and fear algorithms

    用演算法來威脅你,

  • because you trust and fear mathematics.

    讓你相信並懼怕演算法,

  • A lot can go wrong when we put blind faith in big data.

    因為你相信並懼怕數學。

  • This is Kiri Soares. She's a high school principal in Brooklyn.

    當我們盲目相信大數據時, 很多地方都可能出錯。

  • In 2011, she told me her teachers were being scored

    這位是琦莉索瑞斯, 她是布魯克林的高中校長。

  • with a complex, secret algorithm

    2011 年,她告訴我, 用來評分她的老師的演算法

  • called the "value-added model."

    是一種複雜的秘密演算法,

  • I told her, "Well, figure out what the formula is, show it to me.

    叫做「加值模型」。

  • I'm going to explain it to you."

    我告訴她:「找出那方程式 是什麼,給我看,

  • She said, "Well, I tried to get the formula,

    我就會解釋給你聽。」

  • but my Department of Education contact told me it was math

    她說:「嗯,我試過取得方程式了,

  • and I wouldn't understand it."

    但教育部聯絡人告訴我, 那方程式是數學,

  • It gets worse.

    我也看不懂的。」

  • The New York Post filed a Freedom of Information Act request,

    還有更糟的。

  • got all the teachers' names and all their scores

    紐約郵報提出了一項 資訊自由法案的請求,

  • and they published them as an act of teacher-shaming.

    取得有所有老師的名字 以及他們的分數,

  • When I tried to get the formulas, the source code, through the same means,

    郵報把這些都刊出來, 用來羞辱老師。

  • I was told I couldn't.

    當我試著透過同樣的手段 來找出方程式、原始碼,

  • I was denied.

    我被告知我不可能辦到。

  • I later found out

    我被拒絕了。

  • that nobody in New York City had access to that formula.

    我後來發現,

  • No one understood it.

    紐約市中沒有人能取得那方程式。

  • Then someone really smart got involved, Gary Rubinstein.

    沒有人了解它。

  • He found 665 teachers from that New York Post data

    有個很聰明的人介入: 蓋瑞魯賓斯坦。

  • that actually had two scores.

    他發現紐約郵報資料中 有 665 名老師

  • That could happen if they were teaching

    其實有兩個分數。

  • seventh grade math and eighth grade math.

    如果他們是在教七年級

  • He decided to plot them.

    及八年級數學,是有可能發生。

  • Each dot represents a teacher.

    他決定把他們用圖畫出來。

  • (Laughter)

    每一個點代表一個老師。

  • What is that?

    (笑聲)

  • (Laughter)

    那是什麼?

  • That should never have been used for individual assessment.

    (笑聲)

  • It's almost a random number generator.

    那絕對不該被用來做個人評估用。

  • (Applause)

    它幾乎就是個隨機數產生器。

  • But it was.

    (掌聲)

  • This is Sarah Wysocki.

    但它的確被用了。

  • She got fired, along with 205 other teachers,

    這是莎拉薇沙琪,

  • from the Washington, DC school district,

    她和其他 205 名老師都被開除了,

  • even though she had great recommendations from her principal

    都是在華盛頓特區的學區,

  • and the parents of her kids.

    即使她有校長及 學童家長的強力推薦,

  • I know what a lot of you guys are thinking,

    還是被開除了。

  • especially the data scientists, the AI experts here.

    我很清楚你們在想什麼,

  • You're thinking, "Well, I would never make an algorithm that inconsistent."

    特別是這裡的資料科學家 及人工智慧專家。

  • But algorithms can go wrong,

    你們在想:「我絕對不會寫出 那麼不一致的演算法。」

  • even have deeply destructive effects with good intentions.

    但演算法是可能出錯的,

  • And whereas an airplane that's designed badly

    即使出自好意 仍可能產生毀滅性的效應。

  • crashes to the earth and everyone sees it,

    設計得很糟的飛機墜機,

  • an algorithm designed badly

    每個人都會看到;

  • can go on for a long time, silently wreaking havoc.

    可是,設計很糟的演算法,

  • This is Roger Ailes.

    可以一直運作很長的時間, 靜靜地製造破壞或混亂。

  • (Laughter)

    這位是羅傑艾爾斯。

  • He founded Fox News in 1996.

    (笑聲)

  • More than 20 women complained about sexual harassment.

    他在 1996 年成立了 Fox News。

  • They said they weren't allowed to succeed at Fox News.

    有超過二十位女性投訴性騷擾。

  • He was ousted last year, but we've seen recently

    她們說,她們在 Fox News 不被允許成功。

  • that the problems have persisted.

    他去年被攆走了,但我們看到近期

  • That begs the question:

    這個問題仍然存在。

  • What should Fox News do to turn over another leaf?

    這就帶來一個問題:

  • Well, what if they replaced their hiring process

    Fox News 該做什麼才能改過自新?

  • with a machine-learning algorithm?

    如果他們把僱用的流程換掉,

  • That sounds good, right?

    換成機器學習演算法呢?

  • Think about it.

    聽起來很好,對嗎?

  • The data, what would the data be?

    想想看。

  • A reasonable choice would be the last 21 years of applications to Fox News.

    資料,資料會是什麼?

  • Reasonable.

    一個合理的選擇會是 Fox News 過去 21 年間收到的申請。

  • What about the definition of success?

    很合理。

  • Reasonable choice would be,

    成功的定義呢?

  • well, who is successful at Fox News?

    合理的選擇會是,

  • I guess someone who, say, stayed there for four years

    在 Fox News 有誰是成功的?

  • and was promoted at least once.

    我猜是在那邊待了四年、

  • Sounds reasonable.

    且至少升遷過一次的人。

  • And then the algorithm would be trained.

    聽起來很合理。

  • It would be trained to look for people to learn what led to success,

    接著,演算法就會被訓練。

  • what kind of applications historically led to success

    它會被訓練來找人, 尋找什麼導致成功,

  • by that definition.

    在過去怎樣的申請書會導致成功,

  • Now think about what would happen

    用剛剛的成功定義。

  • if we applied that to a current pool of applicants.

    想想看會發生什麼事,

  • It would filter out women

    如果我們把它用到 目前的一堆申請書上。

  • because they do not look like people who were successful in the past.

    它會把女性過濾掉,

  • Algorithms don't make things fair

    因為在過去,女性 並不像是會成功的人。

  • if you just blithely, blindly apply algorithms.

    如果只是漫不經心、 盲目地運用演算法,

  • They don't make things fair.

    它們並不會讓事情變公平。

  • They repeat our past practices,

    演算法不會讓事情變公平。

  • our patterns.

    它們會重覆我們過去的做法,

  • They automate the status quo.

    我們的模式。

  • That would be great if we had a perfect world,

    它們會把現狀給自動化。

  • but we don't.

    如果我們有個完美的 世界,那就很好了,

  • And I'll add that most companies don't have embarrassing lawsuits,

    但世界不完美。

  • but the data scientists in those companies

    我還要補充,大部份公司 沒有難堪的訴訟,

  • are told to follow the data,

    但在那些公司中的資料科學家

  • to focus on accuracy.

    被告知要遵從資料,

  • Think about what that means.

    著重正確率。

  • Because we all have bias, it means they could be codifying sexism

    想想那意味著什麼。

  • or any other kind of bigotry.

    因為我們都有偏見,那就意味著, 他們可能會把性別偏見

  • Thought experiment,

    或其他偏執給寫到程式中,

  • because I like them:

    來做個思想實驗,

  • an entirely segregated society --

    因為我喜歡思想實驗:

  • racially segregated, all towns, all neighborhoods

    一個完全種族隔離的社會,

  • and where we send the police only to the minority neighborhoods

    所有的城鎮、所有的街坊 都做了種族隔離,

  • to look for crime.

    我們只會針對少數種族 住的街坊派出警力

  • The arrest data would be very biased.

    來尋找犯罪。

  • What if, on top of that, we found the data scientists

    逮捕的資料會非常偏頗。

  • and paid the data scientists to predict where the next crime would occur?

    如果再加上,我們 找到了資料科學家,

  • Minority neighborhood.

    付錢給他們,要他們預測下次 犯罪會發生在哪裡,會如何?

  • Or to predict who the next criminal would be?

    答案:少數種族的街坊。

  • A minority.

    或是去預測下一位犯人會是誰?

  • The data scientists would brag about how great and how accurate

    答案:少數族裔。

  • their model would be,

    資料科學家會吹噓他們的的模型

  • and they'd be right.

    有多了不起、多精準,

  • Now, reality isn't that drastic, but we do have severe segregations

    他們是對的。

  • in many cities and towns,

    現實沒那麼極端,但在許多 城鎮和城市中,我們的確有

  • and we have plenty of evidence

    嚴重的種族隔離,

  • of biased policing and justice system data.

    我們有很多證據可證明

  • And we actually do predict hotspots,

    執法和司法資料是偏頗的。

  • places where crimes will occur.

    我們確實預測了熱點,

  • And we do predict, in fact, the individual criminality,

    犯罪會發生的地方。

  • the criminality of individuals.

    事實上,我們確實預測了 個別的犯罪行為,

  • The news organization ProPublica recently looked into

    個人的犯罪行為。

  • one of those "recidivism risk" algorithms,

    新聞組織 ProPublica 近期調查了

  • as they're called,

    「累犯風險」演算法之一,

  • being used in Florida during sentencing by judges.

    他們是這麼稱呼它的,

  • Bernard, on the left, the black man, was scored a 10 out of 10.

    演算法被用在佛羅里達, 法官在判刑時使用。

  • Dylan, on the right, 3 out of 10.

    左邊的黑人是伯納, 總分十分,他得了十分。

  • 10 out of 10, high risk. 3 out of 10, low risk.

    右邊的狄倫,十分只得了三分。

  • They were both brought in for drug possession.

    十分就得十分,高風險。 十分只得三分,低風險。

  • They both had records,

    他們都因為持有藥品而被逮捕。

  • but Dylan had a felony

    他們都有犯罪記錄,

  • but Bernard didn't.

    但狄倫犯過重罪,

  • This matters, because the higher score you are,

    伯納則沒有。

  • the more likely you're being given a longer sentence.

    這很重要,因為你的得分越高,

  • What's going on?

    你就越可能被判比較長的徒刑。

  • Data laundering.

    發生了什麼事?

  • It's a process by which technologists hide ugly truths

    洗資料。

  • inside black box algorithms

    它是個流程,即技術專家 用黑箱作業的演算法

  • and call them objective;

    來隱藏醜陋的真相,

  • call them meritocratic.

    還宣稱是客觀的;

  • When they're secret, important and destructive,

    是精英領導的。

  • I've coined a term for these algorithms:

    我為這些秘密、重要、

  • "weapons of math destruction."

    又有毀滅性的演算法取了個名字:

  • (Laughter)

    「毀滅性的數學武器」。

  • (Applause)

    (笑聲)

  • They're everywhere, and it's not a mistake.

    (掌聲)

  • These are private companies building private algorithms

    它們無所不在,且不是個過失。

  • for private ends.

    私人公司建立私人演算法,

  • Even the ones I talked about for teachers and the public police,

    來達到私人的目的。

  • those were built by private companies

    即使是我剛談到 對老師和警方用的演算法,

  • and sold to the government institutions.

    也是由私人公司建立的,

  • They call it their "secret sauce" --

    然後再銷售給政府機關。

  • that's why they can't tell us about it.

    他們稱它為「秘方醬料」,

  • It's also private power.

    所以不能跟我們討論它。

  • They are profiting for wielding the authority of the inscrutable.

    它也是種私人的權力。

  • Now you might think, since all this stuff is private

    他們透過行使別人 無法理解的權威來獲利。

  • and there's competition,

    你可能會認為, 所有這些都是私人的,

  • maybe the free market will solve this problem.

    且有競爭存在,

  • It won't.

    也許自由市場會解決這個問題。

  • There's a lot of money to be made in unfairness.

    並不會。

  • Also, we're not economic rational agents.

    從不公平中可以賺取很多錢。

  • We all are biased.

    且,我們不是經濟合法代理人。

  • We're all racist and bigoted in ways that we wish we weren't,

    我們都有偏見。

  • in ways that we don't even know.

    我們都是種族主義的、偏執的, 即使我們也希望不要這樣,

  • We know this, though, in aggregate,

    我們甚至不知道我們是這樣的。

  • because sociologists have consistently demonstrated this

    不過我們確實知道,總的來說,

  • with these experiments they build,

    因為社會學家不斷地用 他們建立的實驗

  • where they send a bunch of applications to jobs out,

    來展現出這一點,

  • equally qualified but some have white-sounding names

    他們寄出一大堆的工作申請書,

  • and some have black-sounding names,

    都有同樣的資格, 但有些用白人人名,

  • and it's always disappointing, the results -- always.

    有些用黑人人名,

  • So we are the ones that are biased,

    結果總是讓人失望的,總是如此。

  • and we are injecting those biases into the algorithms

    所以,我們才是有偏見的人,

  • by choosing what data to collect,

    且我們把這些偏見注入演算法中,

  • like I chose not to think about ramen noodles --

    做法是選擇要收集哪些資料、

  • I decided it was irrelevant.

    比如我選擇不要考量拉麵,

  • But by trusting the data that's actually picking up on past practices

    我決定它不重要。

  • and by choosing the definition of success,

    但透過相信這些資料 真的能了解過去的做法,

  • how can we expect the algorithms to emerge unscathed?

    以及透過選擇成功的定義,

  • We can't. We have to check them.

    我們如何能冀望產生的演算法未受損?

  • We have to check them for fairness.

    不能。我們得要檢查這些演算法。

  • The good news is, we can check them for fairness.

    我們得要檢查它們是否公平。

  • Algorithms can be interrogated,

    好消息是,我們可以 檢查它們是否公平。

  • and they will tell us the truth every time.

    演算法可以被審問,

  • And we can fix them. We can make them better.

    且它們每次都會告訴我們真相。

  • I call this an algorithmic audit,

    我們可以修正它們, 我們可以把它們變更好。

  • and I'll walk you through it.

    我稱這個為演算法稽核,

  • First, data integrity check.

    我會帶大家來了解它。

  • For the recidivism risk algorithm I talked about,

    首先,檢查資料完整性。

  • a data integrity check would mean we'd have to come to terms with the fact

    針對我先前說的累犯風險演算法,

  • that in the US, whites and blacks smoke pot at the same rate

    檢查資料完整性就意味著 我們得接受事實,

  • but blacks are far more likely to be arrested --

    事實是,在美國,白人和黑人 抽大麻的比率是一樣的,

  • four or five times more likely, depending on the area.

    但黑人被逮捕的機率遠高於白人,

  • What is that bias looking like in other crime categories,

    四、五倍高的可能性被捕, 依地區而異。

  • and how do we account for it?

    在其他犯罪類別中, 那樣的偏見會如何呈現?

  • Second, we should think about the definition of success,

    我們要如何處理它?

  • audit that.

    第二,我們要想想成功的定義,

  • Remember -- with the hiring algorithm? We talked about it.

    去稽核它。

  • Someone who stays for four years and is promoted once?

    記得我們剛剛談過的僱用演算法嗎?

  • Well, that is a successful employee,

    待了四年且升遷至少一次?

  • but it's also an employee that is supported by their culture.

    那就是個成功員工,

  • That said, also it can be quite biased.

    但那也是個被其文化所支持的員工。

  • We need to separate those two things.

    儘管如此,它也可能很有偏見。

  • We should look to the blind orchestra audition

    我們得把這兩件事分開。

  • as an example.

    我們應該要把交響樂團的盲眼甄選

  • That's where the people auditioning are behind a sheet.

    當作參考範例。

  • What I want to think about there

    他們的做法是讓試演奏的人 在布幕後演奏。

  • is the people who are listening have decided what's important

    我想探討的重點是

  • and they've decided what's not important,

    那些在聽並且決定什麼重要的人,

  • and they're not getting distracted by that.

    他們也會決定什麼不重要 ,

  • When the blind orchestra auditions started,

    他們不會被不重要的部份給分心。

  • the number of women in orchestras went up by a factor of five.

    當交響樂團開始採用盲眼甄選,

  • Next, we have to consider accuracy.

    團內的女性成員數上升五倍。

  • This is where the value-added model for teachers would fail immediately.

    接著,我們要考量正確率。

  • No algorithm is perfect, of course,

    這就是老師的加值模型 立刻會出問題的地方。

  • so we have to consider the errors of every algorithm.

    當然,沒有演算法是完美的,

  • How often are there errors, and for whom does this model fail?

    所以我們得要考量 每個演算法的錯誤。

  • What is the cost of that failure?

    多常會出現錯誤、這個模型 針對哪些人會發生錯誤?

  • And finally, we have to consider

    發生錯誤的成本多高?

  • the long-term effects of algorithms,

    最後,我們得要考量

  • the feedback loops that are engendering.

    演算法的長期效應,

  • That sounds abstract,

    也就是產生出來的反饋迴圈。

  • but imagine if Facebook engineers had considered that

    那聽起來很抽象,

  • before they decided to show us only things that our friends had posted.

    但想像一下,如果臉書的工程師

  • I have two more messages, one for the data scientists out there.

    決定只讓我們看到朋友的貼文 之前就先考量那一點。

  • Data scientists: we should not be the arbiters of truth.

    我還有兩個訊息要傳遞, 其一是給資料科學家的。

  • We should be translators of ethical discussions that happen

    資料科學家,我們 不應該是真相的仲裁者,

  • in larger society.

    我們應該是翻譯者,

  • (Applause)

    翻譯大社會中發生的每個道德討論。

  • And the rest of you,

    (掌聲)

  • the non-data scientists:

    至於你們其他人,

  • this is not a math test.

    不是資料科學家的人:

  • This is a political fight.

    這不是個數學考試。

  • We need to demand accountability for our algorithmic overlords.

    這是場政治鬥爭。

  • (Applause)

    我們得要求為演算法的超載負責。

  • The era of blind faith in big data must end.

    (掌聲)

  • Thank you very much.

    盲目信仰大數據的時代必須要結束。

  • (Applause)

    非常謝謝。

Algorithms are everywhere.

譯者: Lilian Chiu 審譯者: NAN-KUN WU

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it

B1 US TED 演算法 資料 成功 方程式 偏見

【TED】凱西-奧尼爾:盲目相信大數據的時代必須結束(The era of blind faith in big data must end | Cathy O'Neil)。 (【TED】Cathy O'Neil: The era of blind faith in big data must end (The era of blind faith in big data must end | Cathy O'Neil))

  • 43 11
    Zenn posted on 2021/01/14
Video vocabulary