ABOUT THE SPEAKER
Eli Pariser - Organizer and author
Pioneering online organizer Eli Pariser is the author of "The Filter Bubble," about how personalized search might be narrowing our worldview.

Why you should listen

Shortly after the September 11, 2001, attacks, Eli Pariser created a website calling for a multilateral approach to fighting terrorism. In the following weeks, over half a million people from 192 countries signed on, and Pariser rather unexpectedly became an online organizer. The website merged with MoveOn.org in November 2001, and Pariser -- then 20 years old -- joined the group to direct its foreign policy campaigns. He led what the New York Times Magazine called the "mainstream arm of the peace movement" -- tripling MoveOn's member base and demonstrating how large numbers of small donations could be mobilized through online engagement.

In 2004, Pariser became executive director of MoveOn. Under his leadership, MoveOn.org Political Action has grown to 5 million members and raised over $120 million from millions of small donors to support advocacy campaigns and political candidates. Pariser focused MoveOn on online-to-offline organizing, developing phone-banking tools and precinct programs in 2004 and 2006 that laid the groundwork for Barack Obama's extraordinary web-powered campaign. In 2008, Pariser transitioned the Executive Director role at MoveOn to Justin Ruben and became President of MoveOn’s board; he's now a senior fellow at the Roosevelt Institute.

His book The Filter Bubble is set for release May 12, 2011. In it, he asks how modern search tools -- the filter by which many of see the wider world -- are getting better and better and screening the wider world from us, by returning only the search results it "thinks" we want to see.

More profile about the speaker
Eli Pariser | Speaker | TED.com
TED2011

Eli Pariser: Beware online "filter bubbles"

Eli Pariser: Ettevaatust interneti "filtreeritud mullidega"!

Filmed:
5,309,238 views

Internetiettevõtted tahavad teha oma teenuseid aina personaalsemateks (sh. uudised ja otsingutulemused), et need vastaksid meie isiklikele eelistustele. Sellega kaasneb aga võimalus, et me jääme "filtreeritud mulli" ja ei saa enam uut teavet, mis võiks avardada meie silmaringi. Eli Pariser hoiatab, et see võib halvasti mõjuda nii meile kui demokraatiale laiemalt.
- Organizer and author
Pioneering online organizer Eli Pariser is the author of "The Filter Bubble," about how personalized search might be narrowing our worldview. Full bio

Double-click the English transcript below to play the video.

00:15
MarkMark ZuckerbergZuckerberg,
0
0
2000
Mark Zukerbergi käest
00:17
a journalistajakirjanik was askingpaludes him a questionküsimus about the newsuudised feedsööda.
1
2000
3000
küsis ajakirjanik facebooki uudistevoo kohta.
00:20
And the journalistajakirjanik was askingpaludes him,
2
5000
2000
Ajakirjanik küsis:
00:22
"Why is this so importantoluline?"
3
7000
2000
"Miks see üldse nii oluline on?"
00:24
And ZuckerbergZuckerberg said,
4
9000
2000
Ja Zukerberg vastas:
00:26
"A squirrelorav dyingsuremas in your frontees yardõue
5
11000
2000
"Teie hoovis surev orav
00:28
mayvõib be more relevantasjakohane to your interestshuvid right now
6
13000
3000
võib olla teie huvidest lähtuvalt olulisem
00:31
than people dyingsuremas in AfricaAfrica."
7
16000
3000
kui inimesed, kes surevad Aafrikas."
00:34
And I want to talk about
8
19000
2000
Ja ma tahaksin rääkida milline
00:36
what a WebWeb basedpõhineb on that ideaidee of relevanceasjakohasus mightvõib-olla look like.
9
21000
3000
sellisel ideel põhinev internet välja võiks näha.
00:40
So when I was growingkasvab up
10
25000
2000
Ma kasvasin üles
00:42
in a really ruralmaal areaala in MaineMaine,
11
27000
2000
Maine osariigi maapiirkonnas
00:44
the InternetInternet meanttähendas something very differenterinevad to me.
12
29000
3000
ja internet oli minu jaoks midagi erilist.
00:47
It meanttähendas a connectionühendus to the worldmaailm.
13
32000
2000
See oli ühenduseks maailmaga.
00:49
It meanttähendas something that would connectühendage us all togetherkoos.
14
34000
3000
See oli midagi, mis meid kõiki omavahel ühendab.
00:52
And I was sure that it was going to be great for democracydemokraatia
15
37000
3000
Ma olin kindel, et see tuleb suuresti kasuks demokraatiale
00:55
and for our societyühiskond.
16
40000
3000
ja kogu meie ühiskonnale.
00:58
But there's this shiftnihe
17
43000
2000
Kuid nüüd on toimumas muutus
01:00
in how informationteave is flowingvoolav onlineInternetis,
18
45000
2000
kuidas informatsioon internetis liigub
01:02
and it's invisiblenähtamatu.
19
47000
3000
ja see on nähtamatu.
01:05
And if we don't paymaksma attentiontähelepanu to it,
20
50000
2000
Ning kui me sellele tähelepanu ei pööra
01:07
it could be a realreaalne problemprobleem.
21
52000
3000
võib see probleemiks saada.
01:10
So I first noticedmärkasin this in a placekoht I spendkulutama a lot of time --
22
55000
3000
Märkasin seda esimesena kohas, kus palju aega veedan -
01:13
my FacebookFacebook pageleht.
23
58000
2000
oma Facebooki lehel.
01:15
I'm progressiveprogresseeruv, politicallypoliitiliselt -- bigsuur surpriseüllatus --
24
60000
3000
Olen poliitiliselt liberaal - suur üllatus -
01:18
but I've always goneläinud out of my way to meetkohtuma conservativesKonservatiivid.
25
63000
2000
aga olen teadlikult ka konservatiividega suhelnud.
01:20
I like hearingära kuulata what they're thinkingmõtlesin about;
26
65000
2000
Mulle meeldib kuulata millest nad mõtlevad,
01:22
I like seeingnähes what they linklink to;
27
67000
2000
mulle meeldib näha mida nad jagavad,
01:24
I like learningõppimine a thing or two.
28
69000
2000
mulle meeldib uusi asju teada saada.
01:26
And so I was surprisedüllatunud when I noticedmärkasin one day
29
71000
3000
Ja ma olin üllatunud ühel päeval märgates,
01:29
that the conservativesKonservatiivid had disappearedkadus from my FacebookFacebook feedsööda.
30
74000
3000
et mu konservatiividest sõbrad olid Facebooki uudistevoost kadunud.
01:33
And what it turnedpööratud out was going on
31
78000
2000
Tuli välja, et
01:35
was that FacebookFacebook was looking at whichmis linkslingid I clickedklõpsatud on,
32
80000
4000
Facebook vaatas, millistele linkidele ma vajutan
01:39
and it was noticingmärganud that, actuallytegelikult,
33
84000
2000
ja nägi, et tegelikult,
01:41
I was clickingklõpsates more on my liberalliberaalne friends'Sõprade linkslingid
34
86000
2000
vajutan ma liberaalidest sõprade linkidele rohkem
01:43
than on my conservativekonservatiivne friends'Sõprade linkslingid.
35
88000
3000
kui konservatiivide omadele.
01:46
And withoutilma consultingnõustamine me about it,
36
91000
2000
Ja ilma minuga nõu pidamata,
01:48
it had editedtoimetatud them out.
37
93000
2000
oli ta nad välja jätnud.
01:50
They disappearedkadus.
38
95000
3000
Nad olid kadunud.
01:54
So FacebookFacebook isn't the only placekoht
39
99000
2000
Ega Facebook pole ainuke koht,
01:56
that's doing this kindlaadi of invisiblenähtamatu, algorithmicalgorithmic
40
101000
2000
mis sarnast nähtamatut, algoritmilist
01:58
editingredigeerimine of the WebWeb.
41
103000
3000
toimetamist kasutab.
02:01
Google'sGoogle's doing it too.
42
106000
2000
Google teeb sedasama.
02:03
If I searchotsing for something, and you searchotsing for something,
43
108000
3000
Kui mina ja teie praegu midagi otsiksime,
02:06
even right now at the very samesama time,
44
111000
2000
koos ja samal ajal,
02:08
we mayvõib get very differenterinevad searchotsing resultstulemused.
45
113000
3000
võime saada täiesti erinevad otsingutulemused.
02:11
Even if you're loggedsisse logitud out, one engineerinsener told me,
46
116000
3000
Isegi juhul kui te ei ole sisse loginud, rääkis üks insener mulle,
02:14
there are 57 signalssignaale
47
119000
2000
et on 57 signaali,
02:16
that GoogleGoogle looksnäeb välja at --
48
121000
3000
mida Google kontrollib -
02:19
everything from what kindlaadi of computerarvuti you're on
49
124000
3000
alates millist arvutit kasutatakse,
02:22
to what kindlaadi of browserbrauser you're usingkasutades
50
127000
2000
millist veebilehitsejat,
02:24
to where you're locatedasub --
51
129000
2000
kuni otsija asukohani -
02:26
that it useskasutab to personallyisiklikult tailorrätsep your querypäring resultstulemused.
52
131000
3000
nende abil pannakse kokku otsingu tulemused.
02:29
Think about it for a secondteine:
53
134000
2000
Mõelge natuke selle peale:
02:31
there is no standardstandard GoogleGoogle anymoreenam.
54
136000
4000
tavalist Google'it ei ole enam olemas.
02:35
And you know, the funnynaljakas thing about this is that it's hardraske to see.
55
140000
3000
Teate, naljakas asja juures on see, et seda on keeruline näha.
02:38
You can't see how differenterinevad your searchotsing resultstulemused are
56
143000
2000
Te ei saa ju näha, kui palju teie otsingutulemused
02:40
from anyonekeegi else'smujal.
57
145000
2000
teiste omadest erinevad.
02:42
But a couplepaar of weeksnädalat agotagasi,
58
147000
2000
Kuid paar nädalat tagasi
02:44
I askedküsisin a bunchkamp of friendssõbrad to GoogleGoogle "EgyptEgiptus"
59
149000
3000
palusin oma sõpradel googeldada sõna "Egiptus".
02:47
and to sendsaada me screenekraan shotskaadrid of what they got.
60
152000
3000
Palusin, et nad saadaksid mulle pildid vastustest.
02:50
So here'ssiin on my friendsõber Scott'sScotti screenekraan shottulistas.
61
155000
3000
Siin on minu sõber Scott oma vastustega.
02:54
And here'ssiin on my friendsõber Daniel'sDaniel's screenekraan shottulistas.
62
159000
3000
Ja siin on Daniel enda omadega.
02:57
When you put them side-by-sidekõrvuti,
63
162000
2000
Kui need kõrvuti panna,
02:59
you don't even have to readloe the linkslingid
64
164000
2000
ei pea isegi linke lähemalt lugema,
03:01
to see how differenterinevad these two pagesleheküljed are.
65
166000
2000
et aru saada kui erinevad tulemused on.
03:03
But when you do readloe the linkslingid,
66
168000
2000
Kuid lähemalt vaadates
03:05
it's really quiteüsna remarkabletähelepanuväärne.
67
170000
3000
on see päris erakordne.
03:09
DanielDaniel didn't get anything about the protestsprotestid in EgyptEgiptus at all
68
174000
3000
Daniel ei saanud Egiptuse meeleavalduste kohta
03:12
in his first pageleht of GoogleGoogle resultstulemused.
69
177000
2000
esimesel lehel mitte midagi.
03:14
Scott'sScotti resultstulemused were fulltäis of them.
70
179000
2000
Scotti vastused olid neid aga täis.
03:16
And this was the bigsuur storylugu of the day at that time.
71
181000
2000
Need olid sel ajal aga kõige kuumemad uudised.
03:18
That's how differenterinevad these resultstulemused are becomingmuutumas.
72
183000
3000
Vaadake, kui erinevaks tulemused on muutumas.
03:21
So it's not just GoogleGoogle and FacebookFacebook eitherkas.
73
186000
3000
Ning see ei ole ainult Google ja Facebook.
03:24
This is something that's sweepingpühkimine the WebWeb.
74
189000
2000
See on midagi, mis toimub kogu internetis.
03:26
There are a wholeterve hostvõõrustaja of companiesettevõtted that are doing this kindlaadi of personalizationisikupärastamine.
75
191000
3000
On suur hulk ettevõtteid, kes kasutavad samasugust isikustamist.
03:29
YahooYahoo NewsUudised, the biggestsuurim newsuudised sitesait on the InternetInternet,
76
194000
3000
Yahoo News, interneti suurim uudistelehekülg
03:32
is now personalizedisikupärastatud -- differenterinevad people get differenterinevad things.
77
197000
3000
kasutab nüüd isikustamist - erinevad inimesed näevad erinevaid asju.
03:36
HuffingtonHuffington PostPostitus, the WashingtonWashington PostPostitus, the NewUus YorkYork TimesKorda --
78
201000
3000
Huffington Post, Washington Post ja New York Times -
03:39
all flirtingflirt with personalizationisikupärastamine in variouserinevad waysviisid.
79
204000
3000
semmivad kõik erineval moel isikustamisega.
03:42
And this movesliigub us very quicklykiiresti
80
207000
3000
Ja seeläbi liigume kiiresti
03:45
towardsuunas a worldmaailm in whichmis
81
210000
2000
olukorrani, kus
03:47
the InternetInternet is showingnäidates us what it thinksmõtleb we want to see,
82
212000
4000
internet näitab meile ainult neid asju mida ta eeldab, et näha tahame,
03:51
but not necessarilytingimata what we need to see.
83
216000
3000
mis ei pruugi alati olla asjad, mida nägema peame.
03:54
As EricEric SchmidtSchmidt said,
84
219000
3000
Eric Smith on öelnud:
03:57
"It will be very hardraske for people to watch or consumetarbida something
85
222000
3000
"Inimeste jaoks on keeruline vaadata või tarbida midagi,
04:00
that has not in some sensemeel
86
225000
2000
mis mingis plaanis
04:02
been tailoredkohandatud for them."
87
227000
3000
ei ole nende jaoks sobivaks tehtud."
04:05
So I do think this is a problemprobleem.
88
230000
2000
Ja ma arvan, et see on probleem.
04:07
And I think, if you take all of these filtersfiltrid togetherkoos,
89
232000
3000
Ma arvan ka, et kui võtta kõik need filtrid üheskoos,
04:10
you take all these algorithmsalgoritmid,
90
235000
2000
kui võtta kõik need algoritmid,
04:12
you get what I call a filterfilter bubblemull.
91
237000
3000
saame midagi, mida võiks nimetada filtreeritud mulliks.
04:16
And your filterfilter bubblemull is your ownoma personalisiklik,
92
241000
3000
Ja see filtreeritud mull on Sinu isiklik,
04:19
uniqueunikaalne universeuniversum of informationteave
93
244000
2000
ainulaadne inforuum,
04:21
that you liveelus in onlineInternetis.
94
246000
2000
milles internetis olles elad.
04:23
And what's in your filterfilter bubblemull
95
248000
3000
See, mis seal filtreeritud mullis on,
04:26
dependssõltub on who you are, and it dependssõltub on what you do.
96
251000
3000
sõltub sellest, kes Sa oled ja millega tegeled.
04:29
But the thing is that you don't decideotsustama what getssaab in.
97
254000
4000
Kuid asi on selles, et Sa ei saa otsustada, mis sinna mulli pääseb.
04:33
And more importantlytähtsam,
98
258000
2000
Veel olulisem on,
04:35
you don't actuallytegelikult see what getssaab editedtoimetatud out.
99
260000
3000
Sa ei saa ka tegelikult näha, mis sealt mullist välja jäetakse.
04:38
So one of the problemsprobleemid with the filterfilter bubblemull
100
263000
2000
Üks filtreeritud mulli probleeme
04:40
was discoveredavastati by some researchersteadlased at NetflixNetflix.
101
265000
3000
avastati Netflixi analüütikute poolt.
04:43
And they were looking at the NetflixNetflix queuesjärjekorrad, and they noticedmärkasin something kindlaadi of funnynaljakas
102
268000
3000
Nad vaatasid Netflixi videote järjekordi ja avastasid midagi päris huvitavat,
04:46
that a lot of us probablytõenäoliselt have noticedmärkasin,
103
271000
2000
mida me ilmselt ka ise oleme märganud,
04:48
whichmis is there are some moviesfilme
104
273000
2000
et on mõned filmid
04:50
that just sortsorteerima of zipzip right up and out to our housesmajad.
105
275000
3000
mis otsejoones meile kodudesse jõuavad.
04:53
They entersisenema the queuejärjekorda, they just zipzip right out.
106
278000
3000
Nad on järjekorras ja hüppavad kohe välja.
04:56
So "IronRaud Man" zipstõmblukud right out,
107
281000
2000
Näiteks "Raudmees" hüppab kohe välja,
04:58
and "WaitingOotel for SupermanSuperman"
108
283000
2000
aga "Oodates Supermani"
05:00
can wait for a really long time.
109
285000
2000
võib päris kaua aega oodata.
05:02
What they discoveredavastati
110
287000
2000
Nad avastasid,
05:04
was that in our NetflixNetflix queuesjärjekorrad
111
289000
2000
et meie Netflixi järjekordades
05:06
there's this epicEpic strugglevõitlus going on
112
291000
3000
käib suur võitlus
05:09
betweenvahel our futuretulevik aspirationalpüüdluslik selvesselves
113
294000
3000
meie tuleviku poole püüdlevate "minade"
05:12
and our more impulsiveimpulsiivne presentpresent selvesselves.
114
297000
3000
ja meie impulsiivsemate, olevikus elavate "minade" vahel.
05:15
You know we all want to be someonekeegi
115
300000
2000
Et me kõik tahaksime öelda, et
05:17
who has watchedvaatasin "RashomonRashomon,"
116
302000
2000
oleme juba näinud kultusfilmi "Rashomon",
05:19
but right now
117
304000
2000
kuid praegu
05:21
we want to watch "AceAKE VenturaVentura" for the fourthneljas time.
118
306000
3000
tahame juba neljandat korda vaadata filmi "Ace Ventura".
05:24
(LaughterNaer)
119
309000
3000
(Naer)
05:27
So the bestparim editingredigeerimine givesannab us a bitnatuke of bothmõlemad.
120
312000
2000
Ehk parim filtreerimine annab meile pisut mõlemat.
05:29
It givesannab us a little bitnatuke of JustinJustin BieberBieber
121
314000
2000
See annab natukene Justin Bieberit
05:31
and a little bitnatuke of AfghanistanAfganistan.
122
316000
2000
ja pisut Afganistani.
05:33
It givesannab us some informationteave vegetablesköögiviljad;
123
318000
2000
See annab meile olulist teavet
05:35
it givesannab us some informationteave dessertmagustoit.
124
320000
3000
ja pisut kollast sinna hulka.
05:38
And the challengeväljakutse with these kindsliiki of algorithmicalgorithmic filtersfiltrid,
125
323000
2000
Selliste algoritmiliste filtrite probleem,
05:40
these personalizedisikupärastatud filtersfiltrid,
126
325000
2000
selliste isikustatud filtrite väljakutse
05:42
is that, because they're mainlypeamiselt looking
127
327000
2000
on, et nad vaatavad põhiliselt,
05:44
at what you clickkliki on first,
128
329000
4000
kuhu vajutatakse esimesena
05:48
it can throwviska off that balancetasakaal.
129
333000
4000
ja võivad tasakaalust välja minna.
05:52
And insteadselle asemel of a balancedtasakaalustatud informationteave dietdieet,
130
337000
3000
Ning tasakaalustatud info asemel
05:55
you can endlõpp up surroundedümbritsetud
131
340000
2000
võite olla ümbritsetud
05:57
by informationteave junkrämps foodtoit.
132
342000
2000
informatsioon kiirtoidust.
05:59
What this suggestssoovitab
133
344000
2000
See viitab tõsiasjale,
06:01
is actuallytegelikult that we mayvõib have the storylugu about the InternetInternet wrongvalesti.
134
346000
3000
et oleme tegelikult internetist valesti aru saanud.
06:04
In a broadcastsaade societyühiskond --
135
349000
2000
Teabeühiskonnas -
06:06
this is how the foundingasutaja mythologymütoloogia goesläheb --
136
351000
2000
selline müüt on käibel -
06:08
in a broadcastsaade societyühiskond,
137
353000
2000
teabeühiskonnas
06:10
there were these gatekeepersjuurdepääsu reguleerivad, the editorstoimetajad,
138
355000
2000
olid nn. "väravavalvurid", toimetajad,
06:12
and they controlledkontrollitud the flowsvoolab of informationteave.
139
357000
3000
kes kontrollisid kogu informatsiooni liikumist.
06:15
And alongmööda cametuli the InternetInternet and it sweptpühib them out of the way,
140
360000
3000
Koos internetiga muutusid nad tähtsusetuks,
06:18
and it allowedlubatud all of us to connectühendage togetherkoos,
141
363000
2000
see andis kõigile võimaluse omavahel ühenduses olla
06:20
and it was awesomevinge.
142
365000
2000
ja see oli super.
06:22
But that's not actuallytegelikult what's happeningjuhtub right now.
143
367000
3000
Kuid nüüd toimub hoopis midagi muud.
06:26
What we're seeingnähes is more of a passingmööduv of the torchtõrvik
144
371000
3000
Nüüd oleme tunnistajaks,
06:29
from humaninimene gatekeepersjuurdepääsu reguleerivad
145
374000
2000
et inimestest "väravavalvurid"
06:31
to algorithmicalgorithmic onesneed.
146
376000
3000
lihtsalt asendatakse algoritmilistega.
06:34
And the thing is that the algorithmsalgoritmid
147
379000
3000
Pprobleem on selles, et algoritmidel
06:37
don't yetveel have the kindlaadi of embeddedmanustatud ethicseetika
148
382000
3000
ei ole veel sisse ehitatud eetilist kompassi,
06:40
that the editorstoimetajad did.
149
385000
3000
mis toimetajatel olemas oli.
06:43
So if algorithmsalgoritmid are going to curatekureerida the worldmaailm for us,
150
388000
3000
Ehk, kui algoritmid maailma meie eest kureerivad,
06:46
if they're going to decideotsustama what we get to see and what we don't get to see,
151
391000
3000
kui nemad otsustavad, mida me näeme ja mida mitte,
06:49
then we need to make sure
152
394000
2000
siis peame olema kindlad,
06:51
that they're not just keyedsisestatud to relevanceasjakohasus.
153
396000
3000
et nad ei otsiks ainult olulisi samasusi.
06:54
We need to make sure that they alsoka shownäidata us things
154
399000
2000
Peame olema kindlad, et nad näitaksid meile ka asju,
06:56
that are uncomfortableebamugav or challengingväljakutse or importantoluline --
155
401000
3000
mis võivad olla ebamugavad, või keerulised, või tähtsad -
06:59
this is what TEDTED does --
156
404000
2000
see on see, mida TED teeb -
07:01
other pointspunkte of viewvaade.
157
406000
2000
teistsugused seisukohad.
07:03
And the thing is, we'veme oleme actuallytegelikult been here before
158
408000
2000
Me oleme seda korra juba läbi teinud
07:05
as a societyühiskond.
159
410000
2000
ühiskonnana.
07:08
In 1915, it's not like newspapersajalehed were sweatinghigistamine a lot
160
413000
3000
1915.a. ei mõelnud ajalehed just palju
07:11
about theiroma civickodanik responsibilitieskohustused.
161
416000
3000
oma sotsiaalse vastutuse peale.
07:14
Then people noticedmärkasin
162
419000
2000
Kuid siis märkasid inimesed,
07:16
that they were doing something really importantoluline.
163
421000
3000
et ajalehtedel on tegelikult oluline roll.
07:19
That, in factfakt, you couldn'tei saanud have
164
424000
2000
Sest ei saa olla
07:21
a functioningtoimimine democracydemokraatia
165
426000
2000
toimivat demokraatiat
07:23
if citizenskodanikud didn't get a good flowvoolu of informationteave,
166
428000
4000
ilma, et kodanikeni jõuaks olulist teavet.
07:28
that the newspapersajalehed were criticalkriitiline because they were actingtegutsedes as the filterfilter,
167
433000
3000
Ajalehed olid tähtsad, sest toimisid sellise filtrina
07:31
and then journalisticajakirjandusliku ethicseetika developedarenenud.
168
436000
2000
ja tekkiski ajakirjanduslik eetika.
07:33
It wasn'tei olnud perfecttäiuslik,
169
438000
2000
See ei olnud ideaalne,
07:35
but it got us throughläbi the last centurysajandil.
170
440000
3000
aga see aitas meid läbi eelmise sajandi.
07:38
And so now,
171
443000
2000
Nüüdsiis
07:40
we're kindlaadi of back in 1915 on the WebWeb.
172
445000
3000
oleme internetiga justkui aastas 1915.
07:44
And we need the newuus gatekeepersjuurdepääsu reguleerivad
173
449000
3000
Me vajame uusi "väravahoidjaid",
07:47
to encodekodeerida that kindlaadi of responsibilityvastutus
174
452000
2000
kes kirjutaksid sellise vastutustunde
07:49
into the codekood that they're writingkirjalikult.
175
454000
2000
ka loodavatesse koodidesse.
07:51
I know that there are a lot of people here from FacebookFacebook and from GoogleGoogle --
176
456000
3000
Tean, et siin on palju inimesi Facebookist ja Googlest -
07:54
LarryLarry and SergeySergei --
177
459000
2000
Larry ja Sergei -
07:56
people who have helpedaitas buildehitama the WebWeb as it is,
178
461000
2000
inimesed, kes on aidanud interneti ehitada selliseks nagu ta täna on,
07:58
and I'm gratefultänulik for that.
179
463000
2000
ja ma olen selle eest tänulik.
08:00
But we really need you to make sure
180
465000
3000
Kuid me peame olema kindlad,
08:03
that these algorithmsalgoritmid have encodedkodeeritud in them
181
468000
3000
et nende poolt kirjutatud algoritmid sisaldavad
08:06
a sensemeel of the publicavalik life, a sensemeel of civickodanik responsibilityvastutus.
182
471000
3000
arusaama avalikust elust, mingisugust kodanikuvastutust.
08:09
We need you to make sure that they're transparentläbipaistev enoughpiisav
183
474000
3000
Peame kindlad olema, et nad on piisavalt avatud,
08:12
that we can see what the rulesreeglid are
184
477000
2000
et aru saada, milliste reeglitega on tegu,
08:14
that determinemäärama what getssaab throughläbi our filtersfiltrid.
185
479000
3000
mis filtreerimist reguleerivad.
08:17
And we need you to give us some controlkontroll
186
482000
2000
Me tahame, et te annaksite osa kontrollist meile,
08:19
so that we can decideotsustama
187
484000
2000
et saaksime ise otsustada,
08:21
what getssaab throughläbi and what doesn't.
188
486000
3000
mis filtrist läbi pääseb ja mis mitte.
08:24
Because I think
189
489000
2000
Sest ma arvan,
08:26
we really need the InternetInternet to be that thing
190
491000
2000
et internet peab olema see koht,
08:28
that we all dreamedunistanud of it beingolemine.
191
493000
2000
millest me kõik unistanud oleme.
08:30
We need it to connectühendage us all togetherkoos.
192
495000
3000
See peab meid kõiki omavahel ühendama.
08:33
We need it to introducetutvustama us to newuus ideasideid
193
498000
3000
See peab tutvustama meile uusi ideid,
08:36
and newuus people and differenterinevad perspectivesperspektiivid.
194
501000
3000
uusi inimesi, erinevaid seisukohti.
08:40
And it's not going to do that
195
505000
2000
Seda ei juhtu,
08:42
if it leaveslehed us all isolatedisoleeritud in a WebWeb of one.
196
507000
3000
kui internet meid hoopis isoleerima hakkab.
08:45
Thank you.
197
510000
2000
Aitäh.
08:47
(ApplauseAplaus)
198
512000
11000
(Aplaus)
Translated by Alo-Jarmo Küppas
Reviewed by Piret Hion

▲Back to top

ABOUT THE SPEAKER
Eli Pariser - Organizer and author
Pioneering online organizer Eli Pariser is the author of "The Filter Bubble," about how personalized search might be narrowing our worldview.

Why you should listen

Shortly after the September 11, 2001, attacks, Eli Pariser created a website calling for a multilateral approach to fighting terrorism. In the following weeks, over half a million people from 192 countries signed on, and Pariser rather unexpectedly became an online organizer. The website merged with MoveOn.org in November 2001, and Pariser -- then 20 years old -- joined the group to direct its foreign policy campaigns. He led what the New York Times Magazine called the "mainstream arm of the peace movement" -- tripling MoveOn's member base and demonstrating how large numbers of small donations could be mobilized through online engagement.

In 2004, Pariser became executive director of MoveOn. Under his leadership, MoveOn.org Political Action has grown to 5 million members and raised over $120 million from millions of small donors to support advocacy campaigns and political candidates. Pariser focused MoveOn on online-to-offline organizing, developing phone-banking tools and precinct programs in 2004 and 2006 that laid the groundwork for Barack Obama's extraordinary web-powered campaign. In 2008, Pariser transitioned the Executive Director role at MoveOn to Justin Ruben and became President of MoveOn’s board; he's now a senior fellow at the Roosevelt Institute.

His book The Filter Bubble is set for release May 12, 2011. In it, he asks how modern search tools -- the filter by which many of see the wider world -- are getting better and better and screening the wider world from us, by returning only the search results it "thinks" we want to see.

More profile about the speaker
Eli Pariser | Speaker | TED.com

Data provided by TED.

This site was created in May 2015 and the last update was on January 12, 2020. It will no longer be updated.

We are currently creating a new site called "eng.lish.video" and would be grateful if you could access it.

If you have any questions or suggestions, please feel free to write comments in your language on the contact form.

Privacy Policy

Developer's Blog

Buy Me A Coffee