ABOUT THE SPEAKER
Julian Assange - Whistleblower
Internet activist Julian Assange serves as spokesperson for WikiLeaks, a controversial, volunteer-driven website that publishes and comments on leaked documents alleging government and corporate misconduct.

Why you should listen

You could say Australian-born Julian Assange has swapped his long-time interest in network security flaws for the far-more-suspect flaws of even bigger targets: governments and corporations. Since his early 20s, he has been using network technology to prod and probe the vulnerable edges of administrative systems, but though he was a computing hobbyist first (in 1991 he was the target of hacking charges after he accessed the computers of an Australian telecom), he's now taken off his "white hat" and launched a career as one of the world's most visible human-rights activists.

He calls himself "editor in chief." He travels the globe as its spokesperson. Yet Assange's part in WikiLeaks is clearly dicier than that: he's become the face of creature that, simply, many powerful organizations would rather see the world rid of. His Wikipedia entry says he is "constantly on the move," and some speculate that his role in publishing decrypted US military video has put him in personal danger. A controversial figure, pundits debate whether his work is reckless and does more harm than good. Amnesty International recognized him with an International Media Award in 2009.

Assange studied physics and mathematics at the University of Melbourne. He wrote Strobe, the first free and open-source port scanner, and contributed to the book Underground: Tales of Hacking, Madness and Obsession on the Electronic Frontier.

More profile about the speaker
Julian Assange | Speaker | TED.com
TEDGlobal 2010

Julian Assange: Why the world needs WikiLeaks

Filmed:
2,716,210 views

The controversial website WikiLeaks collects and posts highly classified documents and video. Founder Julian Assange, who's reportedly being sought for questioning by US authorities, talks to TED's Chris Anderson about how the site operates, what it has accomplished -- and what drives him. The interview includes graphic footage of a recent US airstrike in Baghdad.
- Whistleblower
Internet activist Julian Assange serves as spokesperson for WikiLeaks, a controversial, volunteer-driven website that publishes and comments on leaked documents alleging government and corporate misconduct. Full bio

Double-click the English transcript below to play the video.

00:15
Chris Anderson: Julian, welcome.
0
0
2000
00:17
It's been reported that WikiLeaks, your baby,
1
2000
2000
00:19
has, in the last few years
2
4000
2000
00:21
has released more classified documents
3
6000
3000
00:24
than the rest of the world's media combined.
4
9000
2000
00:26
Can that possibly be true?
5
11000
2000
00:28
Julian Assange: Yeah, can it possibly be true?
6
13000
2000
00:30
It's a worry -- isn't it? -- that the rest of the world's media
7
15000
3000
00:33
is doing such a bad job
8
18000
2000
00:35
that a little group of activists
9
20000
2000
00:37
is able to release more
10
22000
2000
00:39
of that type of information
11
24000
2000
00:41
than the rest of the world press combined.
12
26000
2000
00:43
CA: How does it work?
13
28000
2000
00:45
How do people release the documents?
14
30000
3000
00:48
And how do you secure their privacy?
15
33000
3000
00:51
JA: So these are -- as far as we can tell --
16
36000
2000
00:53
classical whistleblowers,
17
38000
2000
00:55
and we have a number of ways for them
18
40000
2000
00:57
to get information to us.
19
42000
2000
00:59
So we use this state-of-the-art encryption
20
44000
2000
01:01
to bounce stuff around the Internet, to hide trails,
21
46000
2000
01:03
pass it through legal jurisdictions
22
48000
2000
01:05
like Sweden and Belgium
23
50000
3000
01:08
to enact those legal protections.
24
53000
3000
01:12
We get information in the mail,
25
57000
2000
01:14
the regular postal mail,
26
59000
3000
01:17
encrypted or not,
27
62000
2000
01:19
vet it like a regular news organization, format it --
28
64000
3000
01:22
which is sometimes something that's quite hard to do,
29
67000
3000
01:25
when you're talking about
30
70000
2000
01:27
giant databases of information --
31
72000
2000
01:29
release it to the public
32
74000
2000
01:31
and then defend ourselves
33
76000
2000
01:33
against the inevitable legal and political attacks.
34
78000
3000
01:36
CA: So you make an effort to ensure
35
81000
2000
01:38
the documents are legitimate,
36
83000
2000
01:40
but you actually
37
85000
2000
01:42
almost never know who the identity of the source is?
38
87000
3000
01:45
JA: That's right, yeah. Very rarely do we ever know,
39
90000
3000
01:49
and if we find out at some stage
40
94000
3000
01:52
then we destroy that information as soon as possible.
41
97000
3000
01:55
(Phone ring) God damn it.
42
100000
2000
01:57
(Laughter)
43
102000
4000
02:01
CA: I think that's the CIA asking what the code is
44
106000
2000
02:03
for a TED membership.
45
108000
2000
02:05
(Laughter)
46
110000
3000
02:08
So let's take [an] example, actually.
47
113000
2000
02:10
This is something
48
115000
2000
02:12
you leaked a few years ago.
49
117000
2000
02:14
If we can have this document up ...
50
119000
2000
02:16
So this was a story in Kenya a few years ago.
51
121000
2000
02:18
Can you tell us what you leaked and what happened?
52
123000
3000
02:21
JA: So this is the Kroll Report.
53
126000
2000
02:23
This was a secret intelligence report
54
128000
3000
02:26
commissioned by the Kenyan government
55
131000
2000
02:28
after its election in 2004.
56
133000
3000
02:31
Prior to 2004, Kenya was ruled
57
136000
2000
02:33
by Daniel arap Moi
58
138000
2000
02:35
for about 18 years.
59
140000
2000
02:37
He was a soft dictator of Kenya.
60
142000
3000
02:40
And when Kibaki got into power --
61
145000
2000
02:42
through a coalition of forces that were trying
62
147000
2000
02:44
to clean up corruption in Kenya --
63
149000
2000
02:46
they commissioned this report,
64
151000
2000
02:48
spent about two million pounds
65
153000
2000
02:50
on this and an associated report.
66
155000
2000
02:52
And then the government sat on it
67
157000
3000
02:55
and used it for political leverage on Moi,
68
160000
2000
02:57
who was the richest man --
69
162000
2000
02:59
still is the richest man -- in Kenya.
70
164000
3000
03:02
It's the Holy Grail of Kenyan journalism.
71
167000
3000
03:05
So I went there in 2007,
72
170000
3000
03:08
and we managed to get hold of this
73
173000
2000
03:10
just prior to the election --
74
175000
2000
03:12
the national election, December 28.
75
177000
3000
03:17
When we released that report,
76
182000
3000
03:20
we did so three days after the new president, Kibaki,
77
185000
3000
03:23
had decided to pal up with
78
188000
2000
03:25
the man that he was going to clean out,
79
190000
2000
03:27
Daniel arap Moi,
80
192000
2000
03:29
so this report then
81
194000
3000
03:32
became a dead albatross
82
197000
2000
03:34
around President Kibaki's neck.
83
199000
3000
03:38
CA: And -- I mean, to cut a long story short --
84
203000
3000
03:41
word of the report leaked into Kenya,
85
206000
3000
03:44
not from the official media, but indirectly,
86
209000
3000
03:47
and in your opinion, it actually shifted the election.
87
212000
3000
03:50
JA: Yeah. So this became front page of the Guardian
88
215000
3000
03:53
and was then printed in all the surrounding countries of Kenya,
89
218000
3000
03:56
in Tanzanian and South African press.
90
221000
3000
03:59
And so it came in from the outside.
91
224000
2000
04:01
And that, after a couple of days,
92
226000
2000
04:03
made the Kenyan press feel safe to talk about it.
93
228000
2000
04:05
And it ran for 20 nights straight on Kenyan TV,
94
230000
3000
04:08
shifted the vote by 10 percent,
95
233000
3000
04:11
according to a Kenyan intelligence report,
96
236000
2000
04:13
which changed the result of the election.
97
238000
2000
04:15
CA: Wow, so your leak
98
240000
2000
04:17
really substantially changed the world?
99
242000
2000
04:19
JA: Yep.
100
244000
2000
04:21
(Applause)
101
246000
4000
04:25
CA: Here's -- We're going to just show
102
250000
2000
04:27
a short clip from this
103
252000
3000
04:30
Baghdad airstrike video.
104
255000
2000
04:32
The video itself is longer,
105
257000
2000
04:34
but here's a short clip.
106
259000
2000
04:36
This is -- this is intense material, I should warn you.
107
261000
3000
04:39
Radio: ... just fuckin', once you get on 'em just open 'em up.
108
264000
3000
04:42
I see your element, uh, got about four Humvees, uh, out along ...
109
267000
4000
04:46
You're clear. All right. Firing.
110
271000
3000
04:49
Let me know when you've got them. Let's shoot.
111
274000
3000
04:52
Light 'em all up.
112
277000
2000
04:54
C'mon, fire!
113
279000
2000
04:56
(Machine gun fire)
114
281000
3000
04:59
Keep shoot 'n. Keep shoot 'n.
115
284000
3000
05:02
(Machine gun fire)
116
287000
3000
05:05
Keep shoot 'n.
117
290000
3000
05:08
Hotel ... Bushmaster Two-Six, Bushmaster Two-Six,
118
293000
2000
05:10
we need to move, time now!
119
295000
2000
05:12
All right, we just engaged all eight individuals.
120
297000
3000
05:15
Yeah, we see two birds [helicopters], and we're still firing.
121
300000
3000
05:18
Roger. I got 'em.
122
303000
2000
05:20
Two-Six, this is Two-Six, we're mobile.
123
305000
2000
05:22
Oops, I'm sorry. What was going on?
124
307000
2000
05:24
God damn it, Kyle. All right, hahaha. I hit 'em.
125
309000
2000
05:29
CA: So, what was the impact of that?
126
314000
3000
05:32
JA: The impact on the people who worked on it
127
317000
3000
05:35
was severe.
128
320000
2000
05:37
We ended up sending two people to Baghdad
129
322000
2000
05:39
to further research that story.
130
324000
2000
05:41
So this is just the first of three attacks
131
326000
3000
05:44
that occurred in that scene.
132
329000
2000
05:46
CA: So, I mean, 11 people died in that attack, right,
133
331000
2000
05:48
including two Reuters employees?
134
333000
2000
05:50
JA: Yeah. Two Reuters employees,
135
335000
2000
05:52
two young children were wounded.
136
337000
3000
05:55
There were between 18 and 26 people killed all together.
137
340000
3000
05:58
CA: And releasing this caused
138
343000
2000
06:00
widespread outrage.
139
345000
2000
06:02
What was the key element of this
140
347000
2000
06:04
that actually caused the outrage, do you think?
141
349000
3000
06:07
JA: I don't know. I guess people can see
142
352000
2000
06:09
the gross disparity in force.
143
354000
3000
06:12
You have guys walking in a relaxed way down the street,
144
357000
2000
06:14
and then an Apache helicopter sitting up at one kilometer
145
359000
3000
06:17
firing 30-millimeter cannon shells
146
362000
2000
06:19
on everyone --
147
364000
2000
06:21
looking for any excuse to do so --
148
366000
3000
06:24
and killing people rescuing the wounded.
149
369000
2000
06:26
And there was two journalists involved that clearly weren't insurgents
150
371000
3000
06:29
because that's their full-time job.
151
374000
2000
06:33
CA: I mean, there's been this U.S. intelligence analyst,
152
378000
3000
06:36
Bradley Manning, arrested,
153
381000
2000
06:38
and it's alleged that he confessed in a chat room
154
383000
3000
06:41
to have leaked this video to you,
155
386000
3000
06:44
along with 280,000
156
389000
2000
06:46
classified U.S. embassy cables.
157
391000
2000
06:48
I mean, did he?
158
393000
3000
06:51
JA: We have denied receiving those cables.
159
396000
2000
06:53
He has been charged,
160
398000
2000
06:55
about five days ago,
161
400000
2000
06:57
with obtaining 150,000 cables
162
402000
3000
07:00
and releasing 50.
163
405000
2000
07:02
Now, we had released,
164
407000
3000
07:05
early in the year,
165
410000
2000
07:07
a cable from the Reykjavik U.S. embassy,
166
412000
3000
07:11
but this is not necessarily connected.
167
416000
2000
07:13
I mean, I was a known visitor of that embassy.
168
418000
2000
07:15
CA: I mean, if you did receive thousands
169
420000
2000
07:17
of U.S. embassy diplomatic cables ...
170
422000
3000
07:20
JA: We would have released them. (CA: You would?)
171
425000
2000
07:22
JA: Yeah. (CA: Because?)
172
427000
3000
07:25
JA: Well, because these sort of things
173
430000
2000
07:27
reveal what the true state
174
432000
3000
07:30
of, say,
175
435000
2000
07:32
Arab governments are like,
176
437000
2000
07:34
the true human-rights abuses in those governments.
177
439000
3000
07:37
If you look at declassified cables,
178
442000
2000
07:39
that's the sort of material that's there.
179
444000
2000
07:41
CA: So let's talk a little more broadly about this.
180
446000
2000
07:43
I mean, in general, what's your philosophy?
181
448000
2000
07:45
Why is it right
182
450000
2000
07:47
to encourage leaking of secret information?
183
452000
3000
07:51
JA: Well, there's a question as to what sort of information is important in the world,
184
456000
3000
07:54
what sort of information
185
459000
2000
07:56
can achieve reform.
186
461000
2000
07:58
And there's a lot of information.
187
463000
2000
08:00
So information that organizations
188
465000
2000
08:02
are spending economic effort into concealing,
189
467000
3000
08:05
that's a really good signal
190
470000
2000
08:07
that when the information gets out,
191
472000
2000
08:09
there's a hope of it doing some good --
192
474000
2000
08:11
because the organizations that know it best,
193
476000
2000
08:13
that know it from the inside out,
194
478000
2000
08:15
are spending work to conceal it.
195
480000
3000
08:18
And that's what we've found in practice,
196
483000
2000
08:20
and that's what the history of journalism is.
197
485000
3000
08:23
CA: But are there risks with that,
198
488000
3000
08:26
either to the individuals concerned
199
491000
3000
08:29
or indeed to society at large,
200
494000
2000
08:31
where leaking can actually have
201
496000
2000
08:33
an unintended consequence?
202
498000
2000
08:35
JA: Not that we have seen with anything we have released.
203
500000
2000
08:37
I mean, we have a harm immunization policy.
204
502000
2000
08:39
We have a way of dealing with information
205
504000
2000
08:41
that has sort of personal --
206
506000
2000
08:43
personally identifying information in it.
207
508000
2000
08:46
But there are legitimate secrets --
208
511000
3000
08:49
you know, your records with your doctor;
209
514000
3000
08:52
that's a legitimate secret --
210
517000
2000
08:54
but we deal with whistleblowers that are coming forward
211
519000
2000
08:56
that are really sort of well-motivated.
212
521000
3000
08:59
CA: So they are well-motivated.
213
524000
2000
09:01
And what would you say to, for example,
214
526000
2000
09:03
the, you know, the parent of someone
215
528000
3000
09:06
whose son is out serving the U.S. military,
216
531000
3000
09:09
and he says, "You know what,
217
534000
2000
09:11
you've put up something that someone had an incentive to put out.
218
536000
2000
09:13
It shows a U.S. soldier laughing
219
538000
2000
09:15
at people dying.
220
540000
2000
09:17
That gives the impression, has given the impression,
221
542000
2000
09:19
to millions of people around the world
222
544000
2000
09:21
that U.S. soldiers are inhuman people.
223
546000
2000
09:23
Actually, they're not. My son isn't. How dare you?"
224
548000
2000
09:25
What would you say to that?
225
550000
2000
09:27
JA: Yeah, we do get a lot of that.
226
552000
2000
09:29
But remember, the people in Baghdad,
227
554000
2000
09:31
the people in Iraq, the people in Afghanistan --
228
556000
3000
09:34
they don't need to see the video;
229
559000
2000
09:36
they see it every day.
230
561000
2000
09:38
So it's not going to change their opinion. It's not going to change their perception.
231
563000
3000
09:41
That's what they see every day.
232
566000
2000
09:43
It will change the perception and opinion
233
568000
3000
09:46
of the people who are paying for it all,
234
571000
2000
09:48
and that's our hope.
235
573000
3000
09:51
CA: So you found a way to shine light
236
576000
3000
09:54
into what you see
237
579000
3000
09:57
as these sort of dark secrets in companies and in government.
238
582000
3000
10:01
Light is good.
239
586000
2000
10:03
But do you see any irony in the fact that,
240
588000
2000
10:05
in order for you to shine that light,
241
590000
2000
10:07
you have to, yourself,
242
592000
2000
10:09
create secrecy around your sources?
243
594000
3000
10:12
JA: Not really. I mean, we don't have
244
597000
3000
10:15
any WikiLeaks dissidents yet.
245
600000
3000
10:19
We don't have sources who are dissidents on other sources.
246
604000
3000
10:23
Should they come forward, that would be a tricky situation for us,
247
608000
3000
10:26
but we're presumably acting in such a way
248
611000
3000
10:29
that people feel
249
614000
2000
10:31
morally compelled
250
616000
2000
10:33
to continue our mission, not to screw it up.
251
618000
3000
10:37
CA: I'd actually be interested, just based on what we've heard so far --
252
622000
3000
10:40
I'm curious as to the opinion in the TED audience.
253
625000
3000
10:45
You know, there might be a couple of views
254
630000
2000
10:47
of WikiLeaks and of Julian.
255
632000
2000
10:49
You know, hero -- people's hero --
256
634000
3000
10:52
bringing this important light.
257
637000
3000
10:55
Dangerous troublemaker.
258
640000
2000
10:58
Who's got the hero view?
259
643000
3000
11:02
Who's got the dangerous troublemaker view?
260
647000
3000
11:06
JA: Oh, come on. There must be some.
261
651000
2000
11:09
CA: It's a soft crowd, Julian, a soft crowd.
262
654000
2000
11:11
We have to try better. Let's show them another example.
263
656000
2000
11:13
Now here's something that you haven't yet leaked,
264
658000
3000
11:16
but I think for TED you are.
265
661000
3000
11:19
I mean it's an intriguing story that's just happened, right?
266
664000
2000
11:21
What is this?
267
666000
2000
11:23
JA: So this is a sample of what we do
268
668000
2000
11:25
sort of every day.
269
670000
2000
11:27
So late last year -- in November last year --
270
672000
3000
11:30
there was a series of well blowouts
271
675000
2000
11:32
in Albania,
272
677000
2000
11:34
like the well blowout in the Gulf of Mexico,
273
679000
3000
11:37
but not quite as big.
274
682000
2000
11:39
And we got a report --
275
684000
3000
11:42
a sort of engineering analysis into what happened --
276
687000
3000
11:45
saying that, in fact, security guards
277
690000
3000
11:48
from some rival, various competing oil firms
278
693000
3000
11:51
had, in fact, parked trucks there and blown them up.
279
696000
3000
11:55
And part of the Albanian government was in this, etc., etc.
280
700000
3000
11:59
And the engineering report
281
704000
1000
12:00
had nothing on the top of it,
282
705000
2000
12:02
so it was an extremely difficult document for us.
283
707000
2000
12:04
We couldn't verify it because we didn't know
284
709000
2000
12:06
who wrote it and knew what it was about.
285
711000
2000
12:08
So we were kind of skeptical that maybe it was
286
713000
2000
12:10
a competing oil firm just sort of playing the issue up.
287
715000
2000
12:12
So under that basis, we put it out and said,
288
717000
2000
12:14
"Look, we're skeptical about this thing.
289
719000
2000
12:16
We don't know, but what can we do?
290
721000
2000
12:18
The material looks good, it feels right,
291
723000
2000
12:20
but we just can't verify it."
292
725000
2000
12:22
And we then got a letter
293
727000
3000
12:25
just this week
294
730000
3000
12:28
from the company who wrote it,
295
733000
3000
12:31
wanting to track down the source --
296
736000
3000
12:34
(Laughter)
297
739000
3000
12:38
saying, "Hey, we want to track down the source."
298
743000
3000
12:41
And we were like, "Oh, tell us more.
299
746000
2000
12:43
What document is it, precisely, you're talking about?
300
748000
3000
12:46
Can you show that you had legal authority over that document?
301
751000
3000
12:49
Is it really yours?"
302
754000
2000
12:51
So they sent us this screen shot
303
756000
3000
12:54
with the author
304
759000
2000
12:56
in the Microsoft Word ID.
305
761000
3000
13:01
Yeah.
306
766000
2000
13:03
(Applause)
307
768000
5000
13:08
That's happened quite a lot though.
308
773000
2000
13:10
This is like one of our methods
309
775000
2000
13:12
of identifying, of verifying, what a material is,
310
777000
3000
13:15
is to try and get these guys to write letters.
311
780000
2000
13:17
CA: Yeah. Have you had information
312
782000
3000
13:20
from inside BP?
313
785000
2000
13:22
JA: Yeah, we have a lot, but I mean, at the moment,
314
787000
3000
13:25
we are undergoing a sort of serious fundraising and engineering effort.
315
790000
3000
13:28
So our publication rate
316
793000
2000
13:30
over the past few months
317
795000
2000
13:32
has been sort of minimized
318
797000
2000
13:34
while we're re-engineering our back systems
319
799000
3000
13:37
for the phenomenal public interest that we have.
320
802000
3000
13:40
That's a problem.
321
805000
2000
13:42
I mean, like any sort of growing startup organization,
322
807000
3000
13:45
we are sort of overwhelmed
323
810000
2000
13:47
by our growth,
324
812000
2000
13:49
and that means we're getting enormous quantity
325
814000
2000
13:51
of whistleblower disclosures
326
816000
2000
13:53
of a very high caliber
327
818000
2000
13:55
but don't have enough people to actually
328
820000
2000
13:57
process and vet this information.
329
822000
2000
13:59
CA: So that's the key bottleneck,
330
824000
2000
14:01
basically journalistic volunteers
331
826000
2000
14:03
and/or the funding of journalistic salaries?
332
828000
3000
14:06
JA: Yep. Yeah, and trusted people.
333
831000
2000
14:08
I mean, we're an organization
334
833000
2000
14:10
that is hard to grow very quickly
335
835000
2000
14:12
because of the sort of material we deal with,
336
837000
2000
14:14
so we have to restructure
337
839000
3000
14:17
in order to have people
338
842000
2000
14:19
who will deal with the highest national security stuff,
339
844000
3000
14:22
and then lower security cases.
340
847000
2000
14:24
CA: So help us understand a bit about you personally
341
849000
3000
14:27
and how you came to do this.
342
852000
2000
14:29
And I think I read that as a kid
343
854000
2000
14:31
you went to 37 different schools.
344
856000
3000
14:34
Can that be right?
345
859000
2000
14:36
JA: Well, my parents were in the movie business
346
861000
3000
14:39
and then on the run from a cult,
347
864000
2000
14:41
so the combination between the two ...
348
866000
2000
14:43
(Laughter)
349
868000
4000
14:47
CA: I mean, a psychologist might say
350
872000
2000
14:49
that's a recipe for breeding paranoia.
351
874000
3000
14:52
JA: What, the movie business?
352
877000
2000
14:54
(Laughter)
353
879000
3000
14:57
(Applause)
354
882000
3000
15:00
CA: And you were also -- I mean,
355
885000
2000
15:02
you were also a hacker at an early age
356
887000
2000
15:04
and ran into the authorities early on.
357
889000
3000
15:07
JA: Well, I was a journalist.
358
892000
3000
15:10
You know, I was a very young journalist activist at an early age.
359
895000
2000
15:12
I wrote a magazine,
360
897000
2000
15:14
was prosecuted for it when I was a teenager.
361
899000
3000
15:17
So you have to be careful with hacker.
362
902000
2000
15:19
I mean there's like -- there's a method
363
904000
2000
15:21
that can be deployed for various things.
364
906000
2000
15:23
Unfortunately, at the moment,
365
908000
2000
15:25
it's mostly deployed by the Russian mafia
366
910000
2000
15:27
in order to steal your grandmother's bank accounts.
367
912000
2000
15:29
So this phrase is not,
368
914000
3000
15:32
not as nice as it used to be.
369
917000
2000
15:34
CA: Yeah, well, I certainly don't think
370
919000
2000
15:36
you're stealing anyone's grandmother's bank account,
371
921000
3000
15:39
but what about
372
924000
2000
15:41
your core values?
373
926000
2000
15:43
Can you give us a sense of what they are
374
928000
3000
15:46
and maybe some incident in your life
375
931000
2000
15:48
that helped determine them?
376
933000
3000
15:53
JA: I'm not sure about the incident.
377
938000
2000
15:55
But the core values:
378
940000
3000
15:58
well, capable, generous men
379
943000
3000
16:01
do not create victims;
380
946000
2000
16:03
they nurture victims.
381
948000
2000
16:05
And that's something from my father
382
950000
2000
16:07
and something from other capable, generous men
383
952000
3000
16:10
that have been in my life.
384
955000
3000
16:13
CA: Capable, generous men do not create victims;
385
958000
2000
16:15
they nurture victims?
386
960000
2000
16:17
JA: Yeah. And you know,
387
962000
2000
16:19
I'm a combative person,
388
964000
4000
16:23
so I'm not actually so big on the nurture,
389
968000
2000
16:25
but some way --
390
970000
3000
16:28
there is another way of nurturing victims,
391
973000
3000
16:31
which is to police perpetrators
392
976000
3000
16:34
of crime.
393
979000
2000
16:36
And so that is something
394
981000
2000
16:38
that has been in my character
395
983000
2000
16:40
for a long time.
396
985000
2000
16:42
CA: So just tell us, very quickly in the last minute, the story:
397
987000
3000
16:45
what happened in Iceland?
398
990000
3000
16:48
You basically published something there,
399
993000
3000
16:51
ran into trouble with a bank,
400
996000
3000
16:54
then the news service there
401
999000
2000
16:56
was injuncted from running the story.
402
1001000
3000
16:59
Instead, they publicized your side.
403
1004000
2000
17:01
That made you very high-profile in Iceland. What happened next?
404
1006000
3000
17:04
JA: Yeah, this is a great case, you know.
405
1009000
2000
17:06
Iceland went through this financial crisis.
406
1011000
2000
17:08
It was the hardest hit of any country in the world.
407
1013000
2000
17:10
Its banking sector was 10 times the GDP
408
1015000
2000
17:12
of the rest of the economy.
409
1017000
2000
17:14
Anyway, so we release this report
410
1019000
3000
17:17
in July last year.
411
1022000
3000
17:20
And the national TV station was injuncted
412
1025000
2000
17:22
five minutes before it went on air,
413
1027000
2000
17:24
like out of a movie: injunction landed on the news desk,
414
1029000
2000
17:26
and the news reader was like,
415
1031000
2000
17:28
"This has never happened before. What do we do?"
416
1033000
2000
17:30
Well, we just show the website instead,
417
1035000
2000
17:32
for all that time, as a filler,
418
1037000
3000
17:35
and we became very famous in Iceland,
419
1040000
2000
17:37
went to Iceland and spoke about this issue.
420
1042000
3000
17:40
And there was a feeling in the community
421
1045000
2000
17:42
that that should never happen again,
422
1047000
2000
17:44
and as a result,
423
1049000
2000
17:46
working with Icelandic politicians
424
1051000
2000
17:48
and some other international legal experts,
425
1053000
2000
17:50
we put together a new sort of
426
1055000
2000
17:52
package of legislation for Iceland
427
1057000
3000
17:55
to sort of become an offshore haven
428
1060000
3000
17:58
for the free press,
429
1063000
3000
18:01
with the strongest journalistic protections in the world,
430
1066000
3000
18:04
with a new Nobel Prize
431
1069000
2000
18:06
for freedom of speech.
432
1071000
2000
18:08
Iceland's a Nordic country,
433
1073000
2000
18:10
so, like Norway, it's able to tap into the system.
434
1075000
3000
18:13
And just a month ago,
435
1078000
2000
18:15
this was passed by the Icelandic parliament unanimously.
436
1080000
3000
18:18
CA: Wow.
437
1083000
2000
18:20
(Applause)
438
1085000
6000
18:26
Last question, Julian.
439
1091000
2000
18:28
When you think of the future then,
440
1093000
2000
18:30
do you think it's more likely to be
441
1095000
2000
18:32
Big Brother exerting more control,
442
1097000
2000
18:34
more secrecy,
443
1099000
2000
18:36
or us watching
444
1101000
2000
18:38
Big Brother,
445
1103000
2000
18:40
or it's just all to be played for either way?
446
1105000
3000
18:43
JA: I'm not sure which way it's going to go.
447
1108000
2000
18:45
I mean, there's enormous pressures
448
1110000
2000
18:47
to harmonize freedom of speech legislation
449
1112000
3000
18:50
and transparency legislation around the world --
450
1115000
3000
18:53
within the E.U.,
451
1118000
2000
18:55
between China and the United States.
452
1120000
2000
18:57
Which way is it going to go? It's hard to see.
453
1122000
3000
19:00
That's why it's a very interesting time to be in --
454
1125000
2000
19:02
because with just a little bit of effort,
455
1127000
2000
19:04
we can shift it one way or the other.
456
1129000
3000
19:07
CA: Well, it looks like I'm reflecting the audience's opinion
457
1132000
3000
19:10
to say, Julian, be careful,
458
1135000
2000
19:12
and all power to you.
459
1137000
2000
19:14
JA: Thank you, Chris. (CA: Thank you.)
460
1139000
2000
19:16
(Applause)
461
1141000
10000

▲Back to top

ABOUT THE SPEAKER
Julian Assange - Whistleblower
Internet activist Julian Assange serves as spokesperson for WikiLeaks, a controversial, volunteer-driven website that publishes and comments on leaked documents alleging government and corporate misconduct.

Why you should listen

You could say Australian-born Julian Assange has swapped his long-time interest in network security flaws for the far-more-suspect flaws of even bigger targets: governments and corporations. Since his early 20s, he has been using network technology to prod and probe the vulnerable edges of administrative systems, but though he was a computing hobbyist first (in 1991 he was the target of hacking charges after he accessed the computers of an Australian telecom), he's now taken off his "white hat" and launched a career as one of the world's most visible human-rights activists.

He calls himself "editor in chief." He travels the globe as its spokesperson. Yet Assange's part in WikiLeaks is clearly dicier than that: he's become the face of creature that, simply, many powerful organizations would rather see the world rid of. His Wikipedia entry says he is "constantly on the move," and some speculate that his role in publishing decrypted US military video has put him in personal danger. A controversial figure, pundits debate whether his work is reckless and does more harm than good. Amnesty International recognized him with an International Media Award in 2009.

Assange studied physics and mathematics at the University of Melbourne. He wrote Strobe, the first free and open-source port scanner, and contributed to the book Underground: Tales of Hacking, Madness and Obsession on the Electronic Frontier.

More profile about the speaker
Julian Assange | Speaker | TED.com