ABOUT THE SPEAKER
Bruce Schneier - Security expert
Bruce Schneier thinks hard about security -- as a computer security guru, and as a philosopher of the larger notion of making a safer world.

Why you should listen

Bruce Schneier is an internationally renowned security technologist and author. Described by the Economist as a "security guru," he is best known as a refreshingly candid and lucid security critic and commentator. When people want to know how security really works, they turn to Schneier.

His first bestseller, Applied Cryptography, explained how the arcane science of secret codes actually works, and was described by Wired as "the book the National Security Agency wanted never to be published." His book on computer and network security, Secrets and Lies, was called by Fortune "[a] jewel box of little surprises you can actually use." Beyond Fear tackles the problems of security from the small to the large: personal safety, crime, corporate security, national security. His current book, Schneier on Security, offers insight into everything from the risk of identity theft (vastly overrated) to the long-range security threat of unchecked presidential power and the surprisingly simple way to tamper-proof elections.

Schneier publishes a free monthly newsletter, Crypto-Gram, with over 150,000 readers. In its ten years of regular publication, Crypto-Gram has become one of the most widely read forums for free-wheeling discussions, pointed critiques and serious debate about security. As head curmudgeon at the table, Schneier explains, debunks and draws lessons from security stories that make the news.

More profile about the speaker
Bruce Schneier | Speaker | TED.com
TEDxPSU

Bruce Schneier: The security mirage

Filmed:
958,315 views

The feeling of security and the reality of security don't always match, says computer-security expert Bruce Schneier. In his talk, he explains why we spend billions addressing news story risks, like the "security theater" now playing at your local airport, while neglecting more probable risks -- and how we can break this pattern.
- Security expert
Bruce Schneier thinks hard about security -- as a computer security guru, and as a philosopher of the larger notion of making a safer world. Full bio

Double-click the English transcript below to play the video.

00:15
So security is two different things:
0
0
2000
00:17
it's a feeling, and it's a reality.
1
2000
2000
00:19
And they're different.
2
4000
2000
00:21
You could feel secure
3
6000
2000
00:23
even if you're not.
4
8000
2000
00:25
And you can be secure
5
10000
2000
00:27
even if you don't feel it.
6
12000
2000
00:29
Really, we have two separate concepts
7
14000
2000
00:31
mapped onto the same word.
8
16000
2000
00:33
And what I want to do in this talk
9
18000
2000
00:35
is to split them apart --
10
20000
2000
00:37
figuring out when they diverge
11
22000
2000
00:39
and how they converge.
12
24000
2000
00:41
And language is actually a problem here.
13
26000
2000
00:43
There aren't a lot of good words
14
28000
2000
00:45
for the concepts we're going to talk about.
15
30000
3000
00:48
So if you look at security
16
33000
2000
00:50
from economic terms,
17
35000
2000
00:52
it's a trade-off.
18
37000
2000
00:54
Every time you get some security,
19
39000
2000
00:56
you're always trading off something.
20
41000
2000
00:58
Whether this is a personal decision --
21
43000
2000
01:00
whether you're going to install a burglar alarm in your home --
22
45000
2000
01:02
or a national decision -- where you're going to invade some foreign country --
23
47000
3000
01:05
you're going to trade off something,
24
50000
2000
01:07
either money or time, convenience, capabilities,
25
52000
3000
01:10
maybe fundamental liberties.
26
55000
3000
01:13
And the question to ask when you look at a security anything
27
58000
3000
01:16
is not whether this makes us safer,
28
61000
3000
01:19
but whether it's worth the trade-off.
29
64000
3000
01:22
You've heard in the past several years,
30
67000
2000
01:24
the world is safer because Saddam Hussein is not in power.
31
69000
2000
01:26
That might be true, but it's not terribly relevant.
32
71000
3000
01:29
The question is, was it worth it?
33
74000
3000
01:32
And you can make your own decision,
34
77000
3000
01:35
and then you'll decide whether the invasion was worth it.
35
80000
2000
01:37
That's how you think about security --
36
82000
2000
01:39
in terms of the trade-off.
37
84000
2000
01:41
Now there's often no right or wrong here.
38
86000
3000
01:44
Some of us have a burglar alarm system at home,
39
89000
2000
01:46
and some of us don't.
40
91000
2000
01:48
And it'll depend on where we live,
41
93000
2000
01:50
whether we live alone or have a family,
42
95000
2000
01:52
how much cool stuff we have,
43
97000
2000
01:54
how much we're willing to accept
44
99000
2000
01:56
the risk of theft.
45
101000
2000
01:58
In politics also,
46
103000
2000
02:00
there are different opinions.
47
105000
2000
02:02
A lot of times, these trade-offs
48
107000
2000
02:04
are about more than just security,
49
109000
2000
02:06
and I think that's really important.
50
111000
2000
02:08
Now people have a natural intuition
51
113000
2000
02:10
about these trade-offs.
52
115000
2000
02:12
We make them every day --
53
117000
2000
02:14
last night in my hotel room,
54
119000
2000
02:16
when I decided to double-lock the door,
55
121000
2000
02:18
or you in your car when you drove here,
56
123000
2000
02:20
when we go eat lunch
57
125000
2000
02:22
and decide the food's not poison and we'll eat it.
58
127000
3000
02:25
We make these trade-offs again and again,
59
130000
2000
02:27
multiple times a day.
60
132000
2000
02:29
We often won't even notice them.
61
134000
2000
02:31
They're just part of being alive; we all do it.
62
136000
2000
02:33
Every species does it.
63
138000
3000
02:36
Imagine a rabbit in a field, eating grass,
64
141000
2000
02:38
and the rabbit's going to see a fox.
65
143000
3000
02:41
That rabbit will make a security trade-off:
66
146000
2000
02:43
"Should I stay, or should I flee?"
67
148000
2000
02:45
And if you think about it,
68
150000
2000
02:47
the rabbits that are good at making that trade-off
69
152000
3000
02:50
will tend to live and reproduce,
70
155000
2000
02:52
and the rabbits that are bad at it
71
157000
2000
02:54
will get eaten or starve.
72
159000
2000
02:56
So you'd think
73
161000
2000
02:58
that us, as a successful species on the planet --
74
163000
3000
03:01
you, me, everybody --
75
166000
2000
03:03
would be really good at making these trade-offs.
76
168000
3000
03:06
Yet it seems, again and again,
77
171000
2000
03:08
that we're hopelessly bad at it.
78
173000
3000
03:11
And I think that's a fundamentally interesting question.
79
176000
3000
03:14
I'll give you the short answer.
80
179000
2000
03:16
The answer is, we respond to the feeling of security
81
181000
2000
03:18
and not the reality.
82
183000
3000
03:21
Now most of the time, that works.
83
186000
3000
03:25
Most of the time,
84
190000
2000
03:27
feeling and reality are the same.
85
192000
3000
03:30
Certainly that's true
86
195000
2000
03:32
for most of human prehistory.
87
197000
3000
03:35
We've developed this ability
88
200000
3000
03:38
because it makes evolutionary sense.
89
203000
2000
03:40
One way to think of it
90
205000
2000
03:42
is that we're highly optimized
91
207000
2000
03:44
for risk decisions
92
209000
2000
03:46
that are endemic to living in small family groups
93
211000
3000
03:49
in the East African highlands in 100,000 B.C.
94
214000
3000
03:52
2010 New York, not so much.
95
217000
3000
03:56
Now there are several biases in risk perception.
96
221000
3000
03:59
A lot of good experiments in this.
97
224000
2000
04:01
And you can see certain biases that come up again and again.
98
226000
3000
04:04
So I'll give you four.
99
229000
2000
04:06
We tend to exaggerate spectacular and rare risks
100
231000
3000
04:09
and downplay common risks --
101
234000
2000
04:11
so flying versus driving.
102
236000
3000
04:14
The unknown is perceived
103
239000
2000
04:16
to be riskier than the familiar.
104
241000
3000
04:20
One example would be,
105
245000
2000
04:22
people fear kidnapping by strangers
106
247000
3000
04:25
when the data supports kidnapping by relatives is much more common.
107
250000
3000
04:28
This is for children.
108
253000
2000
04:30
Third, personified risks
109
255000
3000
04:33
are perceived to be greater than anonymous risks --
110
258000
3000
04:36
so Bin Laden is scarier because he has a name.
111
261000
3000
04:39
And the fourth
112
264000
2000
04:41
is people underestimate risks
113
266000
2000
04:43
in situations they do control
114
268000
2000
04:45
and overestimate them in situations they don't control.
115
270000
4000
04:49
So once you take up skydiving or smoking,
116
274000
3000
04:52
you downplay the risks.
117
277000
2000
04:54
If a risk is thrust upon you -- terrorism was a good example --
118
279000
3000
04:57
you'll overplay it because you don't feel like it's in your control.
119
282000
3000
05:02
There are a bunch of other of these biases, these cognitive biases,
120
287000
3000
05:05
that affect our risk decisions.
121
290000
3000
05:08
There's the availability heuristic,
122
293000
2000
05:10
which basically means
123
295000
2000
05:12
we estimate the probability of something
124
297000
3000
05:15
by how easy it is to bring instances of it to mind.
125
300000
4000
05:19
So you can imagine how that works.
126
304000
2000
05:21
If you hear a lot about tiger attacks, there must be a lot of tigers around.
127
306000
3000
05:24
You don't hear about lion attacks, there aren't a lot of lions around.
128
309000
3000
05:27
This works until you invent newspapers.
129
312000
3000
05:30
Because what newspapers do
130
315000
2000
05:32
is they repeat again and again
131
317000
2000
05:34
rare risks.
132
319000
2000
05:36
I tell people, if it's in the news, don't worry about it.
133
321000
2000
05:38
Because by definition,
134
323000
2000
05:40
news is something that almost never happens.
135
325000
3000
05:43
(Laughter)
136
328000
2000
05:45
When something is so common, it's no longer news --
137
330000
3000
05:48
car crashes, domestic violence --
138
333000
2000
05:50
those are the risks you worry about.
139
335000
3000
05:53
We're also a species of storytellers.
140
338000
2000
05:55
We respond to stories more than data.
141
340000
3000
05:58
And there's some basic innumeracy going on.
142
343000
2000
06:00
I mean, the joke "One, Two, Three, Many" is kind of right.
143
345000
3000
06:03
We're really good at small numbers.
144
348000
3000
06:06
One mango, two mangoes, three mangoes,
145
351000
2000
06:08
10,000 mangoes, 100,000 mangoes --
146
353000
2000
06:10
it's still more mangoes you can eat before they rot.
147
355000
3000
06:13
So one half, one quarter, one fifth -- we're good at that.
148
358000
3000
06:16
One in a million, one in a billion --
149
361000
2000
06:18
they're both almost never.
150
363000
3000
06:21
So we have trouble with the risks
151
366000
2000
06:23
that aren't very common.
152
368000
2000
06:25
And what these cognitive biases do
153
370000
2000
06:27
is they act as filters between us and reality.
154
372000
3000
06:30
And the result
155
375000
2000
06:32
is that feeling and reality get out of whack,
156
377000
2000
06:34
they get different.
157
379000
3000
06:37
Now you either have a feeling -- you feel more secure than you are.
158
382000
3000
06:40
There's a false sense of security.
159
385000
2000
06:42
Or the other way,
160
387000
2000
06:44
and that's a false sense of insecurity.
161
389000
2000
06:46
I write a lot about "security theater,"
162
391000
3000
06:49
which are products that make people feel secure,
163
394000
3000
06:52
but don't actually do anything.
164
397000
2000
06:54
There's no real word for stuff that makes us secure,
165
399000
2000
06:56
but doesn't make us feel secure.
166
401000
2000
06:58
Maybe it's what the CIA's supposed to do for us.
167
403000
3000
07:03
So back to economics.
168
408000
2000
07:05
If economics, if the market, drives security,
169
410000
4000
07:09
and if people make trade-offs
170
414000
2000
07:11
based on the feeling of security,
171
416000
3000
07:14
then the smart thing for companies to do
172
419000
2000
07:16
for the economic incentives
173
421000
2000
07:18
are to make people feel secure.
174
423000
3000
07:21
And there are two ways to do this.
175
426000
3000
07:24
One, you can make people actually secure
176
429000
2000
07:26
and hope they notice.
177
431000
2000
07:28
Or two, you can make people just feel secure
178
433000
3000
07:31
and hope they don't notice.
179
436000
3000
07:35
So what makes people notice?
180
440000
3000
07:38
Well a couple of things:
181
443000
2000
07:40
understanding of the security,
182
445000
2000
07:42
of the risks, the threats,
183
447000
2000
07:44
the countermeasures, how they work.
184
449000
3000
07:47
But if you know stuff,
185
452000
2000
07:49
you're more likely to have your feelings match reality.
186
454000
3000
07:52
Enough real world examples helps.
187
457000
3000
07:55
Now we all know the crime rate in our neighborhood,
188
460000
3000
07:58
because we live there, and we get a feeling about it
189
463000
3000
08:01
that basically matches reality.
190
466000
3000
08:04
Security theater's exposed
191
469000
3000
08:07
when it's obvious that it's not working properly.
192
472000
3000
08:10
Okay, so what makes people not notice?
193
475000
4000
08:14
Well, a poor understanding.
194
479000
2000
08:16
If you don't understand the risks, you don't understand the costs,
195
481000
3000
08:19
you're likely to get the trade-off wrong,
196
484000
2000
08:21
and your feeling doesn't match reality.
197
486000
3000
08:24
Not enough examples.
198
489000
2000
08:26
There's an inherent problem
199
491000
2000
08:28
with low probability events.
200
493000
2000
08:30
If, for example,
201
495000
2000
08:32
terrorism almost never happens,
202
497000
2000
08:34
it's really hard to judge
203
499000
2000
08:36
the efficacy of counter-terrorist measures.
204
501000
3000
08:40
This is why you keep sacrificing virgins,
205
505000
3000
08:43
and why your unicorn defenses are working just great.
206
508000
3000
08:46
There aren't enough examples of failures.
207
511000
3000
08:50
Also, feelings that are clouding the issues --
208
515000
3000
08:53
the cognitive biases I talked about earlier,
209
518000
2000
08:55
fears, folk beliefs,
210
520000
3000
08:58
basically an inadequate model of reality.
211
523000
3000
09:02
So let me complicate things.
212
527000
3000
09:05
I have feeling and reality.
213
530000
2000
09:07
I want to add a third element. I want to add model.
214
532000
3000
09:10
Feeling and model in our head,
215
535000
2000
09:12
reality is the outside world.
216
537000
2000
09:14
It doesn't change; it's real.
217
539000
3000
09:17
So feeling is based on our intuition.
218
542000
2000
09:19
Model is based on reason.
219
544000
2000
09:21
That's basically the difference.
220
546000
3000
09:24
In a primitive and simple world,
221
549000
2000
09:26
there's really no reason for a model
222
551000
3000
09:29
because feeling is close to reality.
223
554000
3000
09:32
You don't need a model.
224
557000
2000
09:34
But in a modern and complex world,
225
559000
2000
09:36
you need models
226
561000
2000
09:38
to understand a lot of the risks we face.
227
563000
3000
09:42
There's no feeling about germs.
228
567000
2000
09:44
You need a model to understand them.
229
569000
3000
09:47
So this model
230
572000
2000
09:49
is an intelligent representation of reality.
231
574000
3000
09:52
It's, of course, limited by science,
232
577000
3000
09:55
by technology.
233
580000
2000
09:57
We couldn't have a germ theory of disease
234
582000
3000
10:00
before we invented the microscope to see them.
235
585000
3000
10:04
It's limited by our cognitive biases.
236
589000
3000
10:07
But it has the ability
237
592000
2000
10:09
to override our feelings.
238
594000
2000
10:11
Where do we get these models? We get them from others.
239
596000
3000
10:14
We get them from religion, from culture,
240
599000
3000
10:17
teachers, elders.
241
602000
2000
10:19
A couple years ago,
242
604000
2000
10:21
I was in South Africa on safari.
243
606000
2000
10:23
The tracker I was with grew up in Kruger National Park.
244
608000
3000
10:26
He had some very complex models of how to survive.
245
611000
3000
10:29
And it depended on if you were attacked
246
614000
2000
10:31
by a lion or a leopard or a rhino or an elephant --
247
616000
2000
10:33
and when you had to run away, and when you couldn't run away, and when you had to climb a tree --
248
618000
3000
10:36
when you could never climb a tree.
249
621000
2000
10:38
I would have died in a day,
250
623000
3000
10:41
but he was born there,
251
626000
2000
10:43
and he understood how to survive.
252
628000
2000
10:45
I was born in New York City.
253
630000
2000
10:47
I could have taken him to New York, and he would have died in a day.
254
632000
3000
10:50
(Laughter)
255
635000
2000
10:52
Because we had different models
256
637000
2000
10:54
based on our different experiences.
257
639000
3000
10:58
Models can come from the media,
258
643000
2000
11:00
from our elected officials.
259
645000
3000
11:03
Think of models of terrorism,
260
648000
3000
11:06
child kidnapping,
261
651000
3000
11:09
airline safety, car safety.
262
654000
2000
11:11
Models can come from industry.
263
656000
3000
11:14
The two I'm following are surveillance cameras,
264
659000
2000
11:16
ID cards,
265
661000
2000
11:18
quite a lot of our computer security models come from there.
266
663000
3000
11:21
A lot of models come from science.
267
666000
3000
11:24
Health models are a great example.
268
669000
2000
11:26
Think of cancer, of bird flu, swine flu, SARS.
269
671000
3000
11:29
All of our feelings of security
270
674000
3000
11:32
about those diseases
271
677000
2000
11:34
come from models
272
679000
2000
11:36
given to us, really, by science filtered through the media.
273
681000
3000
11:40
So models can change.
274
685000
3000
11:43
Models are not static.
275
688000
2000
11:45
As we become more comfortable in our environments,
276
690000
3000
11:48
our model can move closer to our feelings.
277
693000
4000
11:53
So an example might be,
278
698000
2000
11:55
if you go back 100 years ago
279
700000
2000
11:57
when electricity was first becoming common,
280
702000
3000
12:00
there were a lot of fears about it.
281
705000
2000
12:02
I mean, there were people who were afraid to push doorbells,
282
707000
2000
12:04
because there was electricity in there, and that was dangerous.
283
709000
3000
12:07
For us, we're very facile around electricity.
284
712000
3000
12:10
We change light bulbs
285
715000
2000
12:12
without even thinking about it.
286
717000
2000
12:14
Our model of security around electricity
287
719000
4000
12:18
is something we were born into.
288
723000
3000
12:21
It hasn't changed as we were growing up.
289
726000
3000
12:24
And we're good at it.
290
729000
3000
12:27
Or think of the risks
291
732000
2000
12:29
on the Internet across generations --
292
734000
2000
12:31
how your parents approach Internet security,
293
736000
2000
12:33
versus how you do,
294
738000
2000
12:35
versus how our kids will.
295
740000
3000
12:38
Models eventually fade into the background.
296
743000
3000
12:42
Intuitive is just another word for familiar.
297
747000
3000
12:45
So as your model is close to reality,
298
750000
2000
12:47
and it converges with feelings,
299
752000
2000
12:49
you often don't know it's there.
300
754000
3000
12:52
So a nice example of this
301
757000
2000
12:54
came from last year and swine flu.
302
759000
3000
12:57
When swine flu first appeared,
303
762000
2000
12:59
the initial news caused a lot of overreaction.
304
764000
4000
13:03
Now it had a name,
305
768000
2000
13:05
which made it scarier than the regular flu,
306
770000
2000
13:07
even though it was more deadly.
307
772000
2000
13:09
And people thought doctors should be able to deal with it.
308
774000
4000
13:13
So there was that feeling of lack of control.
309
778000
2000
13:15
And those two things
310
780000
2000
13:17
made the risk more than it was.
311
782000
2000
13:19
As the novelty wore off, the months went by,
312
784000
3000
13:22
there was some amount of tolerance,
313
787000
2000
13:24
people got used to it.
314
789000
2000
13:26
There was no new data, but there was less fear.
315
791000
3000
13:29
By autumn,
316
794000
2000
13:31
people thought
317
796000
2000
13:33
the doctors should have solved this already.
318
798000
2000
13:35
And there's kind of a bifurcation --
319
800000
2000
13:37
people had to choose
320
802000
2000
13:39
between fear and acceptance --
321
804000
4000
13:43
actually fear and indifference --
322
808000
2000
13:45
they kind of chose suspicion.
323
810000
3000
13:48
And when the vaccine appeared last winter,
324
813000
3000
13:51
there were a lot of people -- a surprising number --
325
816000
3000
13:54
who refused to get it --
326
819000
3000
13:58
as a nice example
327
823000
2000
14:00
of how people's feelings of security change, how their model changes,
328
825000
3000
14:03
sort of wildly
329
828000
2000
14:05
with no new information,
330
830000
2000
14:07
with no new input.
331
832000
2000
14:09
This kind of thing happens a lot.
332
834000
3000
14:12
I'm going to give one more complication.
333
837000
3000
14:15
We have feeling, model, reality.
334
840000
3000
14:18
I have a very relativistic view of security.
335
843000
2000
14:20
I think it depends on the observer.
336
845000
3000
14:23
And most security decisions
337
848000
2000
14:25
have a variety of people involved.
338
850000
4000
14:29
And stakeholders
339
854000
2000
14:31
with specific trade-offs
340
856000
3000
14:34
will try to influence the decision.
341
859000
2000
14:36
And I call that their agenda.
342
861000
2000
14:38
And you see agenda --
343
863000
2000
14:40
this is marketing, this is politics --
344
865000
3000
14:43
trying to convince you to have one model versus another,
345
868000
3000
14:46
trying to convince you to ignore a model
346
871000
2000
14:48
and trust your feelings,
347
873000
3000
14:51
marginalizing people with models you don't like.
348
876000
3000
14:54
This is not uncommon.
349
879000
3000
14:57
An example, a great example, is the risk of smoking.
350
882000
3000
15:01
In the history of the past 50 years, the smoking risk
351
886000
3000
15:04
shows how a model changes,
352
889000
2000
15:06
and it also shows how an industry fights against
353
891000
3000
15:09
a model it doesn't like.
354
894000
2000
15:11
Compare that to the secondhand smoke debate --
355
896000
3000
15:14
probably about 20 years behind.
356
899000
3000
15:17
Think about seat belts.
357
902000
2000
15:19
When I was a kid, no one wore a seat belt.
358
904000
2000
15:21
Nowadays, no kid will let you drive
359
906000
2000
15:23
if you're not wearing a seat belt.
360
908000
2000
15:26
Compare that to the airbag debate --
361
911000
2000
15:28
probably about 30 years behind.
362
913000
3000
15:31
All examples of models changing.
363
916000
3000
15:36
What we learn is that changing models is hard.
364
921000
3000
15:39
Models are hard to dislodge.
365
924000
2000
15:41
If they equal your feelings,
366
926000
2000
15:43
you don't even know you have a model.
367
928000
3000
15:46
And there's another cognitive bias
368
931000
2000
15:48
I'll call confirmation bias,
369
933000
2000
15:50
where we tend to accept data
370
935000
3000
15:53
that confirms our beliefs
371
938000
2000
15:55
and reject data that contradicts our beliefs.
372
940000
3000
15:59
So evidence against our model,
373
944000
2000
16:01
we're likely to ignore, even if it's compelling.
374
946000
3000
16:04
It has to get very compelling before we'll pay attention.
375
949000
3000
16:08
New models that extend long periods of time are hard.
376
953000
2000
16:10
Global warming is a great example.
377
955000
2000
16:12
We're terrible
378
957000
2000
16:14
at models that span 80 years.
379
959000
2000
16:16
We can do to the next harvest.
380
961000
2000
16:18
We can often do until our kids grow up.
381
963000
3000
16:21
But 80 years, we're just not good at.
382
966000
3000
16:24
So it's a very hard model to accept.
383
969000
3000
16:27
We can have both models in our head simultaneously,
384
972000
4000
16:31
right, that kind of problem
385
976000
3000
16:34
where we're holding both beliefs together,
386
979000
3000
16:37
right, the cognitive dissonance.
387
982000
2000
16:39
Eventually,
388
984000
2000
16:41
the new model will replace the old model.
389
986000
3000
16:44
Strong feelings can create a model.
390
989000
3000
16:47
September 11th created a security model
391
992000
3000
16:50
in a lot of people's heads.
392
995000
2000
16:52
Also, personal experiences with crime can do it,
393
997000
3000
16:55
personal health scare,
394
1000000
2000
16:57
a health scare in the news.
395
1002000
2000
16:59
You'll see these called flashbulb events
396
1004000
2000
17:01
by psychiatrists.
397
1006000
2000
17:03
They can create a model instantaneously,
398
1008000
3000
17:06
because they're very emotive.
399
1011000
3000
17:09
So in the technological world,
400
1014000
2000
17:11
we don't have experience
401
1016000
2000
17:13
to judge models.
402
1018000
2000
17:15
And we rely on others. We rely on proxies.
403
1020000
2000
17:17
I mean, this works as long as it's to correct others.
404
1022000
4000
17:21
We rely on government agencies
405
1026000
2000
17:23
to tell us what pharmaceuticals are safe.
406
1028000
5000
17:28
I flew here yesterday.
407
1033000
2000
17:30
I didn't check the airplane.
408
1035000
2000
17:32
I relied on some other group
409
1037000
2000
17:34
to determine whether my plane was safe to fly.
410
1039000
3000
17:37
We're here, none of us fear the roof is going to collapse on us,
411
1042000
3000
17:40
not because we checked,
412
1045000
3000
17:43
but because we're pretty sure
413
1048000
2000
17:45
the building codes here are good.
414
1050000
3000
17:48
It's a model we just accept
415
1053000
2000
17:50
pretty much by faith.
416
1055000
2000
17:52
And that's okay.
417
1057000
3000
17:57
Now, what we want
418
1062000
2000
17:59
is people to get familiar enough
419
1064000
2000
18:01
with better models --
420
1066000
2000
18:03
have it reflected in their feelings --
421
1068000
2000
18:05
to allow them to make security trade-offs.
422
1070000
4000
18:09
Now when these go out of whack,
423
1074000
2000
18:11
you have two options.
424
1076000
2000
18:13
One, you can fix people's feelings,
425
1078000
2000
18:15
directly appeal to feelings.
426
1080000
2000
18:17
It's manipulation, but it can work.
427
1082000
3000
18:20
The second, more honest way
428
1085000
2000
18:22
is to actually fix the model.
429
1087000
3000
18:26
Change happens slowly.
430
1091000
2000
18:28
The smoking debate took 40 years,
431
1093000
3000
18:31
and that was an easy one.
432
1096000
3000
18:34
Some of this stuff is hard.
433
1099000
2000
18:36
I mean really though,
434
1101000
2000
18:38
information seems like our best hope.
435
1103000
2000
18:40
And I lied.
436
1105000
2000
18:42
Remember I said feeling, model, reality;
437
1107000
2000
18:44
I said reality doesn't change. It actually does.
438
1109000
3000
18:47
We live in a technological world;
439
1112000
2000
18:49
reality changes all the time.
440
1114000
3000
18:52
So we might have -- for the first time in our species --
441
1117000
3000
18:55
feeling chases model, model chases reality, reality's moving --
442
1120000
3000
18:58
they might never catch up.
443
1123000
3000
19:02
We don't know.
444
1127000
2000
19:04
But in the long-term,
445
1129000
2000
19:06
both feeling and reality are important.
446
1131000
3000
19:09
And I want to close with two quick stories to illustrate this.
447
1134000
3000
19:12
1982 -- I don't know if people will remember this --
448
1137000
2000
19:14
there was a short epidemic
449
1139000
3000
19:17
of Tylenol poisonings in the United States.
450
1142000
2000
19:19
It's a horrific story. Someone took a bottle of Tylenol,
451
1144000
3000
19:22
put poison in it, closed it up, put it back on the shelf.
452
1147000
3000
19:25
Someone else bought it and died.
453
1150000
2000
19:27
This terrified people.
454
1152000
2000
19:29
There were a couple of copycat attacks.
455
1154000
2000
19:31
There wasn't any real risk, but people were scared.
456
1156000
3000
19:34
And this is how
457
1159000
2000
19:36
the tamper-proof drug industry was invented.
458
1161000
2000
19:38
Those tamper-proof caps, that came from this.
459
1163000
2000
19:40
It's complete security theater.
460
1165000
2000
19:42
As a homework assignment, think of 10 ways to get around it.
461
1167000
2000
19:44
I'll give you one, a syringe.
462
1169000
3000
19:47
But it made people feel better.
463
1172000
3000
19:50
It made their feeling of security
464
1175000
2000
19:52
more match the reality.
465
1177000
2000
19:54
Last story, a few years ago, a friend of mine gave birth.
466
1179000
3000
19:57
I visit her in the hospital.
467
1182000
2000
19:59
It turns out when a baby's born now,
468
1184000
2000
20:01
they put an RFID bracelet on the baby,
469
1186000
2000
20:03
put a corresponding one on the mother,
470
1188000
2000
20:05
so if anyone other than the mother takes the baby out of the maternity ward,
471
1190000
2000
20:07
an alarm goes off.
472
1192000
2000
20:09
I said, "Well, that's kind of neat.
473
1194000
2000
20:11
I wonder how rampant baby snatching is
474
1196000
2000
20:13
out of hospitals."
475
1198000
2000
20:15
I go home, I look it up.
476
1200000
2000
20:17
It basically never happens.
477
1202000
2000
20:19
But if you think about it,
478
1204000
2000
20:21
if you are a hospital,
479
1206000
2000
20:23
and you need to take a baby away from its mother,
480
1208000
2000
20:25
out of the room to run some tests,
481
1210000
2000
20:27
you better have some good security theater,
482
1212000
2000
20:29
or she's going to rip your arm off.
483
1214000
2000
20:31
(Laughter)
484
1216000
2000
20:33
So it's important for us,
485
1218000
2000
20:35
those of us who design security,
486
1220000
2000
20:37
who look at security policy,
487
1222000
3000
20:40
or even look at public policy
488
1225000
2000
20:42
in ways that affect security.
489
1227000
2000
20:44
It's not just reality; it's feeling and reality.
490
1229000
3000
20:47
What's important
491
1232000
2000
20:49
is that they be about the same.
492
1234000
2000
20:51
It's important that, if our feelings match reality,
493
1236000
2000
20:53
we make better security trade-offs.
494
1238000
2000
20:55
Thank you.
495
1240000
2000
20:57
(Applause)
496
1242000
2000

▲Back to top

ABOUT THE SPEAKER
Bruce Schneier - Security expert
Bruce Schneier thinks hard about security -- as a computer security guru, and as a philosopher of the larger notion of making a safer world.

Why you should listen

Bruce Schneier is an internationally renowned security technologist and author. Described by the Economist as a "security guru," he is best known as a refreshingly candid and lucid security critic and commentator. When people want to know how security really works, they turn to Schneier.

His first bestseller, Applied Cryptography, explained how the arcane science of secret codes actually works, and was described by Wired as "the book the National Security Agency wanted never to be published." His book on computer and network security, Secrets and Lies, was called by Fortune "[a] jewel box of little surprises you can actually use." Beyond Fear tackles the problems of security from the small to the large: personal safety, crime, corporate security, national security. His current book, Schneier on Security, offers insight into everything from the risk of identity theft (vastly overrated) to the long-range security threat of unchecked presidential power and the surprisingly simple way to tamper-proof elections.

Schneier publishes a free monthly newsletter, Crypto-Gram, with over 150,000 readers. In its ten years of regular publication, Crypto-Gram has become one of the most widely read forums for free-wheeling discussions, pointed critiques and serious debate about security. As head curmudgeon at the table, Schneier explains, debunks and draws lessons from security stories that make the news.

More profile about the speaker
Bruce Schneier | Speaker | TED.com