ABOUT THE SPEAKER
Tricia Wang - Technology ethnographer
With astronaut eyes and ethnographer curiosity, Tricia Wang helps corporations grow by discovering the unknown about their customers.

Why you should listen

For Tricia Wang, human behavior generates some of the most perplexing questions of our times. She has taught global organizations how to identify new customers and markets hidden behind their data, amplified IDEO's design thinking practice as an expert-in-residence, researched the social evolution of the Chinese internet, and written about the "elastic self," an emergent form of interaction in a virtual world. Wang is the co-founder of Sudden Compass, a consulting firm that helps companies unlock new growth opportunities by putting customer obsession into practice.

Wang's work has been featured in The Atlantic, Al Jazeera, and The Guardian. Fast Company spotlighted her work in China: "What Twitter Can Learn From Weibo: Field Notes From Global Tech Ethnographer Tricia Wang." In her latest op-ed on Slate, she discusses how attempts to stop terrorists on social media can harm our privacy and anonymity. Her Medium post, "Why Big Data Needs Thick Data," is a frequently cited industry piece on the importance of an integrated data approach. One of her favorite essays documents her day in the life of working as a street vendor in China.

Known for her lively presentations that are grounded in her research and observations about human behavior and data, Wang has spoken at organizations such as Proctor & Gamble, Nike, Wrigley, 21st Century Fox and Tumblr. Her most recent talk at Enterprise UX delved into why corporate innovation usually doesn’t work and what to do about it. She delivered the opening keynote at The Conference to a crowd of marketers and creatives, delving into the wild history of linear perspective and its influence on how we think and form organizations.

Wang holds affiliate positions at Data & Society, Harvard University's Berkman Klein Center for Internet Studies and New York University's Interactive Telecommunication Program. She oversees Ethnography Matters, a site that publishes articles about applied ethnography and technology. She co-started a Slack community for anyone who uses ethnographic methods in industry.

Wang began her career as a documentary filmmaker at NASA, an HIV/AIDS activist, and an educator specializing in culturally responsive pedagogy. She is also proud to have co-founded the first national hip-hop education initiative, which turned into the Hip Hop Education Center at New York University, and to have built after-school technology and arts programs for low-income youth at New York City public schools and the Queens Museum of Arts. Her life philosophy is that you have to go to the edge to discover what’s really happening. She's the proud companion of her internet famous dog, #ellethedog.

More profile about the speaker
Tricia Wang | Speaker | TED.com
TEDxCambridge

Tricia Wang: The human insights missing from big data

Filmed:
1,688,539 views

Why do so many companies make bad decisions, even with access to unprecedented amounts of data? With stories from Nokia to Netflix to the oracles of ancient Greece, Tricia Wang demystifies big data and identifies its pitfalls, suggesting that we focus instead on "thick data" -- precious, unquantifiable insights from actual people -- to make the right business decisions and thrive in the unknown.
- Technology ethnographer
With astronaut eyes and ethnographer curiosity, Tricia Wang helps corporations grow by discovering the unknown about their customers. Full bio

Double-click the English transcript below to play the video.

00:12
In ancient Greece,
0
885
1545
00:15
when anyone from slaves to soldiers,
poets and politicians,
1
3436
3943
00:19
needed to make a big decision
on life's most important questions,
2
7403
4004
00:23
like, "Should I get married?"
3
11431
1391
00:24
or "Should we embark on this voyage?"
4
12846
1857
00:26
or "Should our army
advance into this territory?"
5
14727
2928
00:29
they all consulted the oracle.
6
17679
2579
00:33
So this is how it worked:
7
21020
1440
00:34
you would bring her a question
and you would get on your knees,
8
22484
3112
00:37
and then she would go into this trance.
9
25620
1871
00:39
It would take a couple of days,
10
27515
1549
00:41
and then eventually
she would come out of it,
11
29088
2163
00:43
giving you her predictions as your answer.
12
31275
2536
00:46
From the oracle bones of ancient China
13
34910
2566
00:49
to ancient Greece to Mayan calendars,
14
37500
2345
00:51
people have craved for prophecy
15
39869
2296
00:54
in order to find out
what's going to happen next.
16
42189
3137
00:58
And that's because we all want
to make the right decision.
17
46516
3239
01:01
We don't want to miss something.
18
49779
1545
01:03
The future is scary,
19
51892
1743
01:05
so it's much nicer
knowing that we can make a decision
20
53659
2717
01:08
with some assurance of the outcome.
21
56400
1982
01:11
Well, we have a new oracle,
22
59079
1611
01:12
and it's name is big data,
23
60714
2145
01:14
or we call it "Watson"
or "deep learning" or "neural net."
24
62883
3939
01:19
And these are the kinds of questions
we ask of our oracle now,
25
67340
4012
01:23
like, "What's the most efficient way
to ship these phones
26
71376
3922
01:27
from China to Sweden?"
27
75322
1823
01:29
Or, "What are the odds
28
77169
1800
01:30
of my child being born
with a genetic disorder?"
29
78993
3363
01:34
Or, "What are the sales volume
we can predict for this product?"
30
82952
3244
01:40
I have a dog. Her name is Elle,
and she hates the rain.
31
88108
4047
01:44
And I have tried everything
to untrain her.
32
92179
3306
01:47
But because I have failed at this,
33
95509
2771
01:50
I also have to consult
an oracle, called Dark Sky,
34
98304
3286
01:53
every time before we go on a walk,
35
101614
1635
01:55
for very accurate weather predictions
in the next 10 minutes.
36
103273
3577
02:01
She's so sweet.
37
109535
1303
02:03
So because of all of this,
our oracle is a $122 billion industry.
38
111827
5707
02:10
Now, despite the size of this industry,
39
118006
3376
02:13
the returns are surprisingly low.
40
121406
2456
02:16
Investing in big data is easy,
41
124342
2494
02:18
but using it is hard.
42
126860
1933
02:21
Over 73 percent of big data projects
aren't even profitable,
43
129981
4040
02:26
and I have executives
coming up to me saying,
44
134045
2431
02:28
"We're experiencing the same thing.
45
136500
1789
02:30
We invested in some big data system,
46
138313
1753
02:32
and our employees aren't making
better decisions.
47
140090
2968
02:35
And they're certainly not coming up
with more breakthrough ideas."
48
143082
3162
02:38
So this is all really interesting to me,
49
146914
3184
02:42
because I'm a technology ethnographer.
50
150122
2010
02:44
I study and I advise companies
51
152630
2564
02:47
on the patterns
of how people use technology,
52
155218
2483
02:49
and one of my interest areas is data.
53
157725
2678
02:52
So why is having more data
not helping us make better decisions,
54
160427
5193
02:57
especially for companies
who have all these resources
55
165644
2783
03:00
to invest in these big data systems?
56
168451
1736
03:02
Why isn't it getting any easier for them?
57
170211
2398
03:05
So, I've witnessed the struggle firsthand.
58
173990
2634
03:09
In 2009, I started
a research position with Nokia.
59
177374
3484
03:13
And at the time,
60
181232
1158
03:14
Nokia was one of the largest
cell phone companies in the world,
61
182414
3158
03:17
dominating emerging markets
like China, Mexico and India --
62
185596
3202
03:20
all places where I had done
a lot of research
63
188822
2502
03:23
on how low-income people use technology.
64
191348
2676
03:26
And I spent a lot of extra time in China
65
194048
2330
03:28
getting to know the informal economy.
66
196402
2592
03:31
So I did things like working
as a street vendor
67
199018
2401
03:33
selling dumplings to construction workers.
68
201443
2574
03:36
Or I did fieldwork,
69
204041
1358
03:37
spending nights and days
in internet cafés,
70
205423
2958
03:40
hanging out with Chinese youth,
so I could understand
71
208405
2546
03:42
how they were using
games and mobile phones
72
210975
2284
03:45
and using it between moving
from the rural areas to the cities.
73
213283
3370
03:50
Through all of this qualitative evidence
that I was gathering,
74
218335
3927
03:54
I was starting to see so clearly
75
222286
2824
03:57
that a big change was about to happen
among low-income Chinese people.
76
225134
4472
04:03
Even though they were surrounded
by advertisements for luxury products
77
231020
4367
04:07
like fancy toilets --
who wouldn't want one? --
78
235411
3495
04:10
and apartments and cars,
79
238930
2890
04:13
through my conversations with them,
80
241844
1820
04:15
I found out that the ads
the actually enticed them the most
81
243688
3841
04:19
were the ones for iPhones,
82
247553
1996
04:21
promising them this entry
into this high-tech life.
83
249573
3052
04:25
And even when I was living with them
in urban slums like this one,
84
253469
3163
04:28
I saw people investing
over half of their monthly income
85
256656
2996
04:31
into buying a phone,
86
259676
1623
04:33
and increasingly, they were "shanzhai,"
87
261323
2302
04:35
which are affordable knock-offs
of iPhones and other brands.
88
263649
3388
04:40
They're very usable.
89
268303
1625
04:42
Does the job.
90
270890
1322
04:44
And after years of living
with migrants and working with them
91
272750
5789
04:50
and just really doing everything
that they were doing,
92
278563
3434
04:54
I started piecing
all these data points together --
93
282021
3597
04:57
from the things that seem random,
like me selling dumplings,
94
285642
3123
05:00
to the things that were more obvious,
95
288789
1804
05:02
like tracking how much they were spending
on their cell phone bills.
96
290617
3232
05:05
And I was able to create
this much more holistic picture
97
293873
2639
05:08
of what was happening.
98
296536
1156
05:09
And that's when I started to realize
99
297716
1722
05:11
that even the poorest in China
would want a smartphone,
100
299462
3509
05:14
and that they would do almost anything
to get their hands on one.
101
302995
4985
05:21
You have to keep in mind,
102
309073
2404
05:23
iPhones had just come out, it was 2009,
103
311501
3084
05:26
so this was, like, eight years ago,
104
314609
1885
05:28
and Androids had just started
looking like iPhones.
105
316518
2437
05:30
And a lot of very smart
and realistic people said,
106
318979
2507
05:33
"Those smartphones -- that's just a fad.
107
321510
2207
05:36
Who wants to carry around
these heavy things
108
324243
2996
05:39
where batteries drain quickly
and they break every time you drop them?"
109
327263
3487
05:44
But I had a lot of data,
110
332793
1201
05:46
and I was very confident
about my insights,
111
334018
2260
05:48
so I was very excited
to share them with Nokia.
112
336302
2829
05:53
But Nokia was not convinced,
113
341332
2517
05:55
because it wasn't big data.
114
343873
2335
05:59
They said, "We have
millions of data points,
115
347022
2404
06:01
and we don't see any indicators
of anyone wanting to buy a smartphone,
116
349450
4247
06:05
and your data set of 100,
as diverse as it is, is too weak
117
353721
4388
06:10
for us to even take seriously."
118
358133
1714
06:12
And I said, "Nokia, you're right.
119
360908
1605
06:14
Of course you wouldn't see this,
120
362537
1560
06:16
because you're sending out surveys
assuming that people don't know
121
364121
3371
06:19
what a smartphone is,
122
367516
1159
06:20
so of course you're not going
to get any data back
123
368699
2366
06:23
about people wanting to buy
a smartphone in two years.
124
371089
2572
06:25
Your surveys, your methods
have been designed
125
373685
2118
06:27
to optimize an existing business model,
126
375827
2022
06:29
and I'm looking
at these emergent human dynamics
127
377873
2608
06:32
that haven't happened yet.
128
380505
1354
06:33
We're looking outside of market dynamics
129
381883
2438
06:36
so that we can get ahead of it."
130
384345
1631
06:39
Well, you know what happened to Nokia?
131
387373
2244
06:41
Their business fell off a cliff.
132
389641
2365
06:44
This -- this is the cost
of missing something.
133
392791
3727
06:49
It was unfathomable.
134
397163
1999
06:52
But Nokia's not alone.
135
400003
1651
06:54
I see organizations
throwing out data all the time
136
402258
2581
06:56
because it didn't come from a quant model
137
404863
2561
06:59
or it doesn't fit in one.
138
407448
1768
07:02
But it's not big data's fault.
139
410219
2048
07:04
It's the way we use big data;
it's our responsibility.
140
412942
3907
07:09
Big data's reputation for success
141
417730
1911
07:11
comes from quantifying
very specific environments,
142
419665
3759
07:15
like electricity power grids
or delivery logistics or genetic code,
143
423448
4913
07:20
when we're quantifying in systems
that are more or less contained.
144
428385
4318
07:24
But not all systems
are as neatly contained.
145
432727
2969
07:27
When you're quantifying
and systems are more dynamic,
146
435720
3258
07:31
especially systems
that involve human beings,
147
439002
3799
07:34
forces are complex and unpredictable,
148
442825
2426
07:37
and these are things
that we don't know how to model so well.
149
445275
3486
07:41
Once you predict something
about human behavior,
150
449204
2813
07:44
new factors emerge,
151
452041
1855
07:45
because conditions
are constantly changing.
152
453920
2365
07:48
That's why it's a never-ending cycle.
153
456309
1803
07:50
You think you know something,
154
458136
1464
07:51
and then something unknown
enters the picture.
155
459624
2242
07:53
And that's why just relying
on big data alone
156
461890
3322
07:57
increases the chance
that we'll miss something,
157
465236
2849
08:00
while giving us this illusion
that we already know everything.
158
468109
3777
08:04
And what makes it really hard
to see this paradox
159
472406
3856
08:08
and even wrap our brains around it
160
476286
2659
08:10
is that we have this thing
that I call the quantification bias,
161
478969
3691
08:14
which is the unconscious belief
of valuing the measurable
162
482684
3922
08:18
over the immeasurable.
163
486630
1594
08:21
And we often experience this at our work.
164
489222
3284
08:24
Maybe we work alongside
colleagues who are like this,
165
492530
2650
08:27
or even our whole entire
company may be like this,
166
495204
2428
08:29
where people become
so fixated on that number,
167
497656
2546
08:32
that they can't see anything
outside of it,
168
500226
2067
08:34
even when you present them evidence
right in front of their face.
169
502317
3948
08:39
And this is a very appealing message,
170
507123
3371
08:42
because there's nothing
wrong with quantifying;
171
510518
2343
08:44
it's actually very satisfying.
172
512885
1430
08:46
I get a great sense of comfort
from looking at an Excel spreadsheet,
173
514339
4362
08:50
even very simple ones.
174
518725
1401
08:52
(Laughter)
175
520150
1014
08:53
It's just kind of like,
176
521188
1152
08:54
"Yes! The formula worked. It's all OK.
Everything is under control."
177
522364
3504
08:58
But the problem is
178
526792
2390
09:01
that quantifying is addictive.
179
529206
2661
09:03
And when we forget that
180
531891
1382
09:05
and when we don't have something
to kind of keep that in check,
181
533297
3038
09:08
it's very easy to just throw out data
182
536359
2118
09:10
because it can't be expressed
as a numerical value.
183
538501
2718
09:13
It's very easy just to slip
into silver-bullet thinking,
184
541243
2921
09:16
as if some simple solution existed.
185
544188
2579
09:19
Because this is a great moment of danger
for any organization,
186
547600
4062
09:23
because oftentimes,
the future we need to predict --
187
551686
2634
09:26
it isn't in that haystack,
188
554344
2166
09:28
but it's that tornado
that's bearing down on us
189
556534
2538
09:31
outside of the barn.
190
559096
1488
09:34
There is no greater risk
191
562960
2326
09:37
than being blind to the unknown.
192
565310
1666
09:39
It can cause you to make
the wrong decisions.
193
567000
2149
09:41
It can cause you to miss something big.
194
569173
1974
09:43
But we don't have to go down this path.
195
571734
3101
09:47
It turns out that the oracle
of ancient Greece
196
575453
3195
09:50
holds the secret key
that shows us the path forward.
197
578672
3966
09:55
Now, recent geological research has shown
198
583654
2595
09:58
that the Temple of Apollo,
where the most famous oracle sat,
199
586273
3564
10:01
was actually built
over two earthquake faults.
200
589861
3084
10:04
And these faults would release
these petrochemical fumes
201
592969
2886
10:07
from underneath the Earth's crust,
202
595879
1685
10:09
and the oracle literally sat
right above these faults,
203
597588
3866
10:13
inhaling enormous amounts
of ethylene gas, these fissures.
204
601478
3588
10:17
(Laughter)
205
605090
1008
10:18
It's true.
206
606122
1173
10:19
(Laughter)
207
607319
1017
10:20
It's all true, and that's what made her
babble and hallucinate
208
608360
3509
10:23
and go into this trance-like state.
209
611893
1724
10:25
She was high as a kite!
210
613641
1770
10:27
(Laughter)
211
615435
4461
10:31
So how did anyone --
212
619920
2779
10:34
How did anyone get
any useful advice out of her
213
622723
3030
10:37
in this state?
214
625777
1190
10:39
Well, you see those people
surrounding the oracle?
215
627497
2381
10:41
You see those people holding her up,
216
629902
1879
10:43
because she's, like, a little woozy?
217
631805
1717
10:45
And you see that guy
on your left-hand side
218
633546
2308
10:47
holding the orange notebook?
219
635878
1598
10:50
Well, those were the temple guides,
220
638105
1730
10:51
and they worked hand in hand
with the oracle.
221
639859
3016
10:56
When inquisitors would come
and get on their knees,
222
644084
2516
10:58
that's when the temple guides
would get to work,
223
646624
2340
11:00
because after they asked her questions,
224
648988
1864
11:02
they would observe their emotional state,
225
650876
2001
11:04
and then they would ask them
follow-up questions,
226
652901
2324
11:07
like, "Why do you want to know
this prophecy? Who are you?
227
655249
2834
11:10
What are you going to do
with this information?"
228
658107
2264
11:12
And then the temple guides would take
this more ethnographic,
229
660395
3182
11:15
this more qualitative information,
230
663601
2156
11:17
and interpret the oracle's babblings.
231
665781
2075
11:21
So the oracle didn't stand alone,
232
669428
2292
11:23
and neither should our big data systems.
233
671744
2148
11:26
Now to be clear,
234
674630
1161
11:27
I'm not saying that big data systems
are huffing ethylene gas,
235
675815
3459
11:31
or that they're even giving
invalid predictions.
236
679298
2353
11:33
The total opposite.
237
681675
1161
11:34
But what I am saying
238
682860
2068
11:36
is that in the same way
that the oracle needed her temple guides,
239
684952
3832
11:40
our big data systems need them, too.
240
688808
2288
11:43
They need people like ethnographers
and user researchers
241
691120
4109
11:47
who can gather what I call thick data.
242
695253
2506
11:50
This is precious data from humans,
243
698502
2991
11:53
like stories, emotions and interactions
that cannot be quantified.
244
701517
4102
11:57
It's the kind of data
that I collected for Nokia
245
705643
2322
11:59
that comes in in the form
of a very small sample size,
246
707989
2669
12:02
but delivers incredible depth of meaning.
247
710682
2955
12:05
And what makes it so thick and meaty
248
713661
3680
12:10
is the experience of understanding
the human narrative.
249
718445
4029
12:14
And that's what helps to see
what's missing in our models.
250
722498
3639
12:18
Thick data grounds our business questions
in human questions,
251
726851
4045
12:22
and that's why integrating
big and thick data
252
730920
3562
12:26
forms a more complete picture.
253
734506
1689
12:28
Big data is able to offer
insights at scale
254
736772
2881
12:31
and leverage the best
of machine intelligence,
255
739677
2647
12:34
whereas thick data can help us
rescue the context loss
256
742348
3572
12:37
that comes from making big data usable,
257
745944
2098
12:40
and leverage the best
of human intelligence.
258
748066
2181
12:42
And when you actually integrate the two,
that's when things get really fun,
259
750271
3552
12:45
because then you're no longer
just working with data
260
753847
2436
12:48
you've already collected.
261
756307
1196
12:49
You get to also work with data
that hasn't been collected.
262
757527
2737
12:52
You get to ask questions about why:
263
760288
1719
12:54
Why is this happening?
264
762031
1317
12:55
Now, when Netflix did this,
265
763778
1379
12:57
they unlocked a whole new way
to transform their business.
266
765181
3035
13:01
Netflix is known for their really great
recommendation algorithm,
267
769406
3956
13:05
and they had this $1 million prize
for anyone who could improve it.
268
773386
4797
13:10
And there were winners.
269
778207
1314
13:12
But Netflix discovered
the improvements were only incremental.
270
780255
4323
13:17
So to really find out what was going on,
271
785404
1964
13:19
they hired an ethnographer,
Grant McCracken,
272
787392
3741
13:23
to gather thick data insights.
273
791157
1546
13:24
And what he discovered was something
that they hadn't seen initially
274
792727
3924
13:28
in the quantitative data.
275
796675
1355
13:31
He discovered that people loved
to binge-watch.
276
799072
2728
13:33
In fact, people didn't even
feel guilty about it.
277
801824
2353
13:36
They enjoyed it.
278
804201
1255
13:37
(Laughter)
279
805480
1026
13:38
So Netflix was like,
"Oh. This is a new insight."
280
806530
2356
13:40
So they went to their data science team,
281
808910
1938
13:42
and they were able to scale
this big data insight
282
810872
2318
13:45
in with their quantitative data.
283
813214
2587
13:47
And once they verified it
and validated it,
284
815825
3170
13:51
Netflix decided to do something
very simple but impactful.
285
819019
4761
13:56
They said, instead of offering
the same show from different genres
286
824834
6492
14:03
or more of the different shows
from similar users,
287
831350
3888
14:07
we'll just offer more of the same show.
288
835262
2554
14:09
We'll make it easier
for you to binge-watch.
289
837840
2105
14:11
And they didn't stop there.
290
839969
1486
14:13
They did all these things
291
841479
1474
14:14
to redesign their entire
viewer experience,
292
842977
2959
14:17
to really encourage binge-watching.
293
845960
1758
14:20
It's why people and friends disappear
for whole weekends at a time,
294
848230
3241
14:23
catching up on shows
like "Master of None."
295
851495
2343
14:25
By integrating big data and thick data,
they not only improved their business,
296
853862
4173
14:30
but they transformed how we consume media.
297
858059
2812
14:32
And now their stocks are projected
to double in the next few years.
298
860895
4552
14:38
But this isn't just about
watching more videos
299
866280
3830
14:42
or selling more smartphones.
300
870134
1620
14:44
For some, integrating thick data
insights into the algorithm
301
872143
4050
14:48
could mean life or death,
302
876217
2263
14:50
especially for the marginalized.
303
878504
2146
14:53
All around the country,
police departments are using big data
304
881738
3434
14:57
for predictive policing,
305
885196
1963
14:59
to set bond amounts
and sentencing recommendations
306
887183
3084
15:02
in ways that reinforce existing biases.
307
890291
3147
15:06
NSA's Skynet machine learning algorithm
308
894296
2423
15:08
has possibly aided in the deaths
of thousands of civilians in Pakistan
309
896743
5444
15:14
from misreading cellular device metadata.
310
902211
2721
15:19
As all of our lives become more automated,
311
907131
3403
15:22
from automobiles to health insurance
or to employment,
312
910558
3080
15:25
it is likely that all of us
313
913662
2350
15:28
will be impacted
by the quantification bias.
314
916036
2989
15:32
Now, the good news
is that we've come a long way
315
920972
2621
15:35
from huffing ethylene gas
to make predictions.
316
923617
2450
15:38
We have better tools,
so let's just use them better.
317
926091
3070
15:41
Let's integrate the big data
with the thick data.
318
929185
2323
15:43
Let's bring our temple guides
with the oracles,
319
931532
2261
15:45
and whether this work happens
in companies or nonprofits
320
933817
3376
15:49
or government or even in the software,
321
937217
2469
15:51
all of it matters,
322
939710
1792
15:53
because that means
we're collectively committed
323
941526
3023
15:56
to making better data,
324
944573
2191
15:58
better algorithms, better outputs
325
946788
1836
16:00
and better decisions.
326
948648
1643
16:02
This is how we'll avoid
missing that something.
327
950315
3558
16:07
(Applause)
328
955222
3948

▲Back to top

ABOUT THE SPEAKER
Tricia Wang - Technology ethnographer
With astronaut eyes and ethnographer curiosity, Tricia Wang helps corporations grow by discovering the unknown about their customers.

Why you should listen

For Tricia Wang, human behavior generates some of the most perplexing questions of our times. She has taught global organizations how to identify new customers and markets hidden behind their data, amplified IDEO's design thinking practice as an expert-in-residence, researched the social evolution of the Chinese internet, and written about the "elastic self," an emergent form of interaction in a virtual world. Wang is the co-founder of Sudden Compass, a consulting firm that helps companies unlock new growth opportunities by putting customer obsession into practice.

Wang's work has been featured in The Atlantic, Al Jazeera, and The Guardian. Fast Company spotlighted her work in China: "What Twitter Can Learn From Weibo: Field Notes From Global Tech Ethnographer Tricia Wang." In her latest op-ed on Slate, she discusses how attempts to stop terrorists on social media can harm our privacy and anonymity. Her Medium post, "Why Big Data Needs Thick Data," is a frequently cited industry piece on the importance of an integrated data approach. One of her favorite essays documents her day in the life of working as a street vendor in China.

Known for her lively presentations that are grounded in her research and observations about human behavior and data, Wang has spoken at organizations such as Proctor & Gamble, Nike, Wrigley, 21st Century Fox and Tumblr. Her most recent talk at Enterprise UX delved into why corporate innovation usually doesn’t work and what to do about it. She delivered the opening keynote at The Conference to a crowd of marketers and creatives, delving into the wild history of linear perspective and its influence on how we think and form organizations.

Wang holds affiliate positions at Data & Society, Harvard University's Berkman Klein Center for Internet Studies and New York University's Interactive Telecommunication Program. She oversees Ethnography Matters, a site that publishes articles about applied ethnography and technology. She co-started a Slack community for anyone who uses ethnographic methods in industry.

Wang began her career as a documentary filmmaker at NASA, an HIV/AIDS activist, and an educator specializing in culturally responsive pedagogy. She is also proud to have co-founded the first national hip-hop education initiative, which turned into the Hip Hop Education Center at New York University, and to have built after-school technology and arts programs for low-income youth at New York City public schools and the Queens Museum of Arts. Her life philosophy is that you have to go to the edge to discover what’s really happening. She's the proud companion of her internet famous dog, #ellethedog.

More profile about the speaker
Tricia Wang | Speaker | TED.com