-
Notifications
You must be signed in to change notification settings - Fork 0
/
TASM2019_slides.Rmd
863 lines (503 loc) · 22.5 KB
/
TASM2019_slides.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
---
title: "\nPersonalisation Algorithms &<br>Extremist Content Online"
subtitle: 'Dr Alastair Reed <a href="https://twitter.com/reed_alastair">@reed\\_alastair</a><br>Joe Whittaker <a href="https://twitter.com/CTProject_JW">@CTProject\\_JW</a><br>Fabio Votta <a href="https://twitter.com/favstats">@favstats</a><br>Seán Looney <a href="https://twitter.com/_Sean_Looney_">@\\_Sean\\_Looney\_</a>'
date: "<br>26th June 2019<br>Slides available here:<br>[xrw-and-algorithms.netlify.com](https://xrw-and-algorithms.netlify.com/)"
output:
xaringan::moon_reader:
lib_dir: libs
css: xaringan-themer.css
nature:
titleSlideClass: [top, center]
highlightStyle: github
highlightLines: true
countIncrementalSlides: false
slideNumberFormat: "%current%<br>"
---
class: center, middle, inverse
## What are Filter bubbles?
### And what's the problem?
<style>
.onehundredtwenty {
font-size: 120%;
}
<style>
.ninety {
font-size: 90%;
}
.eightyseven {
font-size: 87%;
}
.eightyfive {
font-size: 85%;
}
.eighty {
font-size: 80%;
}
.seventyfive {
font-size: 75%;
}
.seventy {
font-size: 70%;
}
.twentyfive {
font-size: 25%;
}
</style>
---
### What are Filter bubbles?
<br>
![](images/filter upset.png)
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
## Research so far
### and linking filter bubbles to extremism
---
### Research so far
The empirical evidence of a "filter bubble" effect is less clear and decidedly less pessimistic.
+ Study on Facebook suggests filter bubbles are generated less by algorithms than by individual user decisions (Bakshy, Messing and Adamic, 2015).
+ Research analysing Google news recommendation suggests essential information is not omitted (Haim, Graefe and Brosius, 2018)
+ Personalised recommendations show no reduction in diversity over human editors (Möller et al., 2018).
+ Research on Google search results also finds factors such as time of search were more explanatory than prior behaviour and preferences (Courtois, Slechten and Coenen, 2018).
Why the discrepancy?
+ "Echo chamber about echo chambers" (Guess et al. 2018)
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Filter Bubbles and Extremism
There is a paucity of research studying the effects of personalisation algorithms on extremist content.
.pull-left[
![](images/recommend_friends.png)
]
.eightyseven[
<br>
+ YouTube's recommended videos can propel users into an immersive bubble of right-wing extremism (O’Callaghan et al., 2015)
+ Twitter’s “Who to Follow” suggested violent extremist Islamist groups if the user followed al-Qaeda affiliated group (Berger, 2013)
+ Facebook’s “Recommended Friends” function had likely actively connected at least two Islamic State supporters in SE Asia (Waters and Postings, 2018)
]
<br>
<br>
<br>
The architecture of the platforms may facilitate closer interactions than would otherwise exist.
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
## Research Question & Design
---
### Research Question
> Do algorithms promote extremist material once a user begins to interact with such content?
```{r, echo = F}
knitr::include_graphics("images/social_media_logos.png")
```
**How to measure extremist content?**
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
--
+ Hand-coding content with Holbrook's Extremist Media Index
Holbrook's Extremism Media Index (2015)
```{r, echo = F}
knitr::include_graphics("images/holbrooke_index.png")
```
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Research Design & Data Collection
.eightyseven[
+ YouTube/Reddit Research design
+ Created **THREE** identical accounts
+ All follow same 10 XRW channels/subreddtis; 10 Neutral.<sup>1</sup>
+ Each account interacts with different kinds of content
+ Collecting timelines *two times a day* for *two weeks* (28 sessions in total)
+ 2019-01-21 and ended on 2019-02-04
Each account does nothing for a week and after Session 14 we apply different treatments:
1. *Neutral Interaction Account* mostly interacts with neutral content
2. *Extreme Interaction Account* mostly interacts with extreme content
3. *Baseline Account* does nothing to establish a baseline
+ On YouTube: Pull 18 Recommended Videos of each account
+ On Reddit: 25 threads on the *“Best”* personalized timeline for each account
+ Every piece of content gets a unique rank per session and an Extremist Media Index (EMI) Score (Holbrook, 2015)
]
<br>
.seventy[
[1] We follow the policy of Vox Pol and Berger (2018) in not identifying the names of accounts in this research, both for reasons of potentially increasing exposure and privacy.
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Research Design & Methods
**Expected Relationships**
+ Frequency of extreme material increases after interacting with extreme content
+ Extreme content is prioritized by the algorithm when interacting with extreme material
**Methods**
+ Count data modeled with (quasi-)poisson regression
+ Non-Parametric t-tests to estimate ranking differences
+ Satisfactory interrater reliability between two individuals
+ Krippendorff's alpha: 0.77 across all platforms
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
## Results
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
.pull-left[
![](images/emi_yt_props.png)
]
.pull-right[
Overview of the data:
Of the 1443 videos coded on YouTube (749 unique)
+ 65.77% moderate
+ 28.34% fringe
+ 5.89% extremist
]
.seventyfive[
Figure on the left shows the EMI scores for each session with a rank from one to eighteen, depending on where the video appears on the “Recommended Videos” section. Figure on the right shows the percentage distribution of the three categories of content before and after each treatment.
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F, out.height=400, out.width=600, fig.align='center'}
knitr::include_graphics("images/irr_plot.png")
```
.seventyfive[
After **extreme interaction** treatment:
+ The incident rate for *fringe content* is 1.72 times the incident rate for the reference group
+ The incident rate for *extreme content* is 2.47 times the incident rate for the reference group
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F, out.height=400, out.width=600, fig.align='center'}
knitr::include_graphics("images/frq_plot.png")
```
.seventyfive[
After **extreme interaction** treatment:
+ Fringe content 1.37 (p < 0.01) times more likely than before
+ Extreme content 2.00 (p < 0.01) times more likely than before
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F}
knitr::include_graphics("images/pct_extreme.png")
```
.seventyfive[
+ In the **neutral interaction account**
+ **only three sessions** had an extreme content piece after interaction
+ In the **extreme interaction account**
+ **all but one session** had an extreme content piece after interaction
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F, out.height=400, out.width=800, fig.align='center'}
knitr::include_graphics("images/eia_rank_extreme.png")
```
.seventyfive[
+ In the **extreme interaction account**
+ Extreme content ranked sig. higher (p = 0.028) than moderate
+ Almost all extreme content shows up in the upper half (< 8) of the recommendations
(Median Rank = 5)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/reddit-logo.jpg")
```
.pull-left[
![](images/emi_reddit_props.png)
]
.pull-right[
Of the 2100 posts coded on Reddit (834 unique)
+ 78.76% Moderate
+ 19.81% Fringe
+ 1.43% Extreme
No sig. observable in-/decrease of content
No sig. difference in ranking of content
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/Gab-Logo.jpg")
```
.pull-left[
![](images/gab_overview.png)
]
.pull-right[
*Different setup due to technical difficulties*
3 Different News Feeds: “Popular”, “Controversial”, and “Latest”
3 Topics: “Politics”, “News”, “Humour”
Collected data over five sessions
1271 Rated posts (746 unique)
+ 63.73% Moderate
+ 28.8% Fringe
+ 7.47% Extreme
]
No statistically sig. difference between extreme & non-extreme content in any of the timelines
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
## Discussion
---
### Discussion
> Do algorithms promote extremist material once a user begins to interact with such content?
.pull-left[
+ Only YouTube has an effect which, after engaging with extreme content, prioritises it further.
+ Users which engage with extreme and fringe content are more likely to be recommended more of the same.
+ Extreme content is pushed up the ranking of recommended videos on YouTube
+ Support for O’Callaghan et al. (2015)
]
.pull-right[
![](images/yt_screen.png)
<br>
![](images/down_the_white.png)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Discussion
.pull-left[
+ Safe haven for right wing extremists and home to terrorists such as Pittsburgh Synagogue Shooter
+ Anecdotally, by far the most extreme of the platforms.
+ Lack of evidence of Gab’s algorithm suggests that it is the user’s choices which are responsible for this environment.
]
.pull-right[
![](images/Gab-Logo.jpg)
<br>
![](images/robert_bowers.png)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
## Recommendations
---
### Recommendations
**Removing Problematic Content from Recommendations**
.pull-left[
+ Content which does not clearly violate site rules or policies
+ Google’s “limited features” policy
+ Reddit’s “Quarantine” System
+ Opt in content
+ No monetisation or recommendation
+ Constructive balance between freedom of speech and harmful content.
+ Need for Clarity and Consistency
]
.pull-right[
![](images/reddit_offensive.png)
<a href="https://oilab.eu/4chans-youtube-a-fringe-perspective-on-youtubes-great-purge-of-2019/">
<img src="https://oilab.eu/wp-content/uploads/2019/06/nsg_05-06-2019-gif.gif">
</a>
.seventyfive[
[OILAB](https://oilab.eu/4chans-youtube-a-fringe-perspective-on-youtubes-great-purge-of-2019/): Availability of YouTube videos posted in Nazi threads on 4chan/pol/ on 5 and 6 June 2019.
]
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Recommendations
**Ensuring Video Recommendations are from Quality Sources**
.pull-left[
+ Provide users with more context and alternative perspectives
+ Google introduced changes to make quality count and give more context to searches.
+ Provide high quality sources on the same topic
+ Jigsaw’s redirect method as a model.
]
.pull-right[
![](images/redirect.png)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Recommendations
**Greater Transparency**
.pull-left[
+ Users should have a clear option to request why content has been recommended to them.
+ Opportunity for Explainable AI
+ Facebook’s “Why am I seeing this ad/post”
]
.pull-right[
![](images/xai.png)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Greater Transparency
![](images/why.png)
.seventyfive[
[Source: Facebook Newsroom, 31st March 2019](https://newsroom.fb.com/news/2019/03/why-am-i-seeing-this/)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
### Future Research
.pull-left[
+ More accounts over a longer period of time
+ Research Project constrained by the number and type of social media platforms that we could research.
+ Closed nature of the platform (Facebook)
+ Terms of Service Restrictions (Twitter)
+ Increasing knowledge gap which can only be answered through close collaboration with social media companies.
]
.pull-right[
![](images/fb_twitter.png)
]
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
### Thank you for listening!
---
## Literature
Bakshy, E., Messing, S., & Adamic, L. A. (2015). Exposure to ideologically diverse news and opinion on Facebook. Science, 348(6239), 1130-1132.
Berger, J. M. (2013) ‘Zero Degrees of al Qaeda’, Foreign Policy, (August 14). Available at: http://foreignpolicy.com/2013/08/14/zero-degrees-of-al-qaeda/.
Courtois, C., Slechten, L., & Coenen, L. (2018). Challenging Google Search filter bubbles in social and political information: Disconforming evidence from a digital methods case study. Telematics and Informatics, 35(7), 2006-2015.
Guess, A., Lyons, B., Nyhan, B., & Reifler, J. (2018). Avoiding the echo chamber about echo chambers: Why selective exposure to like-minded political news is less prevalent than you think. Document of the Knight Foundation. Retrieved from: https://www.researchgate.net/publication/330144926_Avoiding_the_echo_chamber_about_echo_chambers_Why_selective_exposure_to_like-minded_political_news_is_less_prevalent_than_you_think
Haim, M., Graefe, A., & Brosius, H. B. (2018). Burst of the filter bubble? Effects of personalization on the diversity of Google News. Digital Journalism, 6(3), 330-343.
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
## Literature
Holbrook, D. (2015). Designing and Applying an ‘Extremist Media Index’. Perspectives On Terrorism, 9(5). Retrieved from http://www.terrorismanalysts.com/pt/index.php/pot/article/view/461
Möller, J., Trilling, D., Helberger, N., & van Es, B. (2018). Do not blame it on the algorithm: an empirical assessment of multiple recommender systems and their impact on content diversity. Information, Communication & Society, 21(7), 959-977.
O’Callaghan, Derek, et al. "Down the (white) rabbit hole: The extreme right and online recommender systems." Social Science Computer Review 33.4 (2015): 459-478.
Vīķe-Freiberga, V., Däubler-Gmelin, H., Hammersley, B., Pessoa Maduro, L.M.P. (2013). A free and pluralistic media to sustain European democracy. Retrieved from http://ec.europa.eu/digital-agenda/sites/digital-agenda/files/HLG%20Final%20Report.pdf
Waters, G. and Postings, R. (2018) Spiders of the Caliphate: Mapping the Islamic State’s Global Support Network on Facebook, Counter-Extremism Project.
<div class="my-footer"><span>Personalisation Algorithms & Extremist Content Online
             
             
</span></div>
---
class: center, middle, inverse
### Appendix
---
### Facebook adds "Why am I seeing this" to posts
![](images/why.png)
[Source: Facebook Newsroom, 31st March 2019](https://newsroom.fb.com/news/2019/03/why-am-i-seeing-this/)
---
### What are Personalisation Algorithms?
.pull-left[
![](images/pariser.jpg)
<br>Pariser 2011
]
+ Algorithms are responsible for content that users see in their feeds
+ Eli Pariser suggests that they can create a “filter bubble” effect or “autopropaganda”
+ by controlling what users do and do not see it can – and is in fact, designed to – dramatically amplify confirmation bias
+ The pre-filtering of content leads to bubbles in which people never view or read about opposing viewpoints
+ Creates seperated spaces that make communication between opposing viewpoints harder: undermines democracy itself
---
### What are Personalisation Algorithms?
<br>
The EU Group on Media Freedom and Pluralism notes that:
> Increasing filtering mechanisms makes it more likely for people to only get news on subjects they are interested in, and with the perspective they identify with. It will also tend to create more insulated communities as isolated subsets within the overall public sphere. [...] Such developments undoubtedly have a potentially negative impact on democracy.
*Vīķe-Freiberga, Däubler-Gmelin, Hammersley, & Pessoa Maduro, 2013, p. 27*
<br>
**Filter bubbles are considered a concern at the highest level.**
---
### How to measure extremist content?
Multiple pathways:
+ Sentiment analysis can be used to identify extremist authors (Scrivens et al., 2018)
+ Topic models to identify (far-right) extremist content (O’Callaghan et al., 2015)
+ Hand-coding content for example with Holbrook's Extremist Media Index (Holbrook, 2015)
+ Manually labelled data can also be used as training dataset for machine learning models
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F}
knitr::include_graphics("images/yt_firstrank.png")
```
.seventyfive[
Figure shows the first appearance of a piece of content (Moderate, Fringe or Extreme) for each session.
]
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
```{r, echo = F, warning = F, message=F}
knitr::include_graphics("images/yt_firstrank_extreme.png")
```
.seventyfive[
Figure shows the first appearance of an **extreme** content piece for each session.
+ In the *neutral interaction account*
+ **only three sessions** had an extreme content piece after interaction
+ In the *extreme interaction account*
+ **all but one session** had an extreme content piece after interaction
+ Almost all content shows up in the upper half (< 8) of the recommendation list
(Median Rank = 5)
]
---
```{r, out.width=150, echo = F}
knitr::include_graphics("images/youtube.png")
```
Poisson Regression Results:
```{r, echo = F, warning = F, message=F}
knitr::include_graphics("images/coeff_poisson3.png")
```
---
### Reddit Personalisation
```{r, echo = F, warning = F, message=F}
knitr::include_graphics("images/reddit_personalisation.png")
```