Matrix Factorization

hontolab 3,695 views 34 slides Nov 11, 2019
Slide 1
Slide 1 of 34
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20
Slide 21
21
Slide 22
22
Slide 23
23
Slide 24
24
Slide 25
25
Slide 26
26
Slide 27
27
Slide 28
28
Slide 29
29
Slide 30
30
Slide 31
31
Slide 32
32
Slide 33
33
Slide 34
34

About This Presentation

This is a material for the course lecture "Data Engineering" at Shizuoka University in 2019.


Slide Content

Matrix Factorization:
Beyond Simple Collaborative Filtering
YusukeYamamoto
Lecturer,FacultyofInformatics
[email protected]
Data Engineering (Recommender Systems 3)
2019.11.11

1
2
Problems on
Simple Collaborative Filtering

User-based Collaborative Filtering
3
Predicts a target user’s rating for an item
based on rating tendency of similar users
!"#$%&'(),%=",-+∑,∈123%4((),()7(",,8−",-)
∑,∈123%4((),()
Item5simAverage Rating
Alice?14
User130.852.4
User250.713.8
Similar
users

Idea about Item-based Collaborative Filtering
4
Item1Item2Item3Item4Item5
Alice5344?
User131233
User243435
User333154
User415521
similar
Predicts unknown scores
based on rating tendency for similar items
similar
!"#$%&'(),%:=∑8∈;23%4(%:,%)7",-,8
∑8∈;23%4(%:,%)

Problems on CF approaches (1/3)
5Image reference: https://rafalab.github.io/dsbook/recommendation-systems.html
Real data is quite sparse!!
Even on large e-commerce sites, there are few intersections
between user vectors (& item vectors).

Problems on CF approaches (2/3)
6
The curse of dimensionality
•In high-dimensional space, it’s difficult to handle similarity
•Usually, item/user vectors have quite high dimensionality
(b/c rating matrix is quite large)

Problems on CF approaches (3/3)
7
High computational cost
•The rating matrix is directly used every time systems find
similar user/items and make predictions
•CF approaches do not scale for most real world scenarios
User 1
3 2 4 … 3
5
2 4 … 1

Rating matrix
Similar users
for user 1
3 2 4 … 3
5
2 4 … 1

Rating matrix
Suggested
items
for user 1
Compute Compute
User 2
3 2 4 … 3
5
2 4 … 1

Rating matrix
Similar users
for user 2
3 2 4 … 3
5
2 4 … 1

Rating matrix
Suggested
items
for user 2
Compute Compute

Recent approaches for recommender systems
8
Model-based approach
-Based on offline pre-processing
-At run-time, only the learned model is used for rating prediction
-Learned models can be updated
3 2 4 … 3
5
2 4 … 1

Rating matrix
Learned
Model
Computation
offline
User
Suggested
Items for user
Computation
online

Memory-based approach vs. model-based approach
9
Memory-based
approach
-User-based CF
-Item-based CF
Model-based
approach
-Matrix factorization
-Association rule mining
-Probabilistic model
-O

2
10
Matrix Factorization

UserGod
father
Termin-
ator
Money
gameTitanicBack to
the future…X-men
Alice51443…2
Basic idea 1 (1/2)
11
I don’t like horror…
Sci-Fisoften move me.
I love humaneand
dramatic movies!
UserHorrorSci-FiHumanityDrama…
Alice0.32.16.14.7…
Latent factors (which cannot be observed)
Assumes that latent factors exist in users/items

Basic idea 1 (2/2)
12
Assumes that latent factors exist in users/items
Image from Amazon.com
Latent factors (which cannot be observed)
UserGod
father
Termin-
ator
Money
gameTitanicBack to
the future…X-men
Alice51443…2
UserHorrorSci-FiHumanityDrama…
God
father2.30.44.95.7…

Basic idea 2
13
UserGod
father
Termin-
ator
Money
gameTitanicBack to
the future…X-men
Alice51443…2
UserHorrorSci-FiHumanityDrama…
Alice0.32.16.14.7…
Assumes that rating scores derive from
latent factors of users and items
UserHorrorSci-FiHumanityDrama…
God
father2.30.44.95.7…
×

Summary of matrix factorization
14
3 … 3
5
2 … 1

Rating matrix = R
(m users ×n items)

P
1.3 … 0.7
0.2 … 4.9

×
0.2 1.1 … 0.9
3.1 … 0.7 1

Latent user matrix
(m users ×k latent factors)
Latent item matrix
(k latent factors x n items)
QR≈

•Rating matrix can be decomposed to latent factors
of users and items
•The dimension of latent factors (vectors) is much less
than the number of users and items (k ≪ m, n)



(□= unknown scores)
User
Item
Rating

Prediction using matrix factorization
15
3 … 3
5
2 … 1

Predicted
rating matrix = R*
=
P
1.3 … 0.7
0.2 … 4.9

×
0.2 1.1 … 0.9
3.1 … 0.7 1

Latent user matrix
Latent item matrix
QR=

If we obtain latent user/item matrix, we can predict
unknown scores by multiplying the two latent matrix
How to obtain latent user/item matrix?

2
5
4
□□□

SVD for recommender systems
16
SVD: singular value decomposition
-A famous linear algebra technique for matrix decomposition
-It is often used for dimensionality reduction
-SVD delivers essentially the same result as PCA does
UΣX=
T×V×
m x n matrix
m x m
unitary matrix
n x n
unitary matrix
Rectangular diagonal matrix
(diagonal values are called as singular values)

Example of SVD
17
UΣX=
T×V×
1 3 3 3 0
2 4 2 2 4
1 3 3 5 1
4 5 2 3 3
1 1 5 2 1
-0.369 -0.325 0.282 0.343 0.749
-0.459 0.448 -0.339 -0.584 0.364
-0.468 -0.359 0.544 -0.459 -0.381
-0.563 0.491 0.07 0.562 -0.348
-0.341 -0.569 -0.71 0.118 -0.202
13.368 00 0 0
0 4.708 0 0 0
0 0 2.792 0 0
0 0 0 1.586 0
0 0 0 0 0.904
-0.3250.341 -0.1010.682 -0.55
-0.5620.3450.2730.1530.684
-0.468 -0.642 -0.5770.1250.141
-0.504 -0.3270.601 -0.324 -0.416
-0.3240.496 -0.47 -0.626 -0.189
U=
Σ=
V=

Important features of SVD (1/4)
18
We can approximate a given matrix by focusing
on the largest singular values and the vectors of U
and V which correspond to the singular values
-0.369 -0.325 0.282 0.343 0.749
-0.459 0.448 -0.339 -0.584 0.364
-0.468 -0.359 0.544 -0.459 -0.381
-0.563 0.491 0.07 0.562 -0.348
-0.341 -0.569 -0.71 0.118 -0.202
13.368 00 0 0
0 4.708 0 0 0
0 0 2.792 0 0
0 0 0 1.586 0
0 0 0 0 0.904
-0.3250.341 -0.1010.682 -0.55
-0.5620.3450.2730.1530.684
-0.468 -0.642 -0.5770.1250.141
-0.504 -0.3270.601 -0.324 -0.416
-0.3240.496 -0.47 -0.626 -0.189
U=
Σ=
V=
Largest singular values

Important features of SVD (2/4)
19
-0.369 -0.325 0.282 0.343 0.749
-0.459 0.448 -0.339 -0.584 0.364
-0.468 -0.359 0.544 -0.459 -0.381
-0.563 0.491 0.07 0.562 -0.348
-0.341 -0.569 -0.71 0.118 -0.202
13.368 00 0 0
0 4.708 0 0 0
0 0 2.792 0 0
0 0 0 1.586 0
0 0 0 0 0.904
-0.3250.341 -0.1010.682 -0.55
-0.5620.3450.2730.1530.684
-0.468 -0.642 -0.5770.1250.141
-0.504 -0.3270.601 -0.324 -0.416
-0.3240.496 -0.47 -0.626 -0.189
U=
Σ=
V=
Ignore unimportant values!
We can approximate a given matrix by focusing
on the largest singular values and the vectors of U
and V which correspond to the singular values

13.368 00 0 0
0 4.708 0 0 0
0 0 2.792 0 0
0 0 0 1.586 0
0 0 0 0 0.90
Important features of SVD (3/4)
20
Σ2=
-0.369 -0.325 0.282 0.343 0.749
-0.459 0.448 -0.339 -0.584 0.364
-0.468 -0.359 0.544 -0.459 -0.381
-0.563 0.491 0.07 0.562 -0.348
-0.341 -0.569 -0.71 0.118 -0.202
U2=
-0.3250.341 -0.1010.682 -0.55
-0.5620.3450.2730.1530.684
-0.468 -0.642 -0.5770.1250.141
-0.504 -0.3270.601 -0.324 -0.416
-0.3240.496 -0.47 -0.626 -0.189V2=
We can approximate a given matrix by focusing
on the largest singular values and the vectors of U
and V which correspond to the singular values

Important features of SVD (4/4)
21
U2Σ2
T
×V2×
=
1.08 2.24 3.29 2.98 0.84
2.72 4.17 1.52 2.41 3.04
1.46 2.93 4.02 3.71 1.19
3.24 5.03 2.04 3.04 3.59
0.57 1.64 3.86 3.18 0.15
1 3 3 3 0
2 4 2 2 4
1 3 3 5 1
4 5 2 3 3
1 1 5 2 1
≈= X
We can approximate a given matrix by focusing
on the largest singular values and the vectors of U
and V which correspond to the singular values

Apply SVD for recommender systems (1/2)
22
Item1Item2Item3Item4Item5
Alice1333?
User124224
User213351
User345233
User411521
Regarded as zero
1 3 3 3 0
2 4 2 2 4
1 3 3 5 1
4 5 2 3 3
1 1 5 2 1
= X
To matrix

Apply SVD for recommender systems (2/2)
23
U2Σ2
TV2
UΣX=
T
×V×
Focus on
important features
××
Step 1
Step 2
Run SVD
Multiply three matrixStep 3
=
1.08 2.24 3.29 2.98 0.84
2.72 4.17 1.52 2.41 3.04
1.46 2.93 4.02 3.71 1.19
3.24 5.03 2.04 3.04 3.59
0.57 1.64 3.86 3.18 0.15
Check the values
which were zero
before running SVD
Step 4

Problems on SVD
24
Predicted values are often negative
-SVD does not take rating score range into account
Zero replacement decreases prediction quality
-SVD analyzes the relation between all data in matrix
-The meaning of “zero” is different from that of “unknown”

Bad example of SVD-based recommendation
25
Ghibli 1Ghibli 2Ghibli 3Ghibli 4
User15
User234
User321
User454
User55
5 000
3400
2010
0504
0005
Example from: http://smrmkt.hatenablog.jp/entry/2014/08/23/211555
U2, Σ2, V2
Run SVD
Zero replacement
Multiply matrix
3.531.880.16-0.26
3.624.040.132.17
1.440.760.07-0.12
2.765.410.064.34
1.675.9705.74
-0.262.91-0.063.52

Netflix Prize (2006-2010)
26Image ref: http://blogs.itmedia.co.jp/saito/2009/09/httpjournalmyco.html
Netflix held the open competition to advance collaborative
filtering algorithms and to seek the best algorithm.

Simon Funk’s Matrix Factorization (2006)
27
Without using SVD (with zero replacement),
the Simon’s method learns matrix P and Q
to compute observed values only in R
4%<=,>?
,,8∈@
",,8−A,B8CD+E(A,D+B8D)
Target optimization function
m users
n items
Rating
matrix
R
≈×User
Item
PQ
u: user u; i: item i;
pu: u’s latent vector; qi: i’s latent vector

Various approaches have been developed …
28Ref: https://www.slideshare.net/databricks/deep-learning-for-recommender-systems-with-nick-pentreath
20032006-200920102013
Scalable models
Amazon’s item-based CF
Netflix Prize
The rise of matrix factorization
like Simon Funk’s method
Factorization machine
Generalized matrix factorization
for dealing with various factors
Deep Learning
・Deep Factorization machine
・Content2Vec to get content
embeddings

3
29
Challenges for
recommender systems

Remaining challenges
30
Cold start problem
How to recommend new items? What to recommend to new users?
Serendipity
-Only recommendation of favorable items cannot expand user interests
-How to recommend unexpected and surprising items?
Providing explanations
-Recommendation algorithms just decide what to recommend
-How can systems persuade users to purchase recommended items?
Reviewer’s trust
How to remove spam reviewers? How to find high-quality reviewers?

Remaining challenges
31
Cold start problem
How to recommend new items? What to recommend to new users?
Serendipity
-Only recommendation of favorable items cannot expand user interests
-How to recommend unexpected and surprising items?
Providing explanations
-Recommendation algorithms just decide what to recommend
-How can systems persuade users to purchase recommended items?
Reviewer’s trust
How to remove spam reviewers? How to find high-quality reviewers?

Cold start problem
32
Item1Item2Item3Item4item5
Kate
User13123
User24343
User33315
User41552
New user
New item
CF approaches don’t work for new items/users
New items/users have no clues to predict unknown scores b/c
the CF cannot find neighbor users/items

Possible approach to cold start problem
33
On-the-fly preference prediction
Systems ask/force users to rate a set of items to understand
their preference
Content-based recommendation
Finds similar items based on user interactions and content features
ExtVision: Augmentation of Visual Experiences with
Generation of Context Images for Peripheral Vision Using
Deep Neural Network

ABSTRACT
We propose a system, called ExtVision, to augment visual
experiences by generating and projecting context-images
onto the periphery of the television or computer screen. A
peripheral projection of the context-image is one of the most
effective techniques to enhance visual experiences. However,
the projection is not commonly used at present, because of
the difficulty in preparing the context-image. In this paper,
we propose a deep neural network-based method to generate
context-images for peripheral projection. A user study was
performed to investigate the manner in which the proposed
system augments traditional visual experiences. In addition,
we present applications and future prospects of the
developed system.
Author Keywords
Spatially augmented reality; immersion; augmented video;
large field of view; human vision; video extrapolation
ACM Classification Keywords
H.5.1. Multimedia Information Systems: Artificial, aug-
mented, and virtual realities
INTRODUCTION
Presentation of images to a viewer’s peripheral vision makes
the experience of watching a video more immersive and
exciting. In this paper, we describe how images are displayed
as context-images in peripheral areas to enhance visual
experiences.Illumiroom [10] projects context-images or
visual effects around the TV screen and enhances the
traditional gaming experience. ScreenX [20] and
BarcoEscape [21] provide sub-screens to display video on
both sides of the main screen, providing a more immersive
movie experience.
However, peripheral presentation systems are not very
popular at present, because it is difficult to prepare the
context- images. Researchers have attempted the boundary
value shooting method and computer vision processing to
address this, but a self-sufficient method still remains to be
developed.
In recent years, deep neural network (DNN)-based methods
have shown considerable success in image completion. The
images generated by DNN-based methods are very plausible
but somewhat inaccurate. Human peripheral vision is lower
in resolution than the central visual field and is therefore
weak in distinguishing visual accuracy; therefore, DNN-
based methods are suitable for generating context-images
that meet the requirements of peripheral vision.
In this study, we investigated context-image generation using
DNN. The contributions of this study are as follows:
1. Implementation of two new DNN-based methods to
generate context-images
Naoki Kimura
University of Tokyo
Tokyo, Japan
[email protected]
Jun Rekimoto
University of Tokyo / Sony CSL
Tokyo, Japan
[email protected]

Figure 1. ExtVision augments visual experiences. These images are examples of our application. The context-images are
generated and projected onto the peripheral area around the TV by our system.

Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. Copyrights for
components of this work owned by others than ACM must be honored.
Abstracting with credit is permitted. To copy otherwise, or republish, to
post on servers or to redistribute to lists, requires prior specific permission
and/or a fee. Request permissions from [email protected].

CHI 2018, April 21–26, 2018, Montreal, QC, Canada
© 2018 Association for Computing Machinery.
ACM ISBN 978-1-4503-5620-6/18/04…$15.00
https://doi.org/10.1145/3173574.3174001

CHI 2018 Honourable MentionCHI 2018, April 21–26, 2018, Montréal, QC, Canada
Paper 427 Page 1
Disinformation on the Web: Impact, Characteristics, and
Detection of Wikipedia Hoaxes
Srijan Kumar
!
University of Maryland
[email protected]
Robert West
Stanford University
[email protected]
Jure Leskovec
Stanford University
[email protected]
ABSTRACT
Wikipedia is a major source of information for many people. How-
ever, false information on Wikipedia raises concerns about its cred-
ibility. One way in which false information may be presented on
Wikipedia is in the form of hoax articles,i.e., articles containing
fabricated facts about nonexistent entities or events. In this paper
we study false information on Wikipedia by focusing on the hoax
articles that have been created throughout its history. We make
several contributions. First, we assess the real-world impact of
hoax articles by measuring how long they survive before being de-
bunked, how many pageviews they receive, and how heavily they
are referred to by documents on the Web. We find that, while most
hoaxes are detected quickly and have little impact on Wikipedia,
a small number of hoaxes survive long and are well cited across
the Web. Second, we characterize the nature of successful hoaxes
by comparing them to legitimate articles and to failed hoaxes that
were discovered shortly after being created. We find characteristic
differences in terms of article structure and content, embeddedness
into the rest of Wikipedia, and features of the editor who created
the hoax. Third, we successfully apply our findings to address a
series of classification tasks, most notably to determine whether a
given article is a hoax. And finally, we describe and evaluate a task
involving humans distinguishing hoaxes from non-hoaxes. We find
that humans are not good at solving this task and that our automated
classifier outperforms them by a big margin.
1. INTRODUCTION
The Web is a space for all, where, in principle, everybody can
read, and everybody can publish and share, information. Thus,
knowledge can be transmitted at a speed and breadth unprecedented
in human history, which has had tremendous positive effects on the
lives of billions of people. But there is also a dark side to the un-
reigned proliferation of information over the Web: it has become a
breeding ground for false information [6, 7, 12, 15, 19, 43].
The reasons for communicating false information vary widely:
on the one extreme,misinformationis conveyed in the honest but
!
Research done partly during a visit at Stanford University.
Copyright is held by the International World Wide Web Conference Com-
mittee (IW3C2). IW3C2 reserves the right to provide a hyperlink to the
author’s site if the Material is used in electronic media.
WWW 2016,April 11–15, 2016, Montréal, Québec, Canada.
ACM 978-1-4503-4143-1/16/04.
http://dx.doi.org/10.1145/2872427.2883085.
mistaken belief that the relayed incorrect facts are true; on the other
extreme,disinformationdenotes false facts that are conceived in
order to deliberately deceive or betray an audience [11, 17]. A third
class of false information has been calledbullshit,where the agent’s
primary purpose is not to mislead an audience into believing false
facts, but rather to “convey a certain impression of himself” [14].
All these types of false information are abundant on the Web, and
regardless of whether a fact is fabricated or misrepresented on pur-
pose or not, the effects it has on people’s lives may be detrimental
and even fatal, as in the case of medical lies [16, 20, 22, 30].
Hoaxes.This paper focuses on a specific kind of disinformation,
namelyhoaxes. Wikipedia defines a hoax as “a deliberately fabri-
cated falsehood made to masquerade as truth.” The Oxford English
Dictionary adds another aspect by defining a hoax as “ahumorous
or mischievous deception” (italics ours).
We study hoaxes in the context of Wikipedia, for which there are
two good reasons: first, anyone can insert information into Wiki-
pedia by creating and editing articles; and second, as the world’s
largest encyclopedia and one of the most visited sites on the Web,
Wikipedia is a major source of information for many people. In
other words: Wikipedia has the potential to both attract and spread
false information in general, and hoaxes in particular.
The impact of some Wikipedia hoaxes has been considerable,
and anecdotes are aplenty. The hoax article about a fake language
called “Balboa Creole French”, supposed to be spoken on Balboa
Island in California, is reported to have resulted in “people com-
ing to [. . . ] Balboa Island to study this imaginary language” [38].
Some hoaxes have made it into books, as in the case of the al-
leged (but fake) Aboriginal Australian god “Jar’Edo Wens”, who
inspired a character’s name in a science fiction book [10] and has
been listed as a real god in at least one nonfiction book [24], all
before it came to light in March 2015 that the article was a hoax.
Another hoax (“Bicholim conflict”) was so elaborate that it was of-
ficially awarded “good article” status and maintained it for half a
decade, before finally being debunked in 2012 [27].
The list of extreme cases could be continued, and the popular
press has covered such incidents widely. What is less available,
however, is a more general understanding of Wikipedia hoaxes that
goes beyond such cherry-picked examples.
Our contributions: impact, characteristics, and detection of
Wikipedia hoaxes.This paper takes a broad perspective by start-
ing from the set of all hoax articles ever created on Wikipedia and
illuminating them from several angles. We study over 20,000 hoax
articles, identified by the fact that they were explicitly flagged as
potential hoaxes by a Wikipedia editor at some point and deleted
after a discussion among editors who concluded that the article was
sciencemag.org SCIENCE
ILLUSTRATION: SÉBASTIEN THIBAULT
By David M. J. Lazer, Matthew A. Baum,
Yochai Benkler, Adam J. Berinsky, Kelly
M. Greenhill, Filippo Menczer, Miriam
J. Metzger, Brendan Nyhan, Gordon
Pennycook, David Rothschild, Michael
Schudson, Steven A. Sloman, Cass R.
Sunstein, Emily A. Thorson, Duncan J.
Watts, Jonathan L. Zittrain
T
he rise of fake news highlights the
erosion of long-standing institutional
bulwarks against misinformation in
theMinternet age. Concern over the
problem is global. However, much
remains unknown regarding the vul-
nerabilities of individuals, institutions, and
society to manipulationsMby malicious actors.
A new system of safeguards is needed. Below,
we discuss extant social and computer sci-
ence research regarding belief in fake news
and the mechanisms by which it spreads.
Fake news has a long history, but we focus
on unanswered scientific questions raised by
the proliferation of its most recent, politically
oriented incarnation. Beyond selected refer-
ences in the text, suggested further reading
can be found in the supplementary materials.
WHAT IS FAKE NEWS?
We define “fake news” to be fabricated in-
formation that mimics news media content
in form but not in organizational process or
intent. Fake-news outlets, in turn, lack the
news media’s editorial norms and processes
for ensuring the accuracy and credibility of
information. Fake news overlaps with other
information disorders, such as misinforma-
tion (false or misleading information) and
disinformation (false information that is pur-
posely spread to deceive people).
Fake news has primarily drawn recent at-
tention in a political context but it also has
been documented in information promul-
gated about topics such as vaccination, nu-
trition, and stock values. It is particularly
pernicious in that it is parasitic on standard
news outlets, simultaneously benefiting from
and undermining their credibility.
Some—notably First Draft and Facebook—
favor the term “false news” because of the
use of fake news as a political weapon (1).
We have retained it because of its value as a
scientific construct, and because its politi-
cal salience draws attention to an impor-
tant subject.
THE HISTORICAL SETTING
Journalistic norms of objectivity and bal-
ance arose as a backlash among journalists
against the widespread use of propaganda
in World War I (particularly their own role
in propagating it) and the rise of corporate
public relations in the 1920s. Local and na-
tional oligopolies created by the dominant
20th century technologies of information
distribution (print and broadcast) sustained
these norms. The internet has lowered the
cost of entry to new competitors—many of
which have rejected those norms—and un-
dermined the business models of traditional
news sources that had enjoyed high levels of
public trust and credibility. General trust in
the mass media collapsed to historic lows in
2016, especially on the political right, with
SOCIAL S CIENCE
The science of fake news
Addressing fake news requires a multidisciplinary effort
INSIGHTS
POLICY FORUM
The list of author affiliations is provided in the supplementary
materials. Email: [email protected]
1094 9 MARCH 2018 • VOL 359 ISSUE 6380
DA_0309PolicyForum.indd 1094 3/7/18 12:16 PM
Published by AAAS
on March 13, 2018

http://science.sciencemag.org/
Downloaded from
Paper list
Click to view !!
Textually similar
(on TF-IDF, D2V etc.)
sciencemag.org SCIENCE
ILLUSTRATION: SÉBASTIEN THIBAULT
By David M. J. Lazer, Matthew A. Baum,
Yochai Benkler, Adam J. Berinsky, Kelly
M. Greenhill, Filippo Menczer, Miriam
J. Metzger, Brendan Nyhan, Gordon
Pennycook, David Rothschild, Michael
Schudson, Steven A. Sloman, Cass R.
Sunstein, Emily A. Thorson, Duncan J.
Watts, Jonathan L. Zittrain
T
he rise of fake news highlights the
erosion of long-standing institutional
bulwarks against misinformation in
theMinternet age. Concern over the
problem is global. However, much
remains unknown regarding the vul-
nerabilities of individuals, institutions, and
society to manipulationsMby malicious actors.
A new system of safeguards is needed. Below,
we discuss extant social and computer sci-
ence research regarding belief in fake news
and the mechanisms by which it spreads.
Fake news has a long history, but we focus
on unanswered scientific questions raised by
the proliferation of its most recent, politically
oriented incarnation. Beyond selected refer-
ences in the text, suggested further reading
can be found in the supplementary materials.
WHAT IS FAKE NEWS?
We define “fake news” to be fabricated in-
formation that mimics news media content
in form but not in organizational process or
intent. Fake-news outlets, in turn, lack the
news media’s editorial norms and processes
for ensuring the accuracy and credibility of
information. Fake news overlaps with other
information disorders, such as misinforma-
tion (false or misleading information) and
disinformation (false information that is pur-
posely spread to deceive people).
Fake news has primarily drawn recent at-
tention in a political context but it also has
been documented in information promul-
gated about topics such as vaccination, nu-
trition, and stock values. It is particularly
pernicious in that it is parasitic on standard
news outlets, simultaneously benefiting from
and undermining their credibility.
Some—notably First Draft and Facebook—
favor the term “false news” because of the
use of fake news as a political weapon (1).
We have retained it because of its value as a
scientific construct, and because its politi-
cal salience draws attention to an impor-
tant subject.
THE HISTORICAL SETTING
Journalistic norms of objectivity and bal-
ance arose as a backlash among journalists
against the widespread use of propaganda
in World War I (particularly their own role
in propagating it) and the rise of corporate
public relations in the 1920s. Local and na-
tional oligopolies created by the dominant
20th century technologies of information
distribution (print and broadcast) sustained
these norms. The internet has lowered the
cost of entry to new competitors—many of
which have rejected those norms—and un-
dermined the business models of traditional
news sources that had enjoyed high levels of
public trust and credibility. General trust in
the mass media collapsed to historic lows in
2016, especially on the political right, with
SOCIAL S CIENCE
The science of fake news
Addressing fake news requires a multidisciplinary effort
INSIGHTS
POLICY FORUM
The list of author affiliations is provided in the supplementary
materials. Email: [email protected]
1094 9 MARCH 2018 • VOL 359 ISSUE 6380
DA_0309PolicyForum.indd 1094 3/7/18 12:16 PM
Published by AAAS
on March 13, 2018

http://science.sciencemag.org/
Downloaded from
Similar items
are recommended

Reviewer Trust
34T. Wang, D. Wang. (2014) “Why Amazon’s Ratings Might Mislead You: The Story of Herding Effects”, Journal of Big Data, Vol.2, No.4, pp.196-204.
specific sequence number (Fig. 1C), finding no significant shift
of average ratings over time across all four different categories.
These observations demonstrate that ratings generation is not
an independent, homogeneous process, supporting our hy-
pothesis that rating systems do not simply aggregate indi-
vidual opinions, but create an environment that influences
subsequent ratings in a systematic manner.
While our results are in good agreement with recent experi-
mental findings on social influence,
suggesting that disclosing prior ratings
by other customers exerts pervasive
and consequential herding effects on
subsequent individual opinions, fur-
ther experiments are needed to con-
clude the extent to which the observed
influence can be attributed to herding.
Yet, as we show next, rating trajecto-
ries follow widely reproducible dy-
namical patterns, as such systematic
influence can be effectively learned
and detected from early rating histo-
ries, which can be used to predict subsequent ratings.
Dynamics of Ratings Growth
We start by identifying two fundamental factors that drive
rating trajectories.
Intrinsic qualitypertains to a product’s true value. A prod-
uct’s value depends on so many intangible and subjective
dimensions that it is impossible to quantify them all. Here,
we view intrinsic quality as a collective measure capturing the
product’s value perceived by the customer community when
each customer expresses his/her opinion independently. In a
discrete 1-to-Kstar ratings system, we can represent a
product’s intrinsic quality as a multinomial distribution
g1,g2,...,gK(
P
K
k!1
gk!1) over 1- to K-star, withgkbeing
the probability density atk-star. For
mathematical convenience, we param-
eterize this multinomial distribution
withKvariablesl
1,l
2,...,l
K,
which satisfygkaexp (lk)for
k2f1, 2,...,Kg.
Herding effectsinfluence ratings
generation in a more intricate
manner. Different prior ratings can
excite, suppress, or have negligible
effects on the generation of a new
rating. Given a ratings sequence
r1,r2,r3,..., the first (i– 1) ratings form thehistoryof the
ith ratingri, which we summarize as a vectorxi!
[xi,1,xi,2,...,xi,K]
T
, withxi,kbeing the number ofk-star
ratings among these (i– 1) ratings. This setting is motivated
by the observation that Amazon displays the counts of dif-
ferent ratings in each product’s review section. We use a
general linear model to capture all possible effectsx
ican exert
‘‘A PRODUCT’S VALUE
DEPENDS ON SO MANY
INTANGIBLE AND SUBJECTIVE
DIMENSIONS THAT IT IS
IMPOSSIBLE TO QUANTIFY
THEM ALL.’’
0 0.2 0.4 0.6 0.8 1
0
0.2
0.4
0.6
0.8
1
P!Value
Cumulative Distribution
0 50 100 150 200 250
2
3
4
5
Sequence Number of Rating
Mean Rating


Books
Electronics
Movies & TV
Music
1050 100150200250
!0.3
!0.2
!0.1
0
0.1
0.2
0.3
Length of Rating History
Pearson’s Correlation Coefficient


1-Star Rating
2-Star Rating
3-Star Rating
4-Star Rating
5-Star Rating
Fitting Curve
0.05
A
C
B
FIG. 1.(A)Cumulative distribution ofp-values of Augment Dickey–Fuller test over all rating sequences in the dataset.(B)Average Pearson’s
correlation coefficient between fractions ofk-star (k=1, 2, 3, 4, 5) ratings in history and the next 10 ratings with respect to length of ratings
history.(C)Mean ratings at specific rating sequence number. Each point is the average over all ratings sequences in the same category.
HERDING EFFECTS IN COLLECTIVE RATINGS
Wang and Wang
198BD
BIG DATA DECEMBER 2014
Fig: Average rating scores on Amazon.com
How to find trustworthy reviewers (rating experts)?
-People often give good scores to items
-Some reviewers intentionally give too high/low scores to items
(spammers)