forked from gjord/gwern.net
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathAnchoring.page
315 lines (246 loc) · 24.2 KB
/
Anchoring.page
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
---
title: LW anchoring experiment
description: Do mindless positive/negative comments skew article quality ratings up and down?
created: 27 Feb 2012
tags: statistics, experiments, psychology
status: finished
confidence: unlikely
importance: 1
...
> I do an informal experiment testing whether LessWrong karma scores are susceptible to a form of anchoring based on the first comment posted; a medium-large effect size is found although the data does not fit the assumed normal distribution so there may or may not be any actual anchoring effect.
# Problem
27 February 2012 in `#lesswrong`:
Grognor> I've been reading the highest-scoring articles, and I have noticed a pattern
a MUCH HIGHER PROPORTION of top-scoring articles have "upvoted" in the first two words in the
first comment
(standard disclaimer: correlation is not causation blah blah)
http://lesswrong.com/r/lesswrong/lw/6r6/tendencies_in_reflective_equilibrium/ then I see this,
one of the follow-ups to one of the top-scoring articles like this. the first comment
says "not upvoted"
and it has a much lower score
while reading it, I was wondering "why is this at only 23? this is one of the best articles
I've ever oh look at that comment"
I'm definitely hitting on a real phenomenon here, probably some social thing that says
"hey if he upvoted, I should too" and it seems to be incredibly
strongly related to the first comment
the proportion is really quite astounding
http://lesswrong.com/lw/5n6/i_know_im_biased_but/ compare this to the
Tendencies in Reflective Equilibrium post. Compared to that, it's awful, but it has
nearly the same score. Note the distinct lack of a first comment saying "not upvoted"
(side note: I thought my article, "on saying the obvious", would have a much lower score than
it did. note the first comment: "good points, all of them.")
it seems like it's having more of an effect than I would naively predict
...
gwern> Grognor: hm. maybe I should register a sockpuppet and on every future article I write flip a coin
and write either upvoted or downvoted
quanticle> gwern: Aren't you afraid you'll incur Goodhart's wrath?
gwern> quanticle: no, that would be if I only put in 'upvoted' comments
gwern> Grognor: do these comments tend to include any reasons?
Grognor> gwern: yes
Boxo> you suggesting that the comments cause the upvotes? I'd rather say that the post is just the kind
of post that makes lots people think as their first reaction
"hell yeah imma upvote this", makes upvoting salient to them, and then some of that bubbles up to
the comments
Grognor> Boxo: I'm not suggesting it's entirely that simple, no, but I do think it's obvious now that a
first comment that says "upvoted for reasons x, y, and z"
will cause more people to upvote than otherwise would have, and vice versa
Boxo> (ie. you saw X and Y and though X caused Y, but I think there's a Z that causes both X and Y)
ksotala> Every now and then, I catch myself wanting to upvote something because others have upvoted it
already. It sounds reasonable that having an explicit comment
declaring "I upvoted" might have an even stronger effect.
ksotala> On the other hand, I usually decide to up/downvote before reading the comments.
gwern> ksotala: you should turn on anti-kibitzing then
rmmh> gwern: maybe karma blinding like HN would help
Boxo> I guess any comment about voting could remind people to vote, in whatever direction. Could test this
if you had the total number of votes per post.
Grognor> that too. the effect here is multifarious and complicated and the intricate details could not
possibly be worked out, which is exactly why this proportion of first comments with an 'upvoted'
note surprises me
Such an [anchoring](!Wikipedia) or [social proof](!Wikipedia) effect resulting in a [first-mover advantage](!Wikipedia) seems plausible to me. At least one experiment found something similar online: ["Social Influence Bias: A Randomized Experiment"](http://web.natur.cuni.cz/~houdek3/papers/Muchnik%20et%20al%202013.pdf "Muchnik et al 2013"):
> Our society is increasingly relying on the digitized, aggregated opinions of others to make decisions. We therefore designed and analyzed a large-scale randomized experiment on a social news aggregation Web site to investigate whether knowledge of such aggregates distorts decision-making. Prior ratings created significant bias in individual rating behavior, and positive and negative social influences created asymmetric herding effects. Whereas negative social influence inspired users to correct manipulated ratings, positive social influence increased the likelihood of positive ratings by 32% and created accumulating positive herding that increased final ratings by 25% on average. This positive herding was topic-dependent and affected by whether individuals were viewing the opinions of friends or enemies. A mixture of changing opinion and greater turnout under both manipulations together with a natural tendency to up-vote on the site combined to create the herding effects. Such findings will help interpret collective judgment accurately and avoid social influence bias in collective intelligence in the future.
# Design
So on the 27th, I registered the account ["Rhwawn"](http://lesswrong.com/user/Rhwawn/)^[If you were wondering about the account name: both 'Rhwawn' and 'Gwern' are character names from the Welsh collection _[Mabinogion](!Wikipedia)_. They share the distinctions of being short, nearly unique, and obviously pseudonymous to anyone who Googles them, which is why I also used that name as an alternate account [on Wikipedia](!Wikipedia "User:Rhwawn").]. I made some quality comments and upvotes to seed the account as a legitimate active account.
Thereafter, whenever I wrote an Article or Discussion, after making it public, I flipped a coin and if Heads, I posted a comment as Rhwawn saying "Upvoted" or if Tails, a comment saying "Downvoted" with some additional text (see next section). Needless to say, no actual vote was made. I then made a number of quality comments and votes on other Articles/Discussions to camouflage the experimental intervention. (In no case did I upvote or downvote someone I had already replied to or voted on with my [Gwern](http://lesswrong.com/user/gwern) account.) Finally, I scheduled a reminder on my calendar for 30 days later to record the karma on that Article/Discussion. I don't post *that* often, so I decided to stop after 1 year, on 27 February 2013. I wound up breaking this decision since by September 2012 I had ceased to find it an interesting question, it was an unfinished task that was burdening my mind, and the necessity of making some genuine contributions as Rhwawn to cloak a anchoring comment was a not-so-trivial inconvenience that was stopping me from posting.
To enlarge the sample, I passed [Recent Posts](http://lesswrong.com/recentposts) through `xclip -o | grep '^by ' | cut -d ' ' -f 2 | sort | uniq -c | sort -g` and picking everyone with >=6 posts (8 people excluding me), and I messaged them with a short message explaining my desire for a large sample and the burden of participation ("It would require perhaps half a minute to a minute of your time every time you post an Article or Discussion for the next year, which is for most of you no more than once a week or month.")
For those who replied, I sent a copy of this writeup and explained their procedure would be as follows: every time they posted they would flip a coin and post likewise (the Rhwawn account password having been shared with them); however, as a convenience to them, I would take care of recording the karma a month later. (I subscribed to participants' post RSS feeds; this would not guarantee that I would learn of their posts in time to add a randomized sock comment - hence the need for their active participation - but I could at least handle the scheduling & karma-checking for them.)
## Comment variation
Grognor pointed that the original comments came with reasons, but unfortunately if I came up with reasons for either comment, some criticisms or praise would be better than others and this would be *another* source of variability; so I added generic reasons.
I know from watching them plummet into oblivion that comments which are just "Upvoted" or "Downvoted" are not a good idea for any anchoring question - they'll quickly be hidden, so any effect size will be a lot smaller than usual, and it's possible that hidden comments themselves anchor (my guess: negatively, by making people think "why is this attracting stupid comments?").
While if you go with more carefully rationalized comments, that's sort of like the [XKCD cartoon](http://xkcd.com/810/): quality comments start to draw on the experimenter's own strengths & weaknesses (I'm sure I could make both quality criticisms and praises of psychology-related articles, but not so much technical decision theory articles).
It's a "damned if you do, damned if you don't" sort of dilemma.
I hoped my strategy would be a golden mean of not too trivial to be downvoted into oblivion, but not so high-quality and individualized that comparability was lost. I think I came close, since as we see in the analysis section, the positive anchoring comments saw only a small negative net downvote, indicating LWers may not have regarded it as good enough to upvote but also not so obviously bad as to merit a downvote.
(Of course, I didn't expect the positive and negative comments to be treated differently - they're pretty much the same thing, with a negation. I'm not sure how I would have designed it differently if I had known about the double-standard in advance.)
To avoid issues with some criticisms being accurate and others poor, a fixed list of reasons was used with minor variation to make them fit in:
1. Negative; "Downvoted; ..."
- "too much weight on _n_ studies"
- "too many studies cited"
- "too reliant on personal anecdote"
- "too heavy on math"
- "not enough math"
- "rather obvious"
- "not very interesting"
2. Positive; "Upvoted; ..."
- "good use of _n_ studies"
- "thorough citation of claims"
- "enjoyable anecdotes"
- "rigorous use of math"
- "just enough math"
- "not at all obvious"
- "very interesting"
# Questions
> I will have to do some contemplation of values before I accept or reject. I like getting honest feedback on my posts, I like accumulating karma, and I also like performing experiments.
Randomization suggests that your expected-karma-value would be 0, unless you expect asymmetry between positive and negative.
> What do you anticipate doing with the data accumulated over the course of the experiment?
Oh, it'd be simple enough. Sort articles into one group of karma scores for the positive anchors, the other group for the negative anchors; feed into a two-sample t-test to see if the means differ and if the difference is significant. I can probably copy the R code straight from my [various Zeo-related `R`](Zeo#vitamin-d-analysis) sessions.
If I can hit _p_<0.10 or _p_<0.05 or so, post an Article triumphantly announcing the finding of bias and an object lesson of why one shouldn't take karma too seriously; if I don't, post a Discussion article discussing it and why I thought the results didn't reach significance. (Not enough articles? Too weak assumptions in my t-test?)
> And the ethics?
The post authors are volunteers, and as already pointed out, the expected karma benefit is 0. So no one is harmed, and as for the deception, it does not seem to me to be a big deal. We are already nudged by countless primes and stimuli and biases, so another one, designed to be neutral in total effect, seems harmless to me.
> "What comes before determines what comes after...The thoughts of all men arise from the darkness. If you are the movement of your soul, and the cause of that movement precedes you, then how could you ever call your thoughts your own? How could you be anything other than a slave to the darkness that comes before?...History. Language. Passion. Custom. All these things determine what men say, think, and do. These are the hidden puppet-strings from which all men hang...all men are deceived....So long as what comes before remains shrouded, so long as men are already deceived, what does [deceiving men] matter?" --Kelhus, [R. Scott Bakker](!Wikipedia)'s _[The Darkness That Comes Before](!Wikipedia)_
# Data
The results:
Post Author Date Anchor Post karma Comment karma
-------------------------------------------------------- ------------------ -------- -------- ---------- -------------
[Cashing Out Cognitive Biases as Behavior][] [gwern][] 02 March 0 11 -4
[Heuristics and Biases in Charity][] [Kaj_Sotala][] 02 March 0 19 -6
[I Was Not Almost Wrong But I Was Almost Right][] Kaj_Sotala 08 March 1 50 0
[Emotional regulation, Part I: a problem summary][] [Swimmer963][] 05 March 0 9 -2
[How would you stop Moore's Law?][] gwern 10 March 1 19 -3
[On the etiology of religious belief][] gwern 11 March 0 11 -7
[Decision Theories: A Less Wrong Primer][] [orthonormal][] 13 March 1 62 0
[Schelling fences on slippery slopes][] [Yvain][] 16 March 1 120 1
[Fallacies as weak Bayesian evidence][] Kaj_Sotala 18 March 1 49 0
[Decision Theories: A Semi-Formal Analysis, Part I][] orthonormal 24 March 1 20 -1
[Decision Theories: A Semi-Formal Analysis, Part II][] orthonormal 06 April 0 16 -13
[To like each other, sing and dance in synchrony][] Kaj_Sotala 23 April 0 20 4
[The state of life extension research][] gwern 23 April 1 10 0
[Value of Information: 8 examples][] gwern 18 May 0 45 -8
[Hope Function][] gwern 01 July 1 22 0
[To Learn Critical Thinking, Study Critical Thinking][] gwern 07 July 0 23 -11
[Dragon Ball's Hyperbolic Time Chamber][] gwern 02 Sep 0 33 -9
[Dragon Ball's Hyperbolic Time Chamber]: http://lesswrong.com/r/discussion/lw/ecp/dragon_balls_hyperbolic_time_chamber/
[To Learn Critical Thinking, Study Critical Thinking]: http://lesswrong.com/lw/dhe/to_learn_critical_thinking_study_critical_thinking/
[Hope Function]: http://lesswrong.com/r/discussion/lw/dds/hope_function/
[Value of Information: 8 examples]: http://lesswrong.com/r/discussion/lw/cih/value_of_information_8_examples/
[The state of life extension research]: http://lesswrong.com/r/discussion/lw/bl3/living_forever_is_hard_part_3_the_state_of_life/
[To like each other, sing and dance in synchrony]: http://lesswrong.com/lw/by9/to_like_each_other_sing_and_dance_in_synchrony/
[Decision Theories: A Semi-Formal Analysis, Part II]: http://lesswrong.com/lw/az6/decision_theories_a_semiformal_analysis_part_ii/
[Decision Theories: A Semi-Formal Analysis, Part I]: http://lesswrong.com/lw/axl/decision_theories_a_semiformal_analysis_part_i/
[Fallacies as weak Bayesian evidence]: http://lesswrong.com/r/lesswrong/lw/aq2/fallacies_as_weak_bayesian_evidence/
[Yvain]: http://lesswrong.com/user/Yvain/
[Schelling fences on slippery slopes]: http://lesswrong.com/lw/ase/schelling_fences_on_slippery_slopes/
[orthonormal]: http://lesswrong.com/user/orthonormal/
[Decision Theories: A Less Wrong Primer]: http://lesswrong.com/lw/aq9/decision_theories_a_less_wrong_primer/
[On the etiology of religious belief]: http://lesswrong.com/r/discussion/lw/aq6/on_the_etiology_of_religious_belief/
[How would you stop Moore's Law?]: http://lesswrong.com/r/discussion/lw/apm/how_would_you_stop_moores_law/
[Swimmer963]: http://lesswrong.com/user/Swimmer963/
[Emotional regulation, Part I: a problem summary]: http://lesswrong.com/lw/aks/emotional_regulation_part_i_a_problem_summary/
[I Was Not Almost Wrong But I Was Almost Right]: http://lesswrong.com/lw/akr/i_was_not_almost_wrong_but_i_was_almost_right/
[Kaj_Sotala]: http://lesswrong.com/user/Kaj_Sotala/
[Heuristics and Biases in Charity]: http://lesswrong.com/lw/aid/heuristics_and_biases_in_charity/
[gwern]: http://lesswrong.com/user/gwern
[Cashing Out Cognitive Biases as Behavior]: http://lesswrong.com/r/discussion/lw/ahz/cashing_out_cognitive_biases_as_behavior/
# Analysis
For the analysis, I have 2 questions:
1. Is there a difference in karma between posts that received a negative initial comment and those that received a positive initial comment? (Any difference suggests that one or both are having an effect.)
2. Is there a difference in karma between the two kinds of initial comments (as I began to suspect during the experiment)?
## Article effect
Some Bayesian inference using [BEST](http://www.indiana.edu/~kruschke/BEST/):
~~~{.R}
lw <- data.frame(Anchor = c( 0, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0),
Post.karma = c(11,19,50, 9,19,11,62,120,49,20,16,20,10,45,22,23,33))
source("BEST.R")
neg <- lw[lw$Anchor==0,]$Post.karma
pos <- lw[lw$Anchor==1,]$Post.karma
mcmc = BESTmcmc(neg, pos)
BESTplot(neg, pos, mcmcChain=mcmc)
SUMMARY.INFO
PARAMETER mean median mode HDIlow HDIhigh pcgtZero
mu1 20.1792 20.104 20.0392 10.7631 29.9835 NA
mu2 41.9474 41.640 40.4661 11.0307 75.1056 NA
muDiff -21.7682 -21.519 -22.6345 -55.3222 11.2283 8.143
sigma1 13.1212 12.264 10.9018 5.8229 22.4381 NA
sigma2 40.9768 37.835 33.8565 16.5560 72.6948 NA
sigmaDiff -27.8556 -24.995 -21.7802 -60.9420 -0.9855 0.838
nu 30.0681 21.230 5.6449 1.0001 86.5698 NA
nuLog10 1.2896 1.327 1.4332 0.4332 2.0671 NA
effSz -0.7718 -0.765 -0.7632 -1.8555 0.3322 8.143
~~~
![Graphical summary of BEST results for full dataset](/images/lw-anchoring-yvain.png)
The results are heavily skewed by Yvain's very popular post; we can't trust any results based on such a high scoring post. Let's try omitting Yvain's datapoint. BEST actually crashes displaying the result, perhaps due to making an assumption about there being at least 8 datapoints or something; and it's question whether we should be using a normal-based test like BEST in the first place: just from graphing we can see it's definitely not a normal distribution. So we'll fall back to a distribution-free [Mann-Whitney U](!Wikipedia) two-sample test (rather than a _t_-test):
~~~{.R}
wilcox.test(Post.karma ~ Anchor, data=lw)
Wilcoxon rank sum test with continuity correction
data: Post.karma by Anchor
W = 19, p-value = 0.1117
...
wilcox.test(Post.karma ~ Anchor, data=(lw[lw$Post.karma<100,]))
Wilcoxon rank sum test with continuity correction
data: Post.karma by Anchor
W = 19, p-value = 0.203
~~~
Reasonable. To work around the bug, let's replace Yvain by the mean for that group without him, 33; the new results:
~~~{.R}
SUMMARY.INFO
PARAMETER mean median mode HDIlow HDIhigh pcgtZero
mu1 20.2877 20.2002 20.1374 10.863 29.9532 NA
mu2 32.7912 32.7664 32.8370 15.609 50.4410 NA
muDiff -12.5035 -12.4802 -12.2682 -32.098 7.3301 9.396
sigma1 13.2561 12.3968 10.9385 6.044 22.3085 NA
sigma2 22.4574 20.6784 18.3106 10.449 38.5859 NA
sigmaDiff -9.2013 -8.1031 -7.1115 -28.725 7.6973 11.685
nu 33.2258 24.5819 8.6726 1.143 91.3693 NA
nuLog10 1.3575 1.3906 1.4516 0.555 2.0837 NA
effSz -0.7139 -0.7066 -0.7053 -1.779 0.3607 9.396
~~~
![Graphical summary of BEST results for dataset with Yvain replaced by a mean](/images/lw-anchoring-yvain-omitted.png)
The difference in means has shrunk but not gone away; it's large enough that 10% of the possible effect sizes (of "a negative initial comment rather than positive") may be zero or actually be positive (increase karma) instead. This is a little concerning, but I don't take this too seriously:
1. this is not a lot of data
2. as we've seen there are extreme outliers suggesting that the assumption that karma scores are Gaussian/normal may be badly wrong
3. even at face value, 10 karma points doesn't seem like it's large enough to have any important real-world consequences (like make people leave LW who should've stayed)
Pursuing point number two, there's two main options for dealing with the normal distribution assumption of BEST/_t_-tests being violated: either switch to a test which assumes a distribution more like what we actually see & hope that this new distribution is close enough to true, or use a test which doesn't rely on distributions at all.
For example, we could try to model it as a [Poisson distribution](!Wikipedia) and see if the anchoring variable is an important predictor of the mean of the process sum, which it seems to be:
~~~{.R}
summary(glm(Post.karma ~ Anchor, data=lw, family=poisson))
...
Deviance Residuals:
Min 1Q Median 3Q Max
-6.194 -2.915 -0.396 0.885 9.423
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) 3.0339 0.0731 41.49 <2e-16
Anchor 0.7503 0.0905 8.29 <2e-16
...
summary(glm(Post.karma ~ Anchor, data=(lw[lw$Post.karma<100,]), family=poisson))
...
Deviance Residuals:
Min 1Q Median 3Q Max
-4.724 -2.386 -0.744 2.493 4.594
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) 3.0339 0.0731 41.49 <2e-16
Anchor 0.4669 0.0983 4.75 2e-06
~~~
(On a side note, I regard these _p_-values as evidence for an effect even thought they don't fall under 0.05 or another alpha I defined in advance: with this small sample size and hence low [statistical power](!Wikipedia), to reach _p_<0.05, each anchoring comment would have to have a grotesquely large effect on article karma - but anchoring comments having such an effect is highly unlikely! Anchoring, in psychology, isn't *that* omnipotent: it's relatively subtle. So we have a similar problem as with the [2013 Lewis meditation results](Lewis meditation) - they're the sort of situations where it'd be nicer to be talking in more relative terms like [Bayes factor](!Wikipedia)s.)
## Comment treatment
How did these mindless unsubstantiated comments either praising or criticizing an article get treated by the community? Let's look at the anchoring comment's karma:
~~~{.R}
neg <- lw[lw$Anchor==0,]$Comment.karma
pos <- lw[lw$Anchor==1,]$Comment.karma
mcmc = BESTmcmc(neg, pos)
BESTplot(neg, pos, mcmcChain=mcmc)
SUMMARY.INFO
PARAMETER mean median mode HDIlow HDIhigh pcgtZero
mu1 -6.4278 -6.4535 -6.55032 -10.5214 -2.2350 NA
mu2 -0.2755 -0.2455 -0.01863 -1.3180 0.7239 NA
muDiff -6.1523 -6.1809 -6.25451 -10.3706 -1.8571 0.569
sigma1 5.6508 5.2895 4.70143 2.3262 9.7424 NA
sigma2 1.2614 1.1822 1.07138 0.2241 2.4755 NA
sigmaDiff 4.3893 4.0347 3.53457 1.1012 8.5941 99.836
nu 27.4160 18.1596 4.04827 1.0001 83.9648 NA
nuLog10 1.2060 1.2591 1.41437 0.2017 2.0491 NA
effSz -1.6750 -1.5931 -1.48805 -3.2757 -0.1889 0.569
~~~
![Graphical summary of BEST results for full dataset of how the positive/negative comments were treated](/images/lw-anchoring-comment.png)
As one would hope, neither group of comments ends up with net positive mean score, but they're clearly being treated very differently: the negative comments get downvoted far more than the positive comments. I take this as perhaps implying that LW's reputation for being negative & hostile is a bit overblown: we're negative and hostile to poorly thought out criticisms and arguments, not fluffy praise.
# External links
- [LW discussion](http://lesswrong.com/lw/gft/lw_anchoring_experiment_maybe/)