Last week, I had the pleasure of talking with CMU professor George Loewenstein, one of the top researchers in the area of behavioral economics. I mentioned my idea of using prediction markets to address the weaknesses of online review systems and reputation systems, and he offered two insightful pointers.
The first pointer was to the notion of pluralistic ignorance. As summarized on Wikipedia:
In social psychology, pluralistic ignorance, a term coined by Daniel Katz and Floyd H. Allport in 1931, describes “a situation where a majority of group members privately reject a norm, but assume (incorrectly) that most others accept it…It is, in Krech and Crutchfield’s (1948, pp. 388–89) words, the situation where ‘no one believes, but everyone thinks that everyone believes'”. This, in turn, provides support for a norm that may be, in fact, disliked by most people.
It had not occurred to me that pluralistic ignorance could wreak havoc on the prediction market approach I proposed. Specifically, there is a risk that, even though the majority participants in the market hold a particular opinion, they suppress their individual opinions and instead vote based on mistaken assumptions about the collective opinion of others. Ironically, these participants are pursuing an optimal strategy, given their pluralistic ignorance. Yet the results of such a market would not necessarily reflect the true collective opinion of participants. Clearly there is a need to incorporate people’s true opinions into the equation, and not just their beliefs about others’ opinions.
Which leads me to the second resource to which Loewenstein pointed me: a paper by fellow behavioral economist and MIT professor Drazen Prelec entitled “A Bayesian Truth Serum for Subjective Data“. As per the abstract:
Subjective judgments, an essential information source for science and policy, are problematic because there are no public criteria for assessing judgmental truthfulness. I present a scoring method for eliciting truthful subjective data in situations where objective truth is unknowable. The method assigns high scores not to the most common answers but to the answers that are more common than collectively predicted, with predictions drawn from the same population. This simple adjustment in the scoring criterion removes all bias in favor of consensus: Truthful answers maximize expected score even for respondents who believe that their answer represents a minority view.
Most of the paper is devoted to proving, subject to a few assumptions, that the optimal strategy for players in this game is to tell what they believe to be the truth–that is, the truth-telling strategy is the optimal Bayesian Nash equilibrium for all players.
The assumptions are as follows:
- The sample of respondents is sufficiently large that a single answer cannot appreciably affect the overall results.
- Respondents believe that others sharing their opinion will draw the same inferences about population frequencies.
- All players assume that other players are responding truthfully–which follows if they are rational players.
Prelec sums up his results as follows:
In the absence of reality checks, it is tempting to grant special status to the prevailing consensus. The benefit of explicit scoring is precisely to counteract informal pressures to agree (or perhaps to “stand out” and disagree). Indeed, the mere existence of a truth-inducing scoring system provides methodological reassurance for social science, showing that subjective data can, if needed, be elicited by means of a process that is neither faith-based (“all answers are equally good”) nor biased against the exceptional view.
Unfortunately, I don’t think that Prelec’s assumptions hold for most online review systems and reputation systems. In typical applications (e.g., product and service reviews on sites like Amazon and Yelp), the input is too sparse to even approximate the first assumption, and the other two assumptions probably ascribe too much rationality to the participants.
Still, Bayesian truth serum is a step in the right direction, and perhaps the approach (or some simple variant of it) applies to a useful subset of real-world prediction scenarios. Certainly it gives me hope that we’ll succeed in the quest to mine “subjective truth” from crowds.
6 replies on “Pluralistic Ignorance and Bayesian Truth Serum”
I’ve always been skeptical of prediction markets. While public markets work well for economics, predictions aren’t exactly like that. Prediction markets are at best a lagging indicator, and a lagging indicator doesn’t meet my definition of a prediction. I’m thinking specifically of the prediction markets for picking McCain’s running mate in 2008. Was Palin even on the inTrade list to choose from? I don’t think she was, but I don’t really remember. Same thing with the proposed TIA terrorist act prediction market.
From what I understand, the logic of prediction markets hangs on the Wisdom of Crowds(tm), which comes from the 19th century ox weighing contest. The key component to that contest was that everyone’s guess was secret. This elements the flocking effect which leads to pluralistic ignorance.
Real markets work because it is an actual market. People really want the things being sold. Prediction markets are just a game with nothing to lose and nothing to gain. The amounts “invested” are so low, you might as well be playing Zynga Fantasy Football.
LikeLike
I understand the skepticism. But I think this problem is a little different. It’s not trying to predict an unknown future event, but rather to aggregate present opinions.
Specifically, the goal is to aggregate the disinterested opinions of a representative population sample. But realistically we can only hope to attract a population sample that is neither representative nor disinterested. I thought prediction markets might be a way to get there, but hadn’t considered the problem of pluralistic ignorance–which is a general problem for prediction markets where participants are trying to predict the opinions of others. The Bayesian truth serum approach looks like promising direction for addressing this problem.
LikeLike
You do not need a situation of pluralistic ignorance to lead you to a bad equilibrium; the bad equilibrium always exists for any mechanism if there is no objective reality that can be measured that you are trying to predict, i.e. opinions. As long as the players can coordinate on some report other than the truth (say via a prejudice they all know to be false, and know all others know is false, but every expects to be reported none the less) there is no incentive to break the coordination (so even through I dont hold the view and I know no one else does, if I expect everyone to report it is in my intrest to report it and this is a Nash Equilibrium). This is a fairly generic result for any mechanism; prediction market, or bayesian truth serum, or what not, as long as you only use the reports of others in the mechanisms to score participants. This argument can be formalised (see for example “Crowd & Prjeudice”, http://arxiv.org/abs/1204.3511 , disclaimer: I am a co-author).
A more interesting paper by Witowski and Parker tries to improve on the Bayesian Truth Serum. In particular http://www.eecs.harvard.edu/econcs/pubs/witkowski_aaai12.pdf
From the abstract: “BTS is only incentive compatible for a large enough number of agents, and the particular number of agents required is uncertain because it depends on this private prior. … present a robust BTS for the elicitation of binary information which is incentive compatible for every n 3”
The fact that it is incentive compatible for every n>3 means we can use it as the scoring mechanisms in a prediction-market-like system where every 3 reports we score the respondents and provide a “market price” (so we cant update the price after every user reports, but we can every 3 users).
It would be fascinating to see the RBTS implemented in the wild.
LikeLike
Nikete, thanks for the references! Your paper certainly seems to put the nail in the coffin as a far as prediction markets are concerned. Will have to check out the Witowski and Parker — RBTS certainly sounds encouraging.
LikeLike
[…] and practical questions about how to obtain or incent a representative participant pool — a topic I’ve been passionate about for a long […]
LikeLike
[…] and practical questions about how to obtain or incent a representative participant pool — a topic I’ve been passionate about for a long […]
LikeLike