A few weeks ago I attended Scott Kominers‘s class on Market Design. They were talking about mechanism design and differential privacy so I felt like it would be fun to attend that session. In the class Scott mentioned some interesting work by Nicholas Lambert and Yoav Shoham on Truthful Surveys that appeared at WINE 2008. There’s also some recent work by Aaron Roth and Grant Schoenebeck up on ArXiV.
In Lambert and Shoham’s set up, the opinion distribution of a population is given by some CDF (with a density) on the unit interval
. We can think of
as a level of approval (say of a politician) and
as the proportion of the population which has approval less than
. A surveyor selects
agents
i.i.d. from
and asks them to report their opinion. They can report anything they like, however, so they will report
. In order to incentivize them, the surveyor will issue a payment
to each agent
. How should we structure the payments to incentivize truthful reporting? In particular, can we make a mechanism in which being truthful is a Nash equilibrium (“accurate”) or the only Nash equilibrium (“strongly accurate”)?
Let . They propose partitioning the agents into
groups with
denoting the group of agent $i$, and
as an unbiased estimator of
that uses the points
. The payments are:
This mechanism is accurate and also permutation-invariant with respect to the agents (“anonymous”) and the sum of the payments is 0 (“budget-balanced”).
This is an instance of a more general mechanism for truthfully inducing samples from a collection of distributions that are known — each agent has a distribution and you want to get their sample of that distribution. Here what they do is replace the known distributions with empirical estimates, in a sense. Why is this only accurate and not strongly accurate? It is possible that the agents could collude and pick a different common distribution
and report values from that. Essentially, each group has an incentive to report from the same distribution and then globally the optimal thing is for all the groups to report from the same distribution, but that distribution need not be
if there is global collusion. How do we get around this issue? If there is a set of “trusted” agents
, then the estimators in the payment model can be built using the trusted data and the remaining untrusted agents can be put in a single group whose optimal strategy is now to follow the trusted agents. That mechanism is strongly accurate. In a sense the trusted agents cause the population to “gel” under this payment strategy.
It seems that Roth and Schoenbeck are not aware of Lambert and Shoham’s work, or it is sufficiently unrelated (they certainly don’t cite it). They also look at truth in surveying from a mechanism design perspective. Their model is somewhat more involved (an has Bayesian bits), but may be of interest to readers who like auction design.