Water from Two Rocks: Maximizing the Mutual Information

Our goal is to forecast ground truth using two sources of information , without access to any data labeled with ground truth. That is, we are aiming to learn two predictors/hypotheses such that and provide high quality forecasts for ground truth , without labeled data. We also want to elicit a high quality forecast for from the crowds and pay the crowds immediately, without access to . We build a natural connection between the learning question and the mechanism design question and deal with them using the same information theoretic approach. Learning: With a natural assumption---conditioning on , and are independent, we reduce the learning question to an optimization problem such that solving the learning question is equivalent to picking the that maximize ---the \emph{-mutual information gain} between and . Moreover, we apply our results to the "learning with noisy labels" problem to learn a predictor that forecasts the ground truth label rather than the noisy label with some side information, without pre-estimating the relationship between the ground truth labels and noisy labels. Mechanism design: We design mechanisms that elicit high quality forecasts without verification and have instant rewards for agents by assuming the agents' information is independent conditioning on . In the single-task setting, we propose a forecast elicitation mechanism where truth-telling is a strict equilibrium, in the multi-task setting, we propose a family of forecast elicitation mechanisms where truth-telling is a strict equilibrium and pays better than any other equilibrium.
View on arXiv