59
41

Water from Two Rocks: Maximizing the Mutual Information

Abstract

Our goal is to forecast ground truth YY using two sources of information XA,XBX_A,X_B, without access to any data labeled with ground truth. That is, we are aiming to learn two predictors/hypotheses PA,PBP_A^*,P_B^* such that PA(XA)P_A^*(X_A) and PB(XB)P_B^*(X_B) provide high quality forecasts for ground truth YY, without labeled data. We also want to elicit a high quality forecast for YY from the crowds and pay the crowds immediately, without access to YY. We build a natural connection between the learning question and the mechanism design question and deal with them using the same information theoretic approach. Learning: With a natural assumption---conditioning on YY, XAX_A and XBX_B are independent, we reduce the learning question to an optimization problem maxPA,PBMIGf(PA,PB)\max_{P_A,P_B}MIG^f(P_A,P_B) such that solving the learning question is equivalent to picking the PA,PBP_A^*,P_B^* that maximize MIGf(PA,PB)MIG^f(P_A,P_B)---the \emph{ff-mutual information gain} between PAP_A and PBP_B. Moreover, we apply our results to the "learning with noisy labels" problem to learn a predictor that forecasts the ground truth label rather than the noisy label with some side information, without pre-estimating the relationship between the ground truth labels and noisy labels. Mechanism design: We design mechanisms that elicit high quality forecasts without verification and have instant rewards for agents by assuming the agents' information is independent conditioning on YY. In the single-task setting, we propose a forecast elicitation mechanism where truth-telling is a strict equilibrium, in the multi-task setting, we propose a family of forecast elicitation mechanisms where truth-telling is a strict equilibrium and pays better than any other equilibrium.

View on arXiv
Comments on this paper