Supervised learning with probabilistic morphisms and kernel mean
embeddings
- GAN
In this paper I propose a concept of a correct loss function in a generative model of supervised learning for an input space and a label space , both of which are measurable spaces. A correct loss function in a generative model of supervised learning must accurately measure the discrepancy between elements of a hypothesis space of possible predictors and the supervisor operator, even when the supervisor operator does not belong to . To define correct loss functions, I propose a characterization of a regular conditional probability measure for a probability measure on relative to the projection as a solution of a linear operator equation. If is a separable metrizable topological space with the Borel -algebra $ \mathcal{B} (\mathcal{Y})$, I propose an additional characterization of a regular conditional probability measure as a minimizer of mean square error on the space of Markov kernels, referred to as probabilistic morphisms, from to . This characterization utilizes kernel mean embeddings. Building upon these results and employing inner measure to quantify the generalizability of a learning algorithm, I extend a result due to Cucker-Smale, which addresses the learnability of a regression model, to the setting of a conditional probability estimation problem. Additionally, I present a variant of Vapnik's regularization method for solving stochastic ill-posed problems, incorporating inner measure, and showcase its applications.
View on arXiv