A Probabilistic Framework for Discriminative and Neuro-Symbolic Semi-Supervised Learning
- BDL

In semi-supervised learning (SSL), a rule to predict labels for data is learned from labelled data and unlabelled samples . Strong progress has been made by combining a variety of methods, some of which pertain to , e.g. data augmentation that generates artificial samples from true ; whilst others relate to model outputs , e.g. regularising predictions on unlabelled data to minimise entropy or induce mutual exclusivity. Focusing on the latter, we fill a gap in the standard text by introducing a unifying probabilistic model for discriminative semi-supervised learning, mirroring that for classical generative methods. We show that several SSL methods can be theoretically justified under our model as inducing approximate priors over predicted parameters of . For tasks where labels represent binary attributes, our model leads to a principled approach to neuro-symbolic SSL, bridging the divide between statistical learning and logical rules.
View on arXiv