Part of Advances in Neural Information Processing Systems 18 (NIPS 2005)
Ashish Kapoor, Hyungil Ahn, Yuan Qi, Rosalind Picard
There have been many graph-based approaches for semi-supervised clas- sification. One problem is that of hyperparameter learning: performance depends greatly on the hyperparameters of the similarity graph, trans- formation of the graph Laplacian and the noise model. We present a Bayesian framework for learning hyperparameters for graph-based semi- supervised classification. Given some labeled data, which can contain inaccurate labels, we pose the semi-supervised classification as an in- ference problem over the unknown labels. Expectation Propagation is used for approximate inference and the mean of the posterior is used for classification. The hyperparameters are learned using EM for evidence maximization. We also show that the posterior mean can be written in terms of the kernel matrix, providing a Bayesian classifier to classify new points. Tests on synthetic and real datasets show cases where there are significant improvements in performance over the existing approaches.