ABSTRACT
We compare two recently proposed frameworks for combining generative and discriminative probabilistic classifiers and apply them to semi-supervised classification. In both cases we explore the tradeoff between maximizing a discriminative likelihood of labeled data and a generative likelihood of labeled and unlabeled data. While prominent semi-supervised learning methods assume low density regions between classes or are subject to generative modeling assumptions, we conjecture that hybrid generative/discriminative methods allow semi-supervised learning in the presence of strongly overlapping classes and reduce the risk of modeling structure in the unlabeled data that is irrelevant for the specific classification task of interest. We apply both hybrid approaches within naively structured Markov random field models and provide a thorough empirical comparison with two well-known semi-supervised learning methods on six text classification tasks. A semi-supervised hybrid generative/discriminative method provides the best accuracy in 75% of the experiments, and the multi-conditional learning hybrid approach achieves the highest overall mean accuracy across all tasks.
- M. Belkin, P. Niyogi, and V. Sindhwani. Manifold regularization: A geometric framework for learning from examples. Technical Report TR-2004-06, University of Chicago, 2004.Google Scholar
- A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training. In COLT: Proceedings of the Workshop on Computational Learning Theory, 1998. Google Scholar
Digital Library
- G. Bouchard and B. Triggs. The tradeoff between generative and discriminative classifiers. In J. Antoch, editor, Proceedings in Computational Statistics, 16th Symposium of IASC, volume 16, Prague. Physica-Verlag.Google Scholar
- U. Brefeld, T. Gaertner, T. Scheffer, and S. Wrobel. Efficient co-regularized least squares regression. In ICML06, 23rd International Conference on Machine Learning, Pittsburgh, USA, 2006. Google Scholar
Digital Library
- R. Caruana. Multitask learning. Machine Learning, 28(1):41--75, 1997. Google Scholar
Digital Library
- O. Chapelle, V. Sindhwani, and S. S. Keerthi. Branch and bound for semi-supervised support vector machines. In Advances in Neural Information Processing Systems (NIPS), 2006.Google Scholar
- R. Collobert, F. Sinz, J. Weston, and L. Bottou. Large scale transductive SVMs. The Journal of Machine Learning Research, 7(Aug):1687--1712, 2006. Google Scholar
Digital Library
- T. Joachims. Transductive inference for text classification using support vector machines. In Proc. 16th International Conf. on Machine Learning, pages 200--209. Morgan Kaufmann, San Francisco, CA, 1999. Google Scholar
Digital Library
- C. Kang and J. Tian. A hybrid generative/discriminative bayesian classifier. In Proceedings of the 19th International FLAIRS Conference, 2006.Google Scholar
- B. M. Kelm, C. Pal, and A. McCallum. Combining generative and discriminative methods for pixel classification with multi-conditional learning. ICPR, 2:828--832, 2006. Google Scholar
Digital Library
- J. A. Lasserre, C. M. Bishop, and T. P. Minka. Principled hybrids of generative and discriminative models. In CVPR '06: Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 87--94, Washington, DC, USA, 2006. IEEE Computer Society. Google Scholar
Digital Library
- R. Malouf. A comparison of algorithms for maximum entropy parameter estimation. In In Sixth Conf. on Natural Language Learning, pages 49--55, 2002. Google Scholar
Digital Library
- A. McCallum, C. Pal, G. Druck, and X. Wang. Multi-conditional learning: Generative/discriminative training for clustering and classification. In AAAI '06: American Association for Artificial Intelligence National Conference on Artificial Intelligence, 2006. Google Scholar
Digital Library
- A. K. McCallum. Mallet: A machine learning for language toolkit. http://mallet.cs.umass.edu, 2002.Google Scholar
- A. Ng and M. Jordan. On discriminative vs. generative classifiers: A comparison of logistic regression and naive bayes. In Advances in Neural Information Processing Systems (NIPS), 2002.Google Scholar
- K. Nigam, J. Lafferty, and A. McCallum. Using maximum entropy for text classification, 1999.Google Scholar
- K. Nigam, A. McCallum, S. Thrun, and T. M. Mitchell. Learning to classify text from labeled and unlabeled documents. In AAAI/IAAI, page 792, 1998. Google Scholar
Digital Library
- R. Raina, Y. Shen, A. Y. Ng, and A. McCallum. Classification with hybrid generative/discriminative models. In NIPS, 2003.Google Scholar
- J. Schler, M. Koppel, S. Argamon, and J. Pennebaker. Effects of age and gender on blogging. In 2006 AAAI Spring Symposium on Computational Approaches for Analyzing Weblogs, 2006.Google Scholar
- V. Sindhwani, P. Niyogi, and M. Belkin. A co-regularized approach to semi-supervised learning with multiple views. In Proc. of the 22nd ICML Workshop on Learning with Multiple Views, August 2005.Google Scholar
- D.-Q. Zhang and S.-F. Chang. A generative-discriminative hybrid method for multi-view object detection. CVPR, 2:2017--2024, 2006. Google Scholar
Digital Library
- X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervised learning using Gaussian fields and harmonic functions. In ICML-03, 20th International Conference on Machine Learning, 2003.Google Scholar
Index Terms
Semi-supervised classification with hybrid generative/discriminative methods
Recommendations
A hybrid generative/discriminative method for semi-supervised classification
Training methods for machine learning are often characterized as being generative or discriminative. We present a new co-training style algorithm which employs a generative classifier (Naive Bayes) and a discriminative classifier (Support Vector Machine)...
Combining deep generative and discriminative models for Bayesian semi-supervised learning
Highlights- Modelling framwork that enables Bayesian semi-supervised learning.
- Bayesian ...
AbstractGenerative models can be used for a wide range of tasks, and have the appealing ability to learn from both labelled and unlabelled data. In contrast, discriminative models cannot learn from unlabelled data, but tend to outperform their ...
Semi-supervised multi-label classification using incomplete label information
Highlights- An inductive semi-supervised method called Smile is proposed for multi-label classification using incomplete label information.
AbstractClassifying multi-label instances using incompletely labeled instances is one of the fundamental tasks in multi-label learning. Most existing methods regard this task as supervised weak-label learning problem and assume sufficient ...





Comments