Abstract
We consider a setting for discriminative semi-supervised learning where unlabeled data are used with a generative model to learn effective feature representations for discriminative training. Within this framework, we revisit the two-view feature generation model of co-training and prove that the optimum predictor can be expressed as a linear combination of a few features constructed from unlabeled data. From this analysis, we derive methods that employ two views but are very different from co-training. Experiments show that our approach is more robust than co-training and EM, under various data generation conditions.
Original language | English (US) |
---|---|
Pages | 25-32 |
Number of pages | 8 |
DOIs | |
State | Published - 2007 |
Externally published | Yes |
Event | 24th International Conference on Machine Learning, ICML 2007 - Corvalis, OR, United States Duration: Jun 20 2007 → Jun 24 2007 |
Other
Other | 24th International Conference on Machine Learning, ICML 2007 |
---|---|
Country/Territory | United States |
City | Corvalis, OR |
Period | 6/20/07 → 6/24/07 |
ASJC Scopus subject areas
- Software
- Human-Computer Interaction
- Computer Vision and Pattern Recognition
- Computer Networks and Communications