Abstract
We work with a model of object recognition where words must be placed on image regions. This approach means that large scale experiments are relatively easy, so we can evaluate the effects of various early and mid-level vision algorithms on recognition performance. We evaluate various image segmentation algorithms by determining word prediction accuracy for images segmented in various ways and represented by various features. We take the view that good segmentations respect object boundaries, and so word prediction should be better for a better segmentation. However, it is usually very difficult in practice to obtain segmentations that do not break up objects, so most practitioners attempt to merge segments to get better putative object representations. We demonstrate that our paradigm of word prediction easily allows us to predict potentially useful segment merges, even for segments that do not look similar (for example, merging the black and white halves of a penguin is not possible with feature-based segmentation; the main cue must be "familiar configuration"). These studies focus on unsupervised learning of recognition. However, we show that word prediction can be markedly improved by providing supervised information for a relatively small number of regions together with large quantities of unsupervised information. This supervisory information allows a better and more discriminative choice of features and breaks possible symmetries.
Original language | English (US) |
---|---|
Pages (from-to) | II/675-II/682 |
Journal | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
Volume | 2 |
State | Published - 2003 |
Externally published | Yes |
Event | 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2003 - Madison, WI, United States Duration: Jun 18 2003 → Jun 20 2003 |
ASJC Scopus subject areas
- Software
- Computer Vision and Pattern Recognition