Scene generative models for adaptive video fast forward

Nebojsa Jojic, Nemanja Petrovic, Thomas S. Huang

Research output: Contribution to conferencePaperpeer-review


In this paper, we present a statistical generative model of scenes with multiple objects that can be efficiently used for tasks related to video search, browsing and retrieval. Instead of using a combination of weighted Euclidean distances as a shot similarity measure, we base the retrieval process on the likelihood of a video frame under the generative model trained on the query sequence. This allows for automatic separation and balancing of various causes of variability, such as occlusion, appearance change and motion. In previous work, this usually required complex user intervention. The likelihood models we study in this paper are based on appearances of multiple, possibly occluding objects in a video clip. Given a query, the video is played at a higher rate until similar frames are found. The playback speed is linked to the frame likelihood, so that the speed drops as the likely target frames are starting to come in.

Original languageEnglish (US)
Number of pages4
StatePublished - 2003
EventProceedings: 2003 International Conference on Image Processing, ICIP-2003 - Barcelona, Spain
Duration: Sep 14 2003Sep 17 2003


OtherProceedings: 2003 International Conference on Image Processing, ICIP-2003

ASJC Scopus subject areas

  • Hardware and Architecture
  • Computer Vision and Pattern Recognition
  • Electrical and Electronic Engineering


Dive into the research topics of 'Scene generative models for adaptive video fast forward'. Together they form a unique fingerprint.

Cite this