Video Google: efficient visual search of videos

We describe an approach to object retrieval which searches for and localizes all the occurrences of an object in a video, given a query image of the object. The object is represented by a set of viewpoint invariant region descriptors so that recognition can proceed successfully despite changes in vi...

Ամբողջական նկարագրություն

Մատենագիտական մանրամասներ
Հիմնական հեղինակներ: Sivic, J, Zisserman, A
Ձևաչափ: Book section
Լեզու:English
Հրապարակվել է: Springer 2007
_version_ 1826317070519762944
author Sivic, J
Zisserman, A
author_facet Sivic, J
Zisserman, A
author_sort Sivic, J
collection OXFORD
description We describe an approach to object retrieval which searches for and localizes all the occurrences of an object in a video, given a query image of the object. The object is represented by a set of viewpoint invariant region descriptors so that recognition can proceed successfully despite changes in viewpoint, illumination and partial occlusion. The temporal continuity of the video within a shot is used to track the regions in order to reject those that are unstable. <br> Efficient retrieval is achieved by employing methods from statistical text retrieval, including inverted file systems, and text and document frequency weightings. This requires a visual analogy of a word which is provided here by vector quantizing the region descriptors. The final ranking also depends on the spatial layout of the regions. The result is that retrieval is immediate, returning a ranked list of shots in the manner of Google. <br> We report results for object retrieval on the full length feature films ‘Groundhog Day’ and ‘Casablanca’.
first_indexed 2025-02-19T04:34:19Z
format Book section
id oxford-uuid:25e756f7-c6bc-4fb1-8f6c-960d420bc78a
institution University of Oxford
language English
last_indexed 2025-02-19T04:34:19Z
publishDate 2007
publisher Springer
record_format dspace
spelling oxford-uuid:25e756f7-c6bc-4fb1-8f6c-960d420bc78a2025-01-23T14:52:35ZVideo Google: efficient visual search of videosBook sectionhttp://purl.org/coar/resource_type/c_3248uuid:25e756f7-c6bc-4fb1-8f6c-960d420bc78aEnglishSymplectic ElementsSpringer2007Sivic, JZisserman, AWe describe an approach to object retrieval which searches for and localizes all the occurrences of an object in a video, given a query image of the object. The object is represented by a set of viewpoint invariant region descriptors so that recognition can proceed successfully despite changes in viewpoint, illumination and partial occlusion. The temporal continuity of the video within a shot is used to track the regions in order to reject those that are unstable. <br> Efficient retrieval is achieved by employing methods from statistical text retrieval, including inverted file systems, and text and document frequency weightings. This requires a visual analogy of a word which is provided here by vector quantizing the region descriptors. The final ranking also depends on the spatial layout of the regions. The result is that retrieval is immediate, returning a ranked list of shots in the manner of Google. <br> We report results for object retrieval on the full length feature films ‘Groundhog Day’ and ‘Casablanca’.
spellingShingle Sivic, J
Zisserman, A
Video Google: efficient visual search of videos
title Video Google: efficient visual search of videos
title_full Video Google: efficient visual search of videos
title_fullStr Video Google: efficient visual search of videos
title_full_unstemmed Video Google: efficient visual search of videos
title_short Video Google: efficient visual search of videos
title_sort video google efficient visual search of videos
work_keys_str_mv AT sivicj videogoogleefficientvisualsearchofvideos
AT zissermana videogoogleefficientvisualsearchofvideos