Oxford-IIIT TRECVID 2010 – notebook paper

Our team participated in the “light” version of the semantic indexing task. All runs used a combination of an image-level dense visual words classifier and an object-level part based detector. For each of the ten features, these two methods were ranked based on their performance on a validation set...

Full description

Bibliographic Details
Main Authors: Juneja, M, Chandra, S, Parkhi, OM, Jawahar, CV, Vedaldi, A, Marszalek, M, Zisserman, A
Format: Conference item
Language:English
Published: National Institute of Standards and Technology 2010
_version_ 1824458964841005056
author Juneja, M
Chandra, S
Parkhi, OM
Jawahar, CV
Vedaldi, A
Marszalek, M
Zisserman, A
author_facet Juneja, M
Chandra, S
Parkhi, OM
Jawahar, CV
Vedaldi, A
Marszalek, M
Zisserman, A
author_sort Juneja, M
collection OXFORD
description Our team participated in the “light” version of the semantic indexing task. All runs used a combination of an image-level dense visual words classifier and an object-level part based detector. For each of the ten features, these two methods were ranked based on their performance on a validation set and associated to successive runs by decreasing performance (we also used a number of different techniques to recombine the scores). The two methods yielded a significantly different performance depending on the feature, as expected by their design: The χ<sup>2</sup>-SVM can be used for all feature types, including scene-like features such as Cityscape, Nighttime, Singing, but is outperformed by the object detector for object-like features, such as Boat or ship, Bus, and Person riding a bicycle. <br> Our team did not participate in the collaborative annotation effort. Instead, annotations were carried out internally for all the ten features to control quality and keyframe extraction, and to obtain region-of-interest annotations to train the object detectors. Compared to last year, the image-level classifier was significantly faster due to the use of a fast dense SIFT feature extractor and of an explicit feature map to approximate the χ<sup>2</sup> kernel SVM.
first_indexed 2025-02-19T04:34:16Z
format Conference item
id oxford-uuid:1b8e4efb-9343-4e78-8b77-0b6cbb21aff5
institution University of Oxford
language English
last_indexed 2025-02-19T04:34:16Z
publishDate 2010
publisher National Institute of Standards and Technology
record_format dspace
spelling oxford-uuid:1b8e4efb-9343-4e78-8b77-0b6cbb21aff52025-01-28T17:20:28ZOxford-IIIT TRECVID 2010 – notebook paperConference itemhttp://purl.org/coar/resource_type/c_5794uuid:1b8e4efb-9343-4e78-8b77-0b6cbb21aff5EnglishSymplectic ElementsNational Institute of Standards and Technology2010Juneja, MChandra, SParkhi, OMJawahar, CVVedaldi, AMarszalek, MZisserman, AOur team participated in the “light” version of the semantic indexing task. All runs used a combination of an image-level dense visual words classifier and an object-level part based detector. For each of the ten features, these two methods were ranked based on their performance on a validation set and associated to successive runs by decreasing performance (we also used a number of different techniques to recombine the scores). The two methods yielded a significantly different performance depending on the feature, as expected by their design: The χ<sup>2</sup>-SVM can be used for all feature types, including scene-like features such as Cityscape, Nighttime, Singing, but is outperformed by the object detector for object-like features, such as Boat or ship, Bus, and Person riding a bicycle. <br> Our team did not participate in the collaborative annotation effort. Instead, annotations were carried out internally for all the ten features to control quality and keyframe extraction, and to obtain region-of-interest annotations to train the object detectors. Compared to last year, the image-level classifier was significantly faster due to the use of a fast dense SIFT feature extractor and of an explicit feature map to approximate the χ<sup>2</sup> kernel SVM.
spellingShingle Juneja, M
Chandra, S
Parkhi, OM
Jawahar, CV
Vedaldi, A
Marszalek, M
Zisserman, A
Oxford-IIIT TRECVID 2010 – notebook paper
title Oxford-IIIT TRECVID 2010 – notebook paper
title_full Oxford-IIIT TRECVID 2010 – notebook paper
title_fullStr Oxford-IIIT TRECVID 2010 – notebook paper
title_full_unstemmed Oxford-IIIT TRECVID 2010 – notebook paper
title_short Oxford-IIIT TRECVID 2010 – notebook paper
title_sort oxford iiit trecvid 2010 notebook paper
work_keys_str_mv AT junejam oxfordiiittrecvid2010notebookpaper
AT chandras oxfordiiittrecvid2010notebookpaper
AT parkhiom oxfordiiittrecvid2010notebookpaper
AT jawaharcv oxfordiiittrecvid2010notebookpaper
AT vedaldia oxfordiiittrecvid2010notebookpaper
AT marszalekm oxfordiiittrecvid2010notebookpaper
AT zissermana oxfordiiittrecvid2010notebookpaper