Structured prediction by joint kernel support estimation

Discriminative techniques, such as conditional random fields (CRFs) or structure aware maximum-margin techniques (maximum margin Markov networks (M³N), structured output support vector machines (S-SVM)), are state-of-the-art in the prediction of structured data. However, to achieve good results thes...

Full description

Bibliographic Details
Main Authors: Lampert, C, Blaschko, M
Format: Journal article
Language:English
Published: Springer 2009
Subjects:
_version_ 1826296133389910016
author Lampert, C
Blaschko, M
author_facet Lampert, C
Blaschko, M
author_sort Lampert, C
collection OXFORD
description Discriminative techniques, such as conditional random fields (CRFs) or structure aware maximum-margin techniques (maximum margin Markov networks (M³N), structured output support vector machines (S-SVM)), are state-of-the-art in the prediction of structured data. However, to achieve good results these techniques require complete and reliable ground truth, which is not always available in realistic problems. Furthermore, training either CRFs or margin-based techniques is computationally costly, because the runtime of current training methods depends not only on the size of the training set but also on properties of the output space to which the training samples are assigned. We propose an alternative model for structured output prediction, <em>Joint Kernel Support Estimation</em> (JKSE), which is rather generative in nature as it relies on estimating the joint probability density of samples and labels in the training set. This makes it tolerant against incomplete or incorrect labels and also opens the possibility of learning in situations where more than one output label can be considered correct. At the same time, we avoid typical problems of generative models as we do not attempt to learn the full joint probability distribution, but we model only its support in a joint reproducing kernel Hilbert space. As a consequence, JKSE training is possible by an adaption of the classical one-class SVM procedure. The resulting optimization problem is convex and efficiently solvable even with tens of thousands of training examples. A particular advantage of JKSE is that the training speed depends only on the size of the training set, and not on the total size of the label space. No inference step during training is required (as M³N and S-SVM would) nor do we have calculate a partition function (as CRFs do). Experiments on realistic data show that, for suitable kernel functions, our method works efficiently and robustly in situations that discriminative techniques have problems with or that are computationally infeasible for them.
first_indexed 2024-03-07T04:11:39Z
format Journal article
id oxford-uuid:c806c9d4-21f3-41f2-850b-0389f32c3f7f
institution University of Oxford
language English
last_indexed 2024-03-07T04:11:39Z
publishDate 2009
publisher Springer
record_format dspace
spelling oxford-uuid:c806c9d4-21f3-41f2-850b-0389f32c3f7f2022-03-27T06:49:23ZStructured prediction by joint kernel support estimationJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:c806c9d4-21f3-41f2-850b-0389f32c3f7fRoboticsEngineering & allied sciencesEnglishOxford University Research Archive - ValetSpringer2009Lampert, CBlaschko, MDiscriminative techniques, such as conditional random fields (CRFs) or structure aware maximum-margin techniques (maximum margin Markov networks (M³N), structured output support vector machines (S-SVM)), are state-of-the-art in the prediction of structured data. However, to achieve good results these techniques require complete and reliable ground truth, which is not always available in realistic problems. Furthermore, training either CRFs or margin-based techniques is computationally costly, because the runtime of current training methods depends not only on the size of the training set but also on properties of the output space to which the training samples are assigned. We propose an alternative model for structured output prediction, <em>Joint Kernel Support Estimation</em> (JKSE), which is rather generative in nature as it relies on estimating the joint probability density of samples and labels in the training set. This makes it tolerant against incomplete or incorrect labels and also opens the possibility of learning in situations where more than one output label can be considered correct. At the same time, we avoid typical problems of generative models as we do not attempt to learn the full joint probability distribution, but we model only its support in a joint reproducing kernel Hilbert space. As a consequence, JKSE training is possible by an adaption of the classical one-class SVM procedure. The resulting optimization problem is convex and efficiently solvable even with tens of thousands of training examples. A particular advantage of JKSE is that the training speed depends only on the size of the training set, and not on the total size of the label space. No inference step during training is required (as M³N and S-SVM would) nor do we have calculate a partition function (as CRFs do). Experiments on realistic data show that, for suitable kernel functions, our method works efficiently and robustly in situations that discriminative techniques have problems with or that are computationally infeasible for them.
spellingShingle Robotics
Engineering & allied sciences
Lampert, C
Blaschko, M
Structured prediction by joint kernel support estimation
title Structured prediction by joint kernel support estimation
title_full Structured prediction by joint kernel support estimation
title_fullStr Structured prediction by joint kernel support estimation
title_full_unstemmed Structured prediction by joint kernel support estimation
title_short Structured prediction by joint kernel support estimation
title_sort structured prediction by joint kernel support estimation
topic Robotics
Engineering & allied sciences
work_keys_str_mv AT lampertc structuredpredictionbyjointkernelsupportestimation
AT blaschkom structuredpredictionbyjointkernelsupportestimation