Deep Large-Margin Rank Loss for Multi-Label Image Classification

The large-margin technique has served as the foundation of several successful theoretical and empirical results in multi-label image classification. However, most large-margin techniques are only suitable to shallow multi-label models with preset feature representations and a few large-margin techni...

Full description

Bibliographic Details
Main Authors: Zhongchen Ma, Zongpeng Li, Yongzhao Zhan
Format: Article
Language:English
Published: MDPI AG 2022-12-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/10/23/4584
Description
Summary:The large-margin technique has served as the foundation of several successful theoretical and empirical results in multi-label image classification. However, most large-margin techniques are only suitable to shallow multi-label models with preset feature representations and a few large-margin techniques of neural networks only enforce margins at the output layer, which are not well suitable for deep networks. Based on the large-margin technique, a deep large-margin rank loss function suitable for any network structure is proposed, which is able to impose a margin on any chosen set of layers of a deep network, allows choosing any <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msub><mo>ℓ</mo><mi>p</mi></msub></semantics></math></inline-formula> norm (<inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>p</mi><mo>≥</mo><mn>1</mn></mrow></semantics></math></inline-formula>) on the metric measuring the margin between labels and is applicable to any network architecture. Although the complete computation of deep large-margin rank loss function has the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi mathvariant="script">O</mi><mo>(</mo><msup><mi>C</mi><mn>2</mn></msup><mo>)</mo></mrow></semantics></math></inline-formula> time complexity, where <i>C</i> denotes the size of the label set, which would cause scalability issues when <i>C</i> is large, a negative sampling technique was proposed to make the loss function scale linearly to <i>C</i>. Experimental results on two large-scale datasets, VOC2007 and MS-COCO, show that the deep large-margin ranking function improves the robustness of the model in multi-label image classification tasks while enhancing the model’s anti-noise performance.
ISSN:2227-7390