Soft-Margin Softmax for Deep Classification

Xuezhi Liang, Xiaobo Wang, Zhen Lei, Shengcai Liao, Stan Z. Li

Research output: Chapter in Book/Report/Conference proceedingConference contribution

94 Citations (Scopus)

Abstract

In deep classification, the softmax loss (Softmax) is arguably one of the most commonly used components to train deep convolutional neural networks (CNNs). However, such a widely used loss is limited due to its lack of encouraging the discriminability of features. Recently, the large-margin softmax loss (L-Softmax [1]) is proposed to explicitly enhance the feature discrimination, with hard margin and complex forward and backward computation. In this paper, we propose a novel soft-margin softmax (SM-Softmax) loss to improve the discriminative power of features. Specifically, SM-Softamx only modifies the forward of Softmax by introducing a non-negative real number m, without changing the backward. Thus it can not only adjust the desired continuous soft margin but also be easily optimized by the typical stochastic gradient descent (SGD). Experimental results on three benchmark datasets have demonstrated the superiority of our SM-Softmax over the baseline Softmax, the alternative L-Softmax and several state-of-the-art competitors.

Original languageEnglish
Title of host publicationNeural Information Processing - 24th International Conference, ICONIP 2017, Proceedings
EditorsDongbin Zhao, El-Sayed M. El-Alfy, Derong Liu, Shengli Xie, Yuanqing Li
PublisherSpringer Verlag
Pages413-421
Number of pages9
ISBN (Print)9783319700953
DOIs
Publication statusPublished - 2017
Externally publishedYes
Event24th International Conference on Neural Information Processing, ICONIP 2017 - Guangzhou, China
Duration: Nov 14 2017Nov 18 2017

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume10635 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference24th International Conference on Neural Information Processing, ICONIP 2017
Country/TerritoryChina
CityGuangzhou
Period11/14/1711/18/17

Keywords

  • Classification
  • CNN
  • L-Softmax
  • SM-Softmax
  • Softmax

ASJC Scopus subject areas

  • Theoretical Computer Science
  • General Computer Science

Fingerprint

Dive into the research topics of 'Soft-Margin Softmax for Deep Classification'. Together they form a unique fingerprint.

Cite this