Large Margin - Minimum Classification Error Using Sum of Shifted Sigmoids as the Loss Function

We have developed a novel loss function that embeds largemargin classification into Minimum Classification Error (MCE) training. Unlike previous efforts this approach employs a loss function that is bounded, does not require incremental adjustment of the margin or prior MCE training. It extends the Bayes risk formulation of MCE using Parzen Window estimation to incorporate large– margin classification and develops a loss function that is a sum of shifted sigmoids. Experimental results show improvement in recognition performance when evaluated on the TIDigits database.

[1]  E. McDermott,et al.  Minimum classification error via a Parzen window based estimate of the theoretical Bayes classification risk , 2002, Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing.

[2]  Biing-Hwang Juang,et al.  Minimum classification error rate methods for speech recognition , 1997, IEEE Trans. Speech Audio Process..

[3]  Jonathan Le Roux,et al.  Discriminative Training for Large-Vocabulary Speech Recognition Using Minimum Classification Error , 2007, IEEE Transactions on Audio, Speech, and Language Processing.

[4]  Li Deng,et al.  Large-Margin Discriminative Training of Hidden Markov Models for Speech Recognition , 2007 .

[5]  Biing-Hwang Juang,et al.  Multi-Class Classification Using a New Sigmoid Loss Function for Minimum Classification Error (MCE) , 2010, 2010 Ninth International Conference on Machine Learning and Applications.

[6]  Jinyu Li,et al.  Approximate Test Risk Bound Minimization Through Soft Margin Estimation , 2007, IEEE Transactions on Audio, Speech, and Language Processing.