Towards Discriminability and Diversity: Batch Nuclear-Norm Maximization Under Label Insufficient Situations

The learning of the deep networks largely relies on the data with human-annotated labels. In some label insufficient situations, the performance degrades on the decision boundary with high data density. A common solution is to directly minimize the Shannon Entropy, but the side effect caused by entr...

Full description

Saved in:
Bibliographic Details
Published in:2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp. 3940 - 3949
Main Authors: Cui, Shuhao, Wang, Shuhui, Zhuo, Junbao, Li, Liang, Huang, Qingming, Tian, Qi
Format: Conference Proceeding
Language:English
Published: IEEE 01-01-2020
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The learning of the deep networks largely relies on the data with human-annotated labels. In some label insufficient situations, the performance degrades on the decision boundary with high data density. A common solution is to directly minimize the Shannon Entropy, but the side effect caused by entropy minimization, \it i.e., reduction of the prediction diversity, is mostly ignored. To address this issue, we reinvestigate the structure of classification output matrix of a randomly selected data batch. We find by theoretical analysis that the prediction discriminability and diversity could be separately measured by the Frobenius-norm and rank of the batch output matrix. Besides, the nuclear-norm is an upperbound of the Frobenius-norm, and a convex approximation of the matrix rank. Accordingly, to improve both discriminability and diversity, we propose Batch Nuclear-norm Maximization (BNM) on the output matrix. BNM could boost the learning under typical label insufficient learning scenarios, such as semi-supervised learning, domain adaptation and open domain recognition. On these tasks, extensive experimental results show that BNM outperforms competitors and works well with existing well-known methods. The code is available at https://github.com/cuishuhao/BNM
ISSN:2575-7075
DOI:10.1109/CVPR42600.2020.00400