IEEE Transactions on Pattern Analysis and Machine Intelligence | 2019

Truncated Cauchy Non-Negative Matrix Factorization

 
 
 
 
 

Abstract


Non-negative matrix factorization (NMF) minimizes the euclidean distance between the data matrix and its low rank approximation, and it fails when applied to corrupted data because the loss function is sensitive to outliers. In this paper, we propose a Truncated CauchyNMF loss that handle outliers by truncating large errors, and develop a Truncated CauchyNMF to robustly learn the subspace on noisy datasets contaminated by outliers. We theoretically analyze the robustness of Truncated CauchyNMF comparing with the competing models and theoretically prove that Truncated CauchyNMF has a generalization bound which converges at a rate of order <inline-formula><tex-math notation= LaTeX >$O(\\sqrt{{\\ln n}/{n}})$</tex-math><alternatives><inline-graphic xlink:href= guan-ieq1-2777841.gif /></alternatives></inline-formula> , where <inline-formula><tex-math notation= LaTeX >$n$</tex-math><alternatives> <inline-graphic xlink:href= guan-ieq2-2777841.gif /></alternatives></inline-formula> is the sample size. We evaluate Truncated CauchyNMF by image clustering on both simulated and real datasets. The experimental results on the datasets containing gross corruptions validate the effectiveness and robustness of Truncated CauchyNMF for learning robust subspaces.

Volume 41
Pages 246-259
DOI 10.1109/TPAMI.2017.2777841
Language English
Journal IEEE Transactions on Pattern Analysis and Machine Intelligence

Full Text