Show simple item record

dc.contributor.authorYing, Yiming
dc.contributor.authorHuang, Kaizhu
dc.contributor.authorCampbell, Colin
dc.date.accessioned2013-07-22T15:04:32Z
dc.date.issued2010
dc.description.abstractIn this paper we study the problem of learning a low-rank (sparse) distance matrix. We propose a novel metric learning model which can simultaneously conduct dimension reduction and learn a distance matrix. The sparse representation involves a mixed-norm regularization which is non-convex. We then show that it can be equivalently formulated as a convex saddle (min-max) problem. From this saddle representation, we develop an efficient smooth optimization approach [15] for sparse metric learning, although the learning model is based on a non-differentiable loss function. This smooth optimization approach has an optimal convergence rate of O(1=t2) for smooth problems where t is the iteration number. Finally, we run experiments to validate the effectiveness and efficiency of our sparse metric learning model on various datasets.en_GB
dc.identifier.citationAdvances in Neural Information Processing Systems 22. Proceedings of the 2009 Conference, pp. 2214-2222en_GB
dc.identifier.urihttp://hdl.handle.net/10871/11941
dc.language.isoenen_GB
dc.titleSparse metric learning via smooth optimizationen_GB
dc.typeConference paperen_GB
dc.date.available2013-07-22T15:04:32Z
dc.descriptionCopyright © 2009 NIPS Foundationen_GB
dc.description23rd Annual Conference on Advances in Neural Information Processing Systems (NIPS 2009), Vancouver, Canada, 7-10 December 2009en_GB


Files in this item

This item appears in the following Collection(s)

Show simple item record