Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD)
Gespeichert in:
Verfasser / Beitragende:
[Qi Qian, Rong Jin, Jinfeng Yi, Lijun Zhang, Shenghuo Zhu]
Ort, Verlag, Jahr:
2015
Enthalten in:
Machine Learning, 99/3(2015-06-01), 353-372
Format:
Artikel (online)
Online Zugang:
| LEADER | caa a22 4500 | ||
|---|---|---|---|
| 001 | 605478473 | ||
| 003 | CHVBK | ||
| 005 | 20210128100405.0 | ||
| 007 | cr unu---uuuuu | ||
| 008 | 210128e20150601xx s 000 0 eng | ||
| 024 | 7 | 0 | |a 10.1007/s10994-014-5456-x |2 doi |
| 035 | |a (NATIONALLICENCE)springer-10.1007/s10994-014-5456-x | ||
| 245 | 0 | 0 | |a Efficient distance metric learning by adaptive sampling and mini-batch stochastic gradient descent (SGD) |h [Elektronische Daten] |c [Qi Qian, Rong Jin, Jinfeng Yi, Lijun Zhang, Shenghuo Zhu] |
| 520 | 3 | |a Distance metric learning (DML) is an important task that has found applications in many domains. The high computational cost of DML arises from the large number of variables to be determined and the constraint that a distance metric has to be a positive semi-definite (PSD) matrix. Although stochastic gradient descent (SGD) has been successfully applied to improve the efficiency of DML, it can still be computationally expensive in order to ensure that the solution is a PSD matrix. It has to, at every iteration, project the updated distance metric onto the PSD cone, an expensive operation. We address this challenge by developing two strategies within SGD, i.e. mini-batch and adaptive sampling, to effectively reduce the number of updates (i.e. projections onto the PSD cone) in SGD. We also develop hybrid approaches that combine the strength of adaptive sampling with that of mini-batch online learning techniques to further improve the computational efficiency of SGD for DML. We prove the theoretical guarantees for both adaptive sampling and mini-batch based approaches for DML. We also conduct an extensive empirical study to verify the effectiveness of the proposed algorithms for DML. | |
| 540 | |a The Author(s), 2014 | ||
| 690 | 7 | |a Loss Function |2 nationallicence | |
| 690 | 7 | |a Hybrid Approach |2 nationallicence | |
| 690 | 7 | |a Adaptive Sampling |2 nationallicence | |
| 690 | 7 | |a Pairwise Constraint |2 nationallicence | |
| 690 | 7 | |a Stochastic Gradient Descent |2 nationallicence | |
| 700 | 1 | |a Qian |D Qi |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | |
| 700 | 1 | |a Jin |D Rong |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | |
| 700 | 1 | |a Yi |D Jinfeng |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | |
| 700 | 1 | |a Zhang |D Lijun |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | |
| 700 | 1 | |a Zhu |D Shenghuo |u NEC Laboratories America, 95014, Cupertino, CA, USA |4 aut | |
| 773 | 0 | |t Machine Learning |d Springer US; http://www.springer-ny.com |g 99/3(2015-06-01), 353-372 |x 0885-6125 |q 99:3<353 |1 2015 |2 99 |o 10994 | |
| 856 | 4 | 0 | |u https://doi.org/10.1007/s10994-014-5456-x |q text/html |z Onlinezugriff via DOI |
| 898 | |a BK010053 |b XK010053 |c XK010000 | ||
| 900 | 7 | |a Metadata rights reserved |b Springer special CC-BY-NC licence |2 nationallicence | |
| 908 | |D 1 |a research-article |2 jats | ||
| 949 | |B NATIONALLICENCE |F NATIONALLICENCE |b NL-springer | ||
| 950 | |B NATIONALLICENCE |P 856 |E 40 |u https://doi.org/10.1007/s10994-014-5456-x |q text/html |z Onlinezugriff via DOI | ||
| 950 | |B NATIONALLICENCE |P 700 |E 1- |a Qian |D Qi |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 700 |E 1- |a Jin |D Rong |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 700 |E 1- |a Yi |D Jinfeng |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 700 |E 1- |a Zhang |D Lijun |u Department of Computer Science and Engineering, Michigan State University, 48824, East Lansing, MI, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 700 |E 1- |a Zhu |D Shenghuo |u NEC Laboratories America, 95014, Cupertino, CA, USA |4 aut | ||
| 950 | |B NATIONALLICENCE |P 773 |E 0- |t Machine Learning |d Springer US; http://www.springer-ny.com |g 99/3(2015-06-01), 353-372 |x 0885-6125 |q 99:3<353 |1 2015 |2 99 |o 10994 | ||