TY - JOUR
T1 - Non-uniform stochastic average gradient method for training conditional random fields
AU - Schmidt, Mark
AU - Babanezhad, Reza
AU - Ahemd, M. Osama
AU - Defazio, Aaron
AU - Clifton, Ann
AU - Sarkar, Anoop
N1 - Publisher Copyright:
Copyright 2015 by the authors.
PY - 2015
Y1 - 2015
N2 - We apply stochastic average gradient (SAG) algorithms for training conditional random fields (CRFs). We describe a practical im-plementation that uses structure in the CRF gradient to reduce the memory requirement of this linearly-convergent stochastic gradi-ent method, propose a non-uniform sampling scheme that substantially improves practical performance, and analyze the rate of con-vergence of the SAGA variant under non-uniform sampling. Our experimental results reveal that our method significantly outper-forms existing methods in terms of the training objective, and performs as well or bet-ter than optimally-tuned stochastic gradient methods in terms of test error.
AB - We apply stochastic average gradient (SAG) algorithms for training conditional random fields (CRFs). We describe a practical im-plementation that uses structure in the CRF gradient to reduce the memory requirement of this linearly-convergent stochastic gradi-ent method, propose a non-uniform sampling scheme that substantially improves practical performance, and analyze the rate of con-vergence of the SAGA variant under non-uniform sampling. Our experimental results reveal that our method significantly outper-forms existing methods in terms of the training objective, and performs as well or bet-ter than optimally-tuned stochastic gradient methods in terms of test error.
UR - http://www.scopus.com/inward/record.url?scp=84954318065&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:84954318065
SN - 1532-4435
VL - 38
SP - 819
EP - 828
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
T2 - 18th International Conference on Artificial Intelligence and Statistics, AISTATS 2015
Y2 - 9 May 2015 through 12 May 2015
ER -