Non-uniform stochastic average gradient method for training conditional random fields

Mark Schmidt, Reza Babanezhad, M. Osama Ahemd, Aaron Defazio, Ann Clifton, Anoop Sarkar

    Research output: Contribution to journalConference articlepeer-review

    28 Citations (Scopus)

    Abstract

    We apply stochastic average gradient (SAG) algorithms for training conditional random fields (CRFs). We describe a practical im-plementation that uses structure in the CRF gradient to reduce the memory requirement of this linearly-convergent stochastic gradi-ent method, propose a non-uniform sampling scheme that substantially improves practical performance, and analyze the rate of con-vergence of the SAGA variant under non-uniform sampling. Our experimental results reveal that our method significantly outper-forms existing methods in terms of the training objective, and performs as well or bet-ter than optimally-tuned stochastic gradient methods in terms of test error.

    Original languageEnglish
    Pages (from-to)819-828
    Number of pages10
    JournalJournal of Machine Learning Research
    Volume38
    Publication statusPublished - 2015
    Event18th International Conference on Artificial Intelligence and Statistics, AISTATS 2015 - San Diego, United States
    Duration: 9 May 201512 May 2015

    Fingerprint

    Dive into the research topics of 'Non-uniform stochastic average gradient method for training conditional random fields'. Together they form a unique fingerprint.

    Cite this