DocumentCode :
3105330
Title :
Accelerating Newton Optimization for Log-Linear Models through Feature Redundancy
Author :
Mathur, Arpit ; Chakrabarti, Soumen
Author_Institution :
IIT Bombay, Mumbai
fYear :
2006
fDate :
18-22 Dec. 2006
Firstpage :
404
Lastpage :
413
Abstract :
Log-linear models are widely used for labeling feature vectors and graphical models, typically to estimate robust conditional distributions in presence of a large number of potentially redundant features. Limited-memory quasi-Newton methods like LBFGS or BLMVM are optimization workhorses for such applications, and most of the training time is spent computing the objective and gradient for the optimizer. We propose a simple technique to speed up the training optimization by clustering features dynamically, and interleaving the standard optimizer with another, coarse-grained, faster optimizer that uses far fewer variables. Experiments with logistic regression training for text classification and conditional random field (CRF) training for information extraction show promising speed-ups between 2times and 9times without any systematic or significant degradation in the quality of the estimated models.
Keywords :
Newton method; feature extraction; gradient methods; optimisation; pattern classification; regression analysis; Newton optimization; conditional random field; feature clustering; feature redundancy; feature vectors; graphical models; information extraction; limited-memory quasiNewton method; log-linear model; logistic regression training; optimizer gradient; text classification; Acceleration; Data mining; Degradation; Graphical models; Interleaved codes; Labeling; Logistics; Optimization methods; Robustness; Text categorization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Data Mining, 2006. ICDM '06. Sixth International Conference on
Conference_Location :
Hong Kong
ISSN :
1550-4786
Print_ISBN :
0-7695-2701-7
Type :
conf
DOI :
10.1109/ICDM.2006.11
Filename :
4053067
Link To Document :
بازگشت