Title :
Naive-Bayes Inspired Effective Pre-Conditioner for Speeding-Up Logistic Regression
Author :
Zaidi, Nayyar A. ; Carman, Mark J. ; Cerquides, Jesus ; Webb, Geoffrey I.
Author_Institution :
Fac. of Inf. Technol., Monash Univ., Monash, VIC, Australia
Abstract :
We propose an alternative parameterization of Logistic Regression (LR) for the categorical data, multi-class setting. LR optimizes the conditional log-likelihood over the training data and is based on an iterative optimization procedure to tune this objective function. The optimization procedure employed may be sensitive to scale and hence an effective pre-conditioning method is recommended. Many problems in machine learning involve arbitrary scales or categorical data (where simple standardization of features is not applicable). The problem can be alleviated by using optimization routines that are invariant to scale such as (second-order) Newton methods. However, computing and inverting the Hessian is a costly procedure and not feasible for big data. Thus one must often rely on first-order methods such as gradient descent (GD), stochastic gradient descent (SGD) or approximate second-order such as quasi-Newton (QN) routines, which are not invariant to scale. This paper proposes a simple yet effective pre-conditioner for speeding-up LR based on naive Bayes conditional probability estimates. The idea is to scale each attribute by the log of the conditional probability of that attribute given the class. This formulation substantially speeds-up LR´s convergence. It also provides a weighted naive Bayes formulation which yields an effective framework for hybrid generative-discriminative classification.
Keywords :
Bayes methods; Newton method; convergence; learning (artificial intelligence); optimisation; pattern classification; regression analysis; LR convergence; categorical data; conditional log-likelihood; hybrid generative-discriminative classification; iterative optimization; logistic regression; machine learning; multiclass setting; naive Bayes conditional probability estimates; naive-Bayes inspired effective preconditioner; optimization routines; parameterisation; preconditioning method; second-order Newton methods; weighted naive Bayes formulation; Convergence; Equations; Logistics; Mathematical model; Niobium; Optimization; Training; classification; discriminative-generative learning; logistic regression; pre-conditioning; stochastic gradient descent; weighted naive Bayes;
Conference_Titel :
Data Mining (ICDM), 2014 IEEE International Conference on
Conference_Location :
Shenzhen
Print_ISBN :
978-1-4799-4303-6
DOI :
10.1109/ICDM.2014.53