Title :
Tikhonov or Lasso Regularization: Which Is Better and When
Author :
Fei Wang ; Wei Liu ; Chawla, Sanjay
Author_Institution :
Sch. of IT, Univ. of Sydney, Sydney, NSW, Australia
Abstract :
It is well known that supervised learning problems with ℓ1 (Lasso) and ℓ2 (Tikhonov or Ridge) regularizers will result in very different solutions. For example, the ℓ1 solution vector will be sparser and can potentially be used both for prediction and feature selection. However, given a data set it is often hard to determine which form of regularization is more applicable in a given context. In this paper we use mathematical properties of the two regularization methods followed by detailed experimentation to understand their impact based on four characteristics: non-stationarity of the data generating process, level of noise in the data sensing mechanism, degree of correlation between dependent and independent variables and the shape of the data set. The practical outcome of our research is that it can serve as a guide for practitioners of large scale data mining and machine learning tools in their day-to-day practice.
Keywords :
correlation methods; data mining; feature selection; learning (artificial intelligence); ℓ1 solution vector; Lasso regularization; Lasso regularizers; Ridge regularizers; Tikhonov regularization; Tikhonov regularizers; correlation degree; data generating process; data sensing mechanism; data set shape; feature selection; independent variables; large scale data mining; machine learning tools; mathematical properties; noise level; nonstationarity characteristics; prediction; regularization methods; supervised learning problems; Accuracy; Correlation; Electronic mail; Mathematical model; Robustness; Stability analysis; Training data; Classification; Lasso; Regularization;
Conference_Titel :
Tools with Artificial Intelligence (ICTAI), 2013 IEEE 25th International Conference on
Conference_Location :
Herndon, VA
Print_ISBN :
978-1-4799-2971-9
DOI :
10.1109/ICTAI.2013.122