University of Cape Coast Institutional Repository

Least squares optimization with L1-norm regularization

Show simple item record

dc.contributor.author Nkansah, Henrietta
dc.date.accessioned 2018-07-16T13:35:09Z
dc.date.available 2018-07-16T13:35:09Z
dc.date.issued 2017-10
dc.identifier.issn 23105496
dc.identifier.uri http://hdl.handle.net/123456789/3369
dc.description xv, 240p.: ill. en_US
dc.description.abstract The non-differentiable L1-norm penalty in the L1-norm regularized least squares problem poses a major challenge to obtaining an analytic solution. The study thus explores smoothing and non-smoothing approximations that yields differentiable loss functional that ensures a close-form solution in over-determined systems. Three smoothing approximations to the L1-norm penalty have been examined. These include the Quadratic, Sigmoid and Cubic Hermite. Tikhonov regularization is then applied to the resulting loss function. The approximations are a modification of the Lee’s approximation to the L1-norm term. The regularized solution using this approximation has been presented in various forms. Using the Hilbert 12×12 matrix, it is found that for all three methods, a good approximation to the exact solution converges at a regularization parameter µ = 10−30. The solutions show an accuracy to nine digits. In each approximation, a suitable value of the parameter is obtained for which the absolute value function approximates the L1-norm penalty. The results of the Modified Newton’s method based on the Lee’s approximation however shows an accuracy of at most two digits. The solution of the smoothing methods also compares favourably with l1 ls method. Analytic solution of the L1-norm problem is also obtained by means of the sub-gradient method, after casting the constrained formulation as unconstrained. Attempt at achieving sparsity of the Least Absolute Shrinkage and Selection Operator (LASSO) solution has been made in two ways. The initial solution is expressed in terms of the singular value decomposition so that by truncating smaller singular values,the desired sparsity is achieved using suitable regularization parameter obtained by the K-fold cross validation of the fit. In another way,the LASSO solution itself has been induced to ensure sparsity. The results show that the LASSO formulation and solution must be appropriately designed for certain type of data sets, particularly those that are severely ill-conditioned and those with monotone trends. en_US
dc.language.iso en en_US
dc.publisher University of Cape Coast en_US
dc.subject Smoothing Approximations en_US
dc.subject Sparsity Induced Systems en_US
dc.subject Least Squares Optimization en_US
dc.subject L2-Norm Regularization en_US
dc.subject L1-Norm Regularization en_US
dc.subject Non-Smoothing Approximations en_US
dc.title Least squares optimization with L1-norm regularization en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search UCC IR


Advanced Search

Browse

My Account