A tutorial on the lasso approach to sparse modeling

    65 Citations (Scopus)

    Abstract

    In applied research data are often collected from sources with a high dimensional multivariate output. Analysis of such data is composed of e.g. extraction and characterization of underlying patterns, and often with the aim of finding a small subset of significant variables or features. Variable and feature selection is well-established in the area of regression, whereas for other types of models this seems more difficult. Penalization of the L 1 norm provides an interesting avenue for such a problem, as it produces a sparse solution and hence embeds variable selection. In this paper a brief introduction to the mathematical properties of using the L 1 norm as a penalty is given. Examples of models extended with L 1 norm penalties/constraints are presented. The examples include PCA modeling with sparse loadings which enhance interpretability of single components. Sparse inverse covariance matrix estimation is used to unravel which variables are affecting each other, and a modified PCA to model data with (piecewise) constant responses in e.g. process monitoring is shown. All examples are demonstrated on real or synthetic data. The results indicate that sparse solutions, when appropriate, can enhance model interpretability.

    Original languageEnglish
    JournalChemometrics and Intelligent Laboratory Systems
    Volume119
    Pages (from-to)21-31
    Number of pages11
    ISSN0169-7439
    DOIs
    Publication statusPublished - 1 Oct 2012

    Fingerprint

    Dive into the research topics of 'A tutorial on the lasso approach to sparse modeling'. Together they form a unique fingerprint.

    Cite this