Regression shrinkage and selection via least quantile shrinkage and selection operator.
Over recent years, the state-of-the-art lasso and adaptive lasso have aquired remarkable consideration. Unlike the lasso technique, adaptive lasso welcomes the variables' effects in penalty meanwhile specifying adaptive weights to penalize coefficients in a different manner. However, if the ini...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2023-01-01
|
Series: | PLoS ONE |
Online Access: | https://doi.org/10.1371/journal.pone.0266267 |
_version_ | 1828005202892947456 |
---|---|
author | Alireza Daneshvar Golalizadeh Mousa |
author_facet | Alireza Daneshvar Golalizadeh Mousa |
author_sort | Alireza Daneshvar |
collection | DOAJ |
description | Over recent years, the state-of-the-art lasso and adaptive lasso have aquired remarkable consideration. Unlike the lasso technique, adaptive lasso welcomes the variables' effects in penalty meanwhile specifying adaptive weights to penalize coefficients in a different manner. However, if the initial values presumed for the coefficients are less than one, the corresponding weights would be relatively large, leading to an increase in bias. To dominate such an impediment, a new class of weighted lasso will be introduced that employs all aspects of data. That is to say, signs and magnitudes of the initial coefficients will be taken into account simultaneously for proposing appropriate weights. To allocate a particular form to the suggested penalty, the new method will be nominated as 'lqsso', standing for the least quantile shrinkage and selection operator. In this paper, we demonstate that lqsso encompasses the oracle properties under certain mild conditions and delineate an efficient algorithm for the computation purpose. Simulation studies reveal the predominance of our proposed methodology when compared with other lasso methods from various aspects, particularly in ultra high-dimensional condition. Application of the proposed method is further underlined with real-world problem based on the rat eye dataset. |
first_indexed | 2024-04-10T07:26:18Z |
format | Article |
id | doaj.art-57566c950d57438a9c257337d06e497d |
institution | Directory Open Access Journal |
issn | 1932-6203 |
language | English |
last_indexed | 2024-04-10T07:26:18Z |
publishDate | 2023-01-01 |
publisher | Public Library of Science (PLoS) |
record_format | Article |
series | PLoS ONE |
spelling | doaj.art-57566c950d57438a9c257337d06e497d2023-02-24T05:31:48ZengPublic Library of Science (PLoS)PLoS ONE1932-62032023-01-01182e026626710.1371/journal.pone.0266267Regression shrinkage and selection via least quantile shrinkage and selection operator.Alireza DaneshvarGolalizadeh MousaOver recent years, the state-of-the-art lasso and adaptive lasso have aquired remarkable consideration. Unlike the lasso technique, adaptive lasso welcomes the variables' effects in penalty meanwhile specifying adaptive weights to penalize coefficients in a different manner. However, if the initial values presumed for the coefficients are less than one, the corresponding weights would be relatively large, leading to an increase in bias. To dominate such an impediment, a new class of weighted lasso will be introduced that employs all aspects of data. That is to say, signs and magnitudes of the initial coefficients will be taken into account simultaneously for proposing appropriate weights. To allocate a particular form to the suggested penalty, the new method will be nominated as 'lqsso', standing for the least quantile shrinkage and selection operator. In this paper, we demonstate that lqsso encompasses the oracle properties under certain mild conditions and delineate an efficient algorithm for the computation purpose. Simulation studies reveal the predominance of our proposed methodology when compared with other lasso methods from various aspects, particularly in ultra high-dimensional condition. Application of the proposed method is further underlined with real-world problem based on the rat eye dataset.https://doi.org/10.1371/journal.pone.0266267 |
spellingShingle | Alireza Daneshvar Golalizadeh Mousa Regression shrinkage and selection via least quantile shrinkage and selection operator. PLoS ONE |
title | Regression shrinkage and selection via least quantile shrinkage and selection operator. |
title_full | Regression shrinkage and selection via least quantile shrinkage and selection operator. |
title_fullStr | Regression shrinkage and selection via least quantile shrinkage and selection operator. |
title_full_unstemmed | Regression shrinkage and selection via least quantile shrinkage and selection operator. |
title_short | Regression shrinkage and selection via least quantile shrinkage and selection operator. |
title_sort | regression shrinkage and selection via least quantile shrinkage and selection operator |
url | https://doi.org/10.1371/journal.pone.0266267 |
work_keys_str_mv | AT alirezadaneshvar regressionshrinkageandselectionvialeastquantileshrinkageandselectionoperator AT golalizadehmousa regressionshrinkageandselectionvialeastquantileshrinkageandselectionoperator |