RUHUNA JOURNAL OF SCIENCE 
Vol 13 (1): 14-28, June 2022 
eISSN: 2536-8400                                                 Faculty of Science 
http://doi.org/10.4038/rjs.v13i1.112                                                                                   University of Ruhuna 

Faculty of Science, University of Ruhuna    

Sri Lanka 
14 

Generalized Stochastic Restricted LARS Algorithm  

Manickavasagar Kayanan1,2 and Pushpakanthie Wijekoon3  

1Postgraduate Institute of Science, University of Peradeniya, Peradeniya, Sri Lanka 
2 Department of Physical Science, University of Vavuniya, Vavuniya, Sri Lanka 
3 Department of Statistics and Computer Science, University of Peradeniya, Peradeniya, Sri Lanka 

 

*Correspondence: mgayanan@vau.jfn.ac.lk;  ORCID: https://orcid.org/0000-0003-2662-4383 

 

Received: 29th July 2021; Revised: 21st April 2022; Accepted: 17th May 2022 

Abstract The Least Absolute Shrinkage and Selection Operator (LASSO) is used 

to tackle both the multicollinearity issue and the variable selection concurrently 

in the linear regression model. The Least Angle Regression (LARS) algorithm has 

been used widely to produce LASSO solutions. However, this algorithm is 

unreliable when high multicollinearity exists among regressor variables. One 

solution to improve the estimation of regression parameters when 

multicollinearity exists is adding preliminary information about the regression 

coefficient to the model as either exact linear restrictions or stochastic linear 

restrictions. Based on this solution, this article proposed a generalized version of 

the stochastic restricted LARS algorithm, which combines LASSO with existing 

stochastic restricted estimators. Further, we examined the performance of the 

proposed algorithm by employing a Monte Carlo simulation study and a 

numerical example. 

Keywords: LASSO, LARS, Stochastic Linear Restrictions.  

1   Introduction 

The biased estimators such as Ridge Estimator (RE) (Hoerl and Kennard 1970), 

Almost Unbiased Ridge Estimator (AURE) (Singh et al. 1986), Liu Estimator (LE) 

(Liu 1993), Almost Unbiased Liu Estimator (AULE) (Akdeniz and Kaรงiranlar 1995), 

Principle Component Regression Estimator (PCRE) (Massy 1965), r-k class estimator 

(Baye and Parker 1984), r-d class estimator (Kaรงiranlar and SakallฤฑoฤŸlu 2001) and 

Sample Information Optimal Estimator (SIOE) (Kayanan and Wijekoon 2019) have 

been widely used in literature to resolve multicollinearity issue in the linear regression 

model. However, these estimators yield high bias when the number of explanatory 

variables is high, and they do not consider about irrelevant variables while fitting 

models. For high dimensional data, having many variables in the model and 

multicollinearity are major issues. To tackle these matters, Tibshirani (1996) 

introduced Least Absolute Shrinkage and Selection Operator (LASSO). The LASSO 

https://rjs.ruh.ac.lk/index.php/rjs/index
https://creativecommons.org/licenses/by-nc/4.0/
mailto:mgayanan@vau.jfn.ac.lk
https://orcid.org/0000-0003-2662-4383
https://orcid.org/0000-0003-2662-4383


 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
15 

is a shrinkage method that was originally used for regularization and variable selection 

in the linear regression model. The Least Angle Regression (LARS) (Efron et al. 2004) 

algorithm has been used to obtain the estimates of LASSO. Zou and Hastie (2005) have 

shown that the LASSO is unsteady when severe multicollinearity exists between the 

explanatory variables. Therefore, they suggested Elastic Net (ENet) estimator by 

combining LASSO and RE as a solution for this issue. Furthermore, they proposed 

LARS-EN algorithm to attain ENet solutions, which is a modified version of the LARS 

algorithm. Besides, Kayanan and Wijekoon (2020b) proposed a generalized version of 

LARS (GLARS) algorithm to combine LASSO with RE and the other biased 

estimators based on sample information such as AURE, LE, AULE, PCRE, r-k class 

estimator, and r-d class estimator. Finally, they have shown that the GLARS algorithm 

performs well when it combines LASSO with r-k class and r-d class estimators. 

According to literature, the parameter estimation can be strengthened if prior 

knowledge about the regression coefficient is applied. The prior information on 

regression coefficients can be defined in the form of exact linear restrictions or 

stochastic linear restrictions. Many researchers proposed stochastic restricted 

estimators such as Mixed Regression Estimator (MRE) (Theil and Goldberger 1961) 

Stochastic Restricted Ridge Estimator (SRRE) (Li and Yang 2010), Stochastic 

Restricted Almost Unbiased Ridge Estimator (SRAURE) (Jibo and Hu 2014), 

Stochastic Restricted Liu Estimator (SRLE) (Hubert and Wijekoon 2006), Stochastic 

Restricted Almost Unbiased Liu Estimator (SRAULE) (Jibo and Hu 2014), Stochastic 

Restricted Principle Component Regression Estimator (SRPCRE) (He and Wu 2014), 

Stochastic Restricted r-k class estimator (SRrk) (Jibo 2014), Stochastic Restricted r-d 

class estimator (SRrd) (Jibo 2014), and Stochastic Restricted Optimal Estimator 

(SROE) (Kayanan and Wijekoon 2019) to incorporate prior information to the 

regression coefficient. Stochastic restricted estimators also have the same issue as 

biased estimators when the linear regression model contains numerous predictors. To 

handle this problem, Kayanan and Wijekoon (2020a) proposed a stochastic restricted 

LARS (SRLARS) algorithm to combine LASSO and MRE, and showed the superiority 

of the SRLARS over LARS algorithm. 

This article proposes a generalized version of the stochastic restricted LARS 

algorithm, namely SRGLARS, to combine LASSO with other stochastic restricted 

estimators. The prediction performance of the SRGLARS algorithm was examined by 

employing a Monte-Carlo simulation and using a real-world example in the Root Mean 

Square Error (RMSE) criterion.  

2 Model Specification and the Estimators  

Consider the linear regression model   

๐’š = ๐‘ฟ๐œท + ๐œบ,     (2.1) 
where ๐‘ฟ is ๐‘› ร— ๐‘ matrix of explanatory variables, ๐œท be the ๐‘ ร— 1 vector of unknown 

coefficients, and ๐œบ be the ๐‘› ร— 1 vector of disturbances such that ๐œบ โˆผ ๐‘ (0,๐œŽ2 ๐‘ฐ). 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
16 

Assume that there exists prior information on ฮฒ, which may be expressed as a 

stochastic linear restriction, as (Theil & Goldberger, 1961)  

๐‹ =  ๐‘น๐œท +  ๐’—,    (2.2) 
where ฯ† be the ๐‘ž ร— 1 vector, ๐‘น be the ๐‘ž ร— ๐‘ matrix with rank ๐‘ž, ๐’— be the ๐‘ž ร— 1 

vector of disturbances, such that ๐’— โˆผ ๐‘ (0,๐œŽ2๐‘พ ), ๐‘พ is positive definite, and 
๐ธ(๐’—๐œบโ€ฒ) =  0. 

To make the variable selection and handle multicollinearity issue by incorporating 

prior information defined in model (2.2), Kayanan and Wijekoon (2020a) suggested 

Stochastic Restricted LASSO Type Estimator (SRLASSO) for model (2.1) as 

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐ฟ๐ด๐‘†๐‘†๐‘‚ = argmin
๐œท

{(๐’š โˆ’ ๐‘ฟ๐œท)โ€ฒ(๐’š โˆ’ ๐‘ฟ๐œท)} 

subject to โˆ‘ |๐›ฝ๐‘—|
๐‘
๐‘—=1 โ‰ค ๐‘ก and ๐‘น๐œท = ๐‹ โˆ’ ๐’—,                                              (2.3) 

where ๐‘ก > 0 is a turning parameter. 
 

Further, Kayanan and Wijekoon (2020a) proposed Stochastic Restricted LARS 

(SRLARS) algorithm to find the SRLASSO estimates. Note that SRLARS combines 

LASSO and MRE to find the estimates. To improve the SRLARS solutions, this article 

proposes a generalized version of SRLARS (SRGLARS) to combine LASSO with 

other stochastic restricted estimators such as SRRE, SRAURE, SRLE, SRAULE, 

SRPCRE, SRrk, SRrd and SROE. 

Kayanan and Wijekoon (2018, 2019) proposed a generalized form to express the 

estimators MRE, SRRE, SRAURE, SRLE, SRAULE, SRPCRE, SRrk, SRrd, and 

SROE as 

๏ฟฝฬ‚๏ฟฝ๐บ = ๐‘ฎ(๐‘ฟ
โ€ฒ๐‘ฟ + ๐‘นโ€ฒ๐‘พโˆ’๐Ÿ๐‘น)โˆ’๐Ÿ (๐‘ฟโ€ฒ๐’š + ๐‘นโ€ฒ๐‘พโˆ’๐Ÿ๐‹), (2.4) 

 where  

๏ฟฝฬ‚๏ฟฝ๐บ =

{
 
 
 
 
 

 
 
 
 
 
๏ฟฝฬ‚๏ฟฝ๐‘€๐‘…๐ธ            if   ๐‘ฎ = ๐‘ฐ                                                 

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐‘…๐ธ           if   ๐‘ฎ = (๐‘ฟ
โ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)โˆ’1๐‘ฟโ€ฒ๐‘ฟ                   

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐ด๐‘ˆ๐‘…๐ธ       if   ๐‘ฎ = (๐‘ฐ โˆ’ ๐‘˜
2(๐‘ฟโ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)โˆ’2)          

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐ฟ๐ธ            if   ๐‘ฎ = (๐‘ฟ
โ€ฒ๐‘ฟ + ๐‘ฐ)โˆ’1(๐‘ฟโ€ฒ๐‘ฟ + ๐‘‘๐‘ฐ)        

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐ด๐‘ˆ๐ฟ๐ธ        if   ๐‘ฎ = (๐‘ฐ โˆ’ (1 โˆ’ ๐‘‘)
2(๐‘ฟโ€ฒ๐‘ฟ + ๐‘ฐ)โˆ’2)  

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐‘ƒ๐ถ๐‘…๐ธ        if   ๐‘ฎ = ๐‘ปโ„Ž๐‘ปโ„Ž
โ€ฒ                                             

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐‘Ÿ๐‘˜            if   ๐‘ฎ = ๐‘ปโ„Ž๐‘ปโ„Ž
โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)โˆ’1๐‘ฟโ€ฒ๐‘ฟ          

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐‘Ÿ๐‘‘            if   ๐‘ฎ = ๐‘ปโ„Ž๐‘ปโ„Ž
โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘ฐ)โˆ’1(๐‘ฟโ€ฒ๐‘ฟ + ๐‘‘๐‘ฐ)

๏ฟฝฬ‚๏ฟฝ๐‘†๐‘…๐‘‚๐ธ            if   ๐‘ฎ = ๐œทโˆ—๐œทโˆ—
โ€ฒ(๐œŽ2(๐‘ฟโ€ฒ๐‘ฟ)โˆ’๐Ÿ + ๐œทโˆ—๐œทโˆ—

โ€ฒ)
โˆ’1
 

 

 

Note that ๐‘˜ > 0 and 0 < ๐‘‘ < 1 are the shrinkage/regularization parameters, ๐‘ฐ is the 
๐‘ ร— ๐‘ identity matrix, ๐‘ปโ„Ž = (๐‘ก1, ๐‘ก2. . . ๐‘กโ„Ž)โ€ฒ is the first โ„Ž columns of the eigenvectors of 
๐‘ฟโ€ฒ๐‘ฟ, and ๐œทโˆ—   is the normalized eigenvector corresponding to the largest eigenvalue of 
๐‘ฟโ€ฒ๐‘ฟ. 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
17 

3   SRGLARS Algorithm 

Based on Kayanan and Wijekoon (2021a) and equation (2.4), the SRGLARS algorithm 

for model (2.1) is outlined below: 

  Algorithm 1: SRGLARS  

1: Standardize ๐‘ฟ to have a mean zero with a standard deviation of one, and center the 
๐’š to have a mean zero.   

2: Start with all estimates of the coefficients ๏ฟฝฬ‚๏ฟฝ = 0 with the residuals ๐’“ = ๏ฟฝฬ‚๏ฟฝ and ๐‰ =
๏ฟฝฬ‚๏ฟฝ. 

3: Find the predictor ๐‘ฟ๐’‹ most correlated with ๐’“; ๐‘— =  1,2, . . . ,๐‘. 

4: Move the estimate of ๏ฟฝฬ‚๏ฟฝ๐‘— from 0 towards the ๏ฟฝฬ‚๏ฟฝ๐บ direction until some other predictor 

๐‘ฟ๐’Œ has as large a correlation with the current residual as ๐‘ฟ๐’‹ does. 

5: Move ๏ฟฝฬ‚๏ฟฝ๐‘— and ๏ฟฝฬ‚๏ฟฝ๐‘˜ in the direction defined by their joint ๏ฟฝฬ‚๏ฟฝ๐บ direction of the current 

residual on (๐‘ฟ๐’‹,๐‘ฟ๐’Œ), until some other predictor ๐‘ฟ๐’ eventually earns its way into the 

active set. 

6: If a non-zero coefficient hits zero, drop its variable from the active set of variables 

and recomputed the current joint ๏ฟฝฬ‚๏ฟฝ๐บ direction. 
7: Repeat the steps 5 and 6 until SRGLARS conditions attained. 

 

The mathematical details of the SRLARS algorithm are as follows: 

Let us assume that the estimates of the coefficients ๏ฟฝฬ‚๏ฟฝ and residuals ๐’“ and ๐‰ are 

(๏ฟฝฬ‚๏ฟฝ)
0
= [(๐›ฝ1)0, (๐›ฝ2)0,โ€ฆ,(๐›ฝ๐‘)0

]
โ€ฒ
= ๐ŸŽ, ๐’“0  =  ๐’š and ๐‰0   =  ๐‹. 

Find the predictor (๐‘‹๐‘—)1 most correlated with ๐’“0.  

(๐‘‹๐‘—)1 = max
๐‘—
|๐ถ๐‘œ๐‘Ÿ(๐‘‹๐‘—,๐’“0)|; ๐‘— โˆˆ (1,2,โ€ฆ,๐‘)    (3.1) 

Then, increase the estimate of respective regression coefficient (๐›ฝ๐‘—)1
 from 0 unto 

any other predictor (๐‘‹๐‘—)2 has a high correlation with ๐’“1 as (๐‘‹๐‘—)1 does. At this stage, 

SRGLARS moves in the equiangular direction between (๐‘‹๐‘—)1 and (๐‘‹๐‘—)2 rather than 

proceeding the path based on (๐‘‹๐‘—)1.  

Similarly, in ๐‘–๐‘กโ„Ž run, the variable (๐‘‹๐‘—)1 eventually acquires its path in the active set, 

and then SRGLARS moves in the equiangular direction among (๐‘ฟ)๐‘– =

[(๐‘‹๐‘—)1
, (๐‘‹๐‘—)2

,โ€ฆ,(๐‘‹๐‘—)๐‘–
]
โ€ฒ
.  Proceed to add variables to the active set in this way, running 

in the path established by the least angle direction. During this process, (๏ฟฝฬ‚๏ฟฝ)
๐‘–
=

[(๐›ฝ๐‘—)1
,(๐›ฝ๐‘—)2

,โ€ฆ,(๐›ฝ๐‘—)๐‘–
]
โ€ฒ
 is updating using the following formula: 

(๏ฟฝฬ‚๏ฟฝ)
๐‘–
= (๏ฟฝฬ‚๏ฟฝ)

๐‘–โˆ’1
+ ๐›ผ๐‘–๐’–๐‘–,   (3.2) 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
18 

where ๐›ผ๐‘– โˆˆ [0,1] which signifies how long the estimate runs in the path before 
another predictor enters the model and the path turns anew, and ๐’–๐‘– is the equiangular 
vector. 

The path vector ๐’–๐‘– is computed using the formula given below based on the 
generalized form defined in equation (2.4): 

๐’–๐‘– = ๐‘ฎ๐ธ((๐‘ฌ)๐’Š
โ€ฒ(๐‘ฟโ€ฒ๐‘ฟ + ๐‘นโ€ฒ๐‘พโˆ’๐Ÿ๐‘น)(๐‘ฌ)๐’Š) 

โˆ’๐Ÿ(๐‘ฌ)๐’Š
โ€ฒ(๐‘ฟโ€ฒ๐’“๐’Šโˆ’๐Ÿ + ๐‘นโ€ฒ๐‘พ

โˆ’๐Ÿ๐‰๐’Šโˆ’๐Ÿ), (3.3) 

where (๐‘ฌ)๐’Š = [(๐’†๐‘—)1
,(๐’†๐‘—)2

,โ€ฆ ,(๐’†๐‘—)๐‘–
],  and  (๐‘’๐‘—)๐‘– be  the  ๐‘—

th  standard  unit  vector  

in ๐‘…๐‘, which has the record of selected variables in every succeeding steps, and ๐‘ฎ๐ธ is 
a generalized matrix that can be changed by respective expressions for any of 

stochastic restricted estimators of our interest as outlined in Table 1. 

Then, ๐›ผ๐‘– be calculated as follows: 
๐›ผ๐‘– = min

 
{๐›ผ๐‘–

+,๐›ผ๐‘–
โˆ’,๐›ผ๐‘–

โˆ—},      (3.4)  

where 

๐›ผ๐‘–
ยฑ =

๐ถ๐‘œ๐‘Ÿ((๐‘‹๐’‹)๐‘–
,๐’“๐’Šโˆ’๐Ÿ)ยฑ๐ถ๐‘œ๐‘Ÿ(๐‘‹๐‘—,๐’“๐’Šโˆ’๐Ÿ)

๐ถ๐‘œ๐‘Ÿ((๐‘‹๐’‹)๐‘–
,๐’“๐’Šโˆ’๐Ÿ)ยฑ๐ถ๐‘œ๐‘Ÿ(๐‘‹๐‘—,(๐‘ฟ)๐‘–๐’–๐‘–)

    (3.5) 

for any ๐‘—  such that (๐›ฝ๐‘—)๐‘–โˆ’1
= 0, and 

๐›ผ๐‘–
โˆ— = โˆ’

(๏ฟฝฬ‚๏ฟฝ)
๐‘–โˆ’1

๐’–๐‘–
      (3.6) 

for any ๐‘— such that (๐›ฝ๐‘—)๐‘–โˆ’1
โ‰  0. 

If ๐›ผ๐‘– = ๐›ผ๐‘–
โˆ—, then (๐‘ฌ)๐’Š is reformed by deleting the column ๐‘’๐‘— from (๐‘ฌ)๐’Šโˆ’๐Ÿ.  Then ๐’“๐’Š 

and ๐‰๐’Š can be calculated as 
๐’“๐’Š = ๐’“๐’Šโˆ’๐Ÿ โˆ’ ๐›ผ๐‘–(๐‘ฟ)๐‘–๐’–๐‘–   and     (3.7) 
๐‰๐’Š = ๐‰๐’Šโˆ’๐Ÿ โˆ’ ๐›ผ๐‘–(๐‘น)๐‘–๐’–๐‘–,     (3.8) 

where (๐‘น)๐‘– = [(๐‘น๐’‹)1
,(๐‘น๐’‹)2

,โ€ฆ,(๐‘น๐’‹)๐‘–
]. Then proceed to the next step where (๐‘—)๐‘–+1 

is the value of ๐‘— such that ๐›ผ๐‘– = ๐›ผ๐‘–
+or ๐›ผ๐‘– = ๐›ผ๐‘–

โˆ’ or ๐›ผ๐‘– = ๐›ผ๐‘–
โˆ— . Proceed with the algorithm 

until๐›ผ๐‘– = 1. 
 
Table 1: ๐บ๐ธ of the estimators for SRGLARS. 

Estimators                      ๐‘ฎ๐ธ 
MRE (๐‘ฌ)๐’Š 
SRRE (๐‘ฌ)๐’Š((๐‘ฌ)๐’Š

โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)(๐‘ฌ)๐’Š)
โˆ’1(๐‘ฌ)๐’Š

โ€ฒ ๐‘ฟโ€ฒ๐‘ฟ(๐‘ฌ)๐’Š 
SRAURE (๐‘ฌ)๐’Š(๐‘ฐ๐‘ƒ๐ธ โˆ’ ๐‘˜

2((๐‘ฌ)๐’Š
โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)(๐‘ฌ)๐’Š)

โˆ’2) 
SRLE (๐‘ฌ)๐’Š((๐‘ฌ)๐’Š

โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘ฐ)(๐‘ฌ)๐’Š)
โˆ’1(๐‘ฌ)๐’Š

โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘‘๐‘ฐ)(๐‘ฌ)๐’Š 
SRAULE (๐‘ฌ)๐’Š(๐‘ฐ๐‘ƒ๐ธ โˆ’ (1 โˆ’ ๐‘‘)

2((๐‘ฌ)๐’Š
โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘ฐ)(๐‘ฌ)๐’Š)

โˆ’2) 
SRPCRE ๐‘ปโ„Ž๐ธ๐‘ปโ„Ž๐ธ

โ€ฒ (๐‘ฌ)๐’Š 
SRrk ๐‘ปโ„Ž๐ธ๐‘ปโ„Ž๐ธ

โ€ฒ (๐‘ฌ)๐’Š((๐‘ฌ)๐’Š
โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘˜๐‘ฐ)(๐‘ฌ)๐’Š)

โˆ’1(๐‘ฌ)๐’Š
โ€ฒ ๐‘ฟโ€ฒ๐‘ฟ(๐‘ฌ)๐’Š 

SRrd ๐‘ปโ„Ž๐ธ๐‘ปโ„Ž๐ธ
โ€ฒ (๐‘ฌ)๐’Š((๐‘ฌ)๐’Š

โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘ฐ)(๐‘ฌ)๐’Š)
โˆ’1(๐‘ฌ)๐’Š

โ€ฒ (๐‘ฟโ€ฒ๐‘ฟ + ๐‘‘๐‘ฐ)(๐‘ฌ)๐’Š 
SROE (๐‘ฌ)๐’Š((๐‘ฌ)๐’Š

โ€ฒ ๐œทโˆ—๐œทโˆ—
โ€ฒ(๐‘ฌ)๐’Š(๐œŽ

2((๐‘ฌ)๐’Š
โ€ฒ ๐‘ฟโ€ฒ๐‘ฟ(๐‘ฌ)๐’Š)

โˆ’๐Ÿ + (๐‘ฌ)๐’Š
โ€ฒ ๐œทโˆ—๐œทโˆ—

โ€ฒ(๐‘ฌ)๐’Š)
โˆ’1

 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
19 

In Table 1, ๐‘ฐ๐‘ƒ๐ธ is the ๐‘๐ธ ร— ๐‘๐ธ identity matrix, ๐‘๐ธ is the amount of selected variables 
in each succeeding step, and ๐‘ปโ„Ž๐ธ = (๐‘ก1, ๐‘ก2. . . ๐‘กโ„Ž๐ธ)  is the first โ„Ž๐ธ column of the 
standardized eigenvectors of (๐‘ฌ)๐’Š

โ€ฒ ๐‘ฟโ€ฒ๐‘ฟ(๐‘ฌ)๐’Š. 
We can apply SRGLARS to consolidate LASSO and any of the stochastic restricted 

estimators listed in Table 1. The suitable value of regularization parameter ๐‘˜ or ๐‘‘ of 
the proposed algorithms can be chosen by 10-fold cross-validation for every ๐‘ก as 
outlined in Appendix C. 

We can get a separate algorithm for each stochastic restricted estimator by referring 

SRGLARS as LARS-MRE, LARS-SRRE, LARS-SRAURE, LARS-SRLE, LARS-

SRAULE, LARS-SRPCRE, LARS-SRrk, LARS-SRrd, and LARS-SROE when GE 

equals to the corresponding expression of MRE, SRRE, SRAURE, SRLE, SRAULE, 

SRPCRE, SRrk, SRrd, and SROE, respectively. 

4 Discussion 

The performance of SRGLARS algorithm by considering different combinations of 

LASSO and stochastic restricted estimators were examined using RMSE criterion, 

which is the expected prediction error of the algorithm for each estimator, and is 

defined as 

๐‘…๐‘€๐‘†๐ธ(๏ฟฝฬ‚๏ฟฝ) = โˆš
1

๐‘›โˆ—
  (๐’šโˆ— โˆ’ ๐‘ฟโˆ—๏ฟฝฬ‚๏ฟฝ)

โ€ฒ
(๐’šโˆ— โˆ’ ๐‘ฟโˆ—๏ฟฝฬ‚๏ฟฝ),   (4.1) 

where (๐’šโˆ—,๐‘ฟโˆ—) are the new observations, ๐‘›โˆ— is the size of new observations, and ๏ฟฝฬ‚๏ฟฝ is 
the estimated value of ๐œท by the corresponding algorithm. Note that we cannot provide 
theoretical conditions for the superiority of particular algorithms in variable selection 

methods. Therefore, we use a Monte Carlo simulation and a real-world example to 

compare the SRGLARS algorithms. 

 

4.1 Simulation study 

 

We used the following formula to generate the explanatory variables based on 

McDonald and Galarneau (1975): 

๐‘ฅ๐‘–,๐‘—  = โˆš(1 โˆ’ ๐œŒ
2)๐‘ง๐‘–,๐‘— + ๐œŒ๐‘ง๐‘–,๐‘š+1;    ๐‘– =  1,2, . . . ,๐‘›.     ๐‘— =  1,2, . . . ,๐‘š.   (4.2) 

where ๐‘ง๐‘–,๐‘— is an independent standard normal pseudo-random number, and ๐œŒ is the 

correlation among any two explanatory variables. 

In this study, we have used 100 observations with 20 explanatory variables, in which 

70 observations were used to fit the model and 30 observations were used to compute 

the RMSE.  

 

A dependent variable is generated by using the following equation, 

๐‘ฆ๐‘– = ๐›ฝ1๐‘ฅ๐‘–,1 + ๐›ฝ2๐‘ฅ๐‘–,2 + โ‹ฏ+ ๐›ฝ20๐‘ฅ๐‘–,20 + ๐œ€๐‘–,  (4.3) 

where ๐œ€๐‘– is a normal pseudo-random number with ๐ธ(๐œ€๐‘–) = 0 and ๐‘‰ (๐œ€๐‘–) = ๐œŽ
2 = 1. 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
20 

We choose ๐›ฝ = (๐›ฝ1,๐›ฝ2,โ€ฆ,๐›ฝ20) as the normalized eigenvector corresponding to the 
largest eigenvalue of ๐‘ฟโ€ฒ๐‘ฟ for which ๐œทโ€ฒ๐œท =  1 (McDonald and Galarneau 1975).  Prior 
information was defined based on Nagar and Kakwaniโ€™s (1964) approach, which is 

described in Appendix B. Further, we have assumed the first four elements of OLSE 

estimates as ๐’ƒ (see Appendix B). To study the effects of various degrees of 
multicollinearity on the data, we pick ๐œŒ = (0.5, 0.7, 0.9), which signifies weak, 
moderate, and high multicollinearity, respectively.   Since the execution time for the 

algorithm simulation is long due to cross-validation, we simulated only 50 datasets. 

Future simulation studies will be implemented with a greater number of simulations 

using cluster programming. Figures 1-3 and Tables 2-4 show the cross-validated RMSE 

and the median cross-validated RMSE of the SRGLARS algorithms, respectively, for 

50 simulated data.  

 

 
 

 

 

 

 

 

 

 

 

 

 

 

 

Fig 1. Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.5 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Fig 2. Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.7 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
21 

 

  

 

 

 

 

 

 

 

 

Fig 3. Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.9 

Table 2: Median Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.5. 

Algorithms RMSE (k, d) t Selected variables 

LARS-MRE 3.2805 โ€“ 6.1630 15 

LARS-SRRE 3.3531 0.3 6.1777 16 

LARS-SRAURE   3.2832 0.1 6.1631 16 

LARS-SRLE 3.3531 0.7 6.1773 16 

LARS-SRAULE   3.3012 0.99 6.1630 16 

LARS-SRPCRE    3.3070 โ€“ 6.4343 16 

LARS-SRrk 3.3122 0.1 6.9493 17 

LARS-SRrd 3.2976 0.9 6.9497 16 

LARS-SROE 3.0182 โ€“ 4.1102 18 

Table 3: Median Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.7. 

Algorithms RMSE (k, d) t Selected variables 

LARS-MRE 3.2041 โ€“ 6.8458 15 

LARS-SRRE 3.3575 0.1 7.5043 16 

LARS-SRAURE 3.3884 0.8 6.7785 16 

LARS-SRLE 3.3537 0.9 7.5049 17 

LARS-SRAULE 3.3886 0.2 6.7786 16 

LARS-SRPCRE 3.3135 โ€“ 7.7712 16 

LARS-SRrk 3.2525 1.0 8.5411 17 

LARS-SRrd 3.2534 0.5 8.1097 17 

LARS-SROE 3.0324 โ€“ 4.3864 18 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
22 

Table 4: Median Cross-validated RMSE values of the SRGLARS algorithms when ฯ = 0.9. 

Algorithms RMSE (k, d) t Selected variables 

LARS-MRE 3.3391 โ€“ 11.6628 16 

LARS-SRRE 3.2722 0.1 11.5897 17 

LARS-SRAURE   3.3354 0.1 11.6606 16 

LARS-SRLE 3.3110 0.99 11.6628 17 

LARS-SRAULE   3.3107 0.99 11.6628 16 

LARS-SRPCRE    3.3782 โ€“ 12.2288 16 

LARS-SRrk 3.3531 0.7 12.9540 17 

LARS-SRrd 3.3276 0.3 13.0038 17 

LARS-SROE     2.9878 โ€“ 4.5654 18 

 

From Figure 1-3, we can observe that the LARS-SROE algorithm outperformed 

other SRGLARS algorithms in RMSE criterion under all degrees of multicollinearity. 

From Tables 2-4, we observe that the SROE included more variables than the other 

SRGLARS algorithms, although it has minimum RMSE. Therefore, in practical 

situations, if a researcher intends to reduce the number of variables from the model, 

they may consider other SRGLARS algorithms based on the interested variables and 

prediction performance using the plot of coefficient paths discussed in the real-world 

example. 

 

4.2 Real-world example 
  

As a real-world data set, we considered the Prostate Cancer Data (Stamey et al. 1989), 

which is used by Tibshirani (1996), Efron et al. (2004) and Zou and Hastie (2005) to 

study the performance of LASSO, LARS algorithm and Enet. The Prostate Cancer 

Data contains 97 observations and 8 predictors such as log cancer volume (lcavol), log 

prostate weight (lweight), age, log of the amount of benign prostatic hyperplasia (lbph), 

seminal vesicle invasion (svi), log capsular penetration (lcp), Gleason score (gleason) 

and percentage Gleason score 4 or 5 (pgg45). The dependent variable is the log of 

prostate-specific antigen (lpsa). The Variance Inflation Factor (VIF) values of the 

predictor variables are 3.09, 2.97, 2.47, 2.05, 1.95, 1.37, 1.36 and 1.32, and the 

condition number is 243, which exposes high multicollinearity between the predictor 

variables. This data set is associated in โ€œlasso2โ€ R package. We have used 67 

observations to fitting the model, and 30 observations to compute the RMSE. We have 

assumed (Nagar and Kakwani 1964) that the first three OLSE estimates of Prostate 

Cancer Data are unbiased, and we defined the prior information for this data based on 

Nagar and Kakwaniโ€™s (1964) approach, as described in Appendix B.  

The cross-validated RMSE of the SRGLARS algorithms are shown in Table 5, and 

coefficient paths of respective SRGLARS algorithm are shown in Figure 4 in Appendix 

A. 

 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
23 

Table 5: Cross-validated RMSE values of Prostate Cancer Data using SRGLARS. 

Algorithms RMSE (k, d) t Selected variables 

LARS-MRE 0.77784 โ€“ 1.5632 6 

LARS-SRRE 0.73567 0.38 1.7329 8 

LARS-SRAURE 0.77784 0.01 1.5632 6 

LARS-SRLE 0.73566 0.62 1.7331 8 

LARS-SRAULE 0.77784 0.99 1.5632 6 

LARS-SRPCRE 0.74084 โ€“ 1.5656 7 

LARS-SRrk 0.74083 0.04 1.5643 7 

LARS-SRrd 0.74084 0.96 1.5643 7 

LARS-SROE 0.69897 โ€“ 1.1226 8 

 

Table 5 shows that the LARS-SROE algorithm outperforms other algorithms on 

Prostate Cancer Data, which is agreed the results obtained in the simulation study. 

Further, we can note that LARS-SRRE, LARS-SRLE, and LARS-SROE did not make 

any variable selections on Prostate Cancer Data.  

Figure 4 (see Appendix A) shows that the choice of variables is different for each 

SRGLARS algorithms. Note that in all graphs, the order of selection of variables is the 

same except Figure 4(i). Further, the coefficients relevant to the variables lcp, age and 

gleason are very small for Fig 4((a)-(h)) and those coefficients are almost zero in Fig 

4(i). This indicates that the contribution from those three variables is very low for the 

fitted model.  

5   Conclusions  

This study presented a generalized version of stochastic restricted LARS (SRGLARS) 

algorithm to combine LASSO with existing stochastic restricted estimators. We have 

shown the superiority of SRGLARS when it combines LASSO with SROE (LARS-

SROE) using a Monte Carlo simulation and a real-world example in RMSE criterion. 

However, LARS-SROE is less likely to make variable selections when looking at the 

results directly. However, by examining the size of the coefficients as shown in the 

example, we can drop some irrelevant variables form the model. If a researcher wanted 

to reduce some variables from the model, the other combinations of the SRGLARS 

algorithm could be considered to fit the regression model by studying the RMSE and 

coefficients estimates based on the coefficients path. 

Acknowledgements 

Anonymous reviewers are acknowledged for their comments on the initial manuscript. 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
24 

References 

Akdeniz F, Kaรงiranlar S. 1995. On the almost unbiased generalized liu estimator and unbiased estimation 

of the bias and mse. Communication in Statistics - Theory and Methods 24(7):1789โ€“1797. 

Baye MR, Parker DF. 1984. Combining ridge and principal component regression: A money demand 

illustration. Communications in Statistics-Theory and Methods 13(2):197โ€“205.  

Efron B, Hastie T, Johnstone I, Tibshirani R. 2004. Least angle regression. The Annals of statistics 

32(2):407โ€“499.   

He D, Wu Y. 2014. A stochastic restricted principal components regression estimator in the linear model. 

The Scientific World Journal, vol. 2014, Article ID 231506, 6p. https://doi.org/10.1155/2014/231506 

Hoerl E, Kennard RW. 1970. Ridge regression: Biased estimation for nonorthogonal problems. 

Technometrics 12:55โ€“67.  

Hubert MH, Wijekoon P. 2006. Improvement of the Liu estimator in linear regression model. Statistical 

Papers 47:471โ€“479.   

Jibo W. 2014. On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class 

Estimator in Linear Regression Model. Journal of Applied Mathematics, vol. 2014, Article ID 

173836, 6 p. https://doi.org/10.1155/2014/173836 

Jibo W, Hu Y. 2014. On the stochastic restricted almost unbiased estimators in linear regression model. 

Communications in Statistics-Simulation and Computation 43(2):428โ€“440.  

Kaรงiranlar S, SakallฤฑoฤŸlu S. 2001. Combining the Liu estimator and the principal component Regression 

Estimator. Communications in Statistics-Theory and Methods 30(12):2699โ€“2705.   

Kayanan M, Wijekoon P. 2018. Stochastic Restricted Biased Estimators in Misspecified Regression 

Model with Incomplete Prior Information. Journal of Probability and Statistics, vol. 2018, Article 

ID 1452181, 8 p. https://doi.org/10.1155/2018/1452181.  

Kayanan M, Wijekoon P. 2019. Optimal estimators in misspecified linear regression model with an 

application to real-world data. Communications in Statistics: Case Studies, Data Analysis and 

Applications 4(3-4):151โ€“163.   

Kayanan M, Wijekoon P. 2020a. Stochastic Restricted LASSO-Type Estimator in the Linear Regression 

Model. Journal of Probability and Statistics, vol. 2020, Article ID 7352097, 7 p. 

https://doi.org/10.1155/2020/7352097. 

Kayanan M, Wijekoon P. 2020b. Variable Selection via Biased Estimators in the Linear Regression 

Model. Open Journal of Statistics 10(1):113โ€“126.    

Li Y, Yang H. 2010. A new stochastic mixed ridge estimator in linear regression model. Statistical Papers 

51:315โ€“323.   

Liu K. 1993. A new class of biased estimate in linear regression. Communication in Statistics - Theory 

and Methods 22(2):393โ€“402.   

Massy WF. 1965. Principal components regression in exploratory statistical research. Journal of the 

American Statistical Association 60:234โ€“266. 

McDonald GC, Galarneau DI. 1975. A Monte Carlo Evaluation of Some Ridge-Type Estimators. Journal 

of the American Statistical Association 70:407โ€“416.   

Nagar AL, Kakwani NC. 1964. The Bias and Moment Matrix of a Mixed Regression Estimator. 

Econometrica 32:174โ€“182.   

Singh B, Chaubey YP, Dwivedi TD. 1986. An Almost Unbiased Ridge Estimator. The Indian Journal of 

Statistics 48:342โ€“346. 

Stamey TA, Kabalin JN, Mcneal JE, Johnstone IM, Freiha F, Redwine EA, Yang N. 1989. Prostate 

specific antigen in the diagnosis and treatment of adenocarcinoma of the prostate: II. radical 

prostatectomy treated patients. Journal of Urology 141:1076โ€“1083.  

Theil H, Goldberger AS. 1961. On Pure and Mixed Statistical Estimation in Economics. International 

Economic Review 2:65โ€“78.  

Tibshirani R. 1996. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical 

Society: Series B (Methodological) 58:267โ€“288.  

Zou H, Hastie T. 2005. Regularization and variable selection via the elastic net. Journal of the Royal 

Statistical Society: Series B 67:301โ€“320. 

https://doi.org/10.1155/2014/231506


 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
25 

Appendix A: Figure 4 

 

(a) 

 

(b) 

 

(c) 

 

(d) 

 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
26 

 

(e) 

 

(f) 

 

(g) 

 

(h) 

 



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
27 

 

(i) 

Fig 4. Coefficient paths of the (a) LARS-MRE, (b) LARS-SRRE, (c) LARS- SRAURE, (d) 

LARS-SRLE, (e) LARS-SRAULE, (f) LARS-SRPCRE, (g) LARS-SRrk (h) LARS-SRrd 

and (i) LARS-SROE versus ๐’• = โˆ‘ |๐œท๐’‹|
๐’‘
๐’‹=๐Ÿ   for the Prostate Cancer Data. 

 

 

  



 M. Kayanan and P. Wijekoon                                                         Generalized Stochastic Restricted LARS Algorithm 

Ruhuna Journal of Science 

Vol 13 (1): 14-28, June 2022 
28 

Appendix B: Selection of prior information 

 

According to Nagar and Kakwani (1964), we can define the prior information as follows:  

Let ๐œท1 be a vector of some selected ๐‘ž elements of ๐œท and ๐œท2 is the rest of elements. Assume 
that ๐’ƒ is the known unbiased estimates of ๐œท1. By using the โ€œtwo sigma ruleโ€, now we can write 
the range of ๐œท1 as ๐’ƒ ยฑ 2๐‘†๐ธ(๐’ƒ). Based on that we can set the expressions of equation (2.2) as 

๏ฟฝฬ‚๏ฟฝ = ๐’ƒ,   ๏ฟฝฬ‚๏ฟฝ = (

1 0 โ‹ฏ 0 โ‹ฏ 0
0 1 โ‹ฏ 0 โ‹ฏ 0
โ‹ฎ
0

โ‹ฎ
0

โ‹ฑ
โ‹ฏ

โ‹ฎ
1

โ‹ฏ
โ‹ฏ

0
0

)

๐‘žร—๐‘

, ๐œท = (
๐œท1
๐œท2
)  and  

 ๐œŽ2๏ฟฝฬ‚๏ฟฝ = (

๐‘†๐ธ(๐‘1) 0      โ‹ฏ      0

0
โ‹ฎ
0

๐‘†๐ธ(๐‘2)
โ‹ฎ
0

โ‹ฏ
โ‹ฑ
โ‹ฏ

0
โ‹ฎ

๐‘†๐ธ(๐‘๐‘ž)

)

๐‘žร—๐‘ž

. 

 

Appendix C: K-fold cross-validation to estimate shrinkage parameters, (k, d)  

 

Step 1: Split the data set into ๐พ groups. 
Step 2: For each unique group ๐‘– =  1,2, . . . ,๐พ:  

โ€ข Take one group as a test data set 

โ€ข Take the remaining ๐พ โˆ’ 1 groups as a training data set 

โ€ข Estimate the respective estimator ๏ฟฝฬ‚๏ฟฝ(๐‘˜,๐‘‘)  with shrinkage parameters (๐‘˜,๐‘‘) using 
the training data set. Use intital values of ๐‘˜,๐‘‘ as 0.01, and then compute the RMSE 
cross-validation errors separately as follows: 

๐ถ๐‘‰_๐‘…๐‘€๐‘†๐ธ๐‘–(๏ฟฝฬ‚๏ฟฝ(๐‘˜,๐‘‘))   = โˆš
๐พ

๐‘›
 โˆ‘ (๐‘ฆ๐‘–

โˆ— โˆ’ โˆ‘ ๐‘ฅ๐‘–๐‘—
โˆ— ๏ฟฝฬ‚๏ฟฝ๐‘—(๐‘˜,๐‘‘)

๐‘
๐‘—=1

)
2

๐‘›/๐พ
๐‘–=1

, 

 

where ๐‘ฆ๐‘–
โˆ—  and ๐‘ฅ๐‘–๐‘—

โˆ—   are the variables belongs to the test data set. 

Step 3: Then, find the overall cross-validation errors as follows: 

๐ถ๐‘‰_๐‘…๐‘€๐‘†๐ธ (๏ฟฝฬ‚๏ฟฝ(๐‘˜,๐‘‘)) = 
1

๐พ
โˆ‘๐ถ๐‘‰_๐‘…๐‘€๐‘†๐ธ๐‘– (๏ฟฝฬ‚๏ฟฝ(๐‘˜,๐‘‘))

๐พ

๐‘–=1

    

Step 4: Continue this procedure by increasing the values of ๐‘˜ or ๐‘‘ by a small increment, and 

choose the value of ๐‘˜ or ๐‘‘ that makes ๐ถ๐‘‰_๐‘…๐‘€๐‘†๐ธ (๏ฟฝฬ‚๏ฟฝ(๐‘˜,๐‘‘))  smallest.