
Be the first to like this
Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our User Agreement and Privacy Policy.
Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our Privacy Policy and User Agreement for details.
Published on
Regularized regression technique
s for lin
ear regression have been creat
ed
the last few
ten
year
s to
reduce
the
flaws
of ordinary least squ
ares regression
with regard to prediction accuracy.
In this paper, new
methods
for using regularized regression in model
choice are introduc
ed, and we
distinguish
the condition
s
in whic
h regularized regression develops
our ability to discriminate models.
W
e applied all the five
methods that use penalty

based (regularization) shrinkage to
handle Oxazolines and Oxazoles derivatives
descriptor
dataset
with far more predictors than observations.
The lasso,
ridge,
elasticnet,
lars and relaxed
lasso
further pos
sess the desirable property that they simultaneously sele
ct relevant predictive descriptor
s
and optimally
estimate their effects.
Here, we comparatively evaluate the performance of five regularized
linear regression methods
The assessment of the performanc
e of each model by means of benchmark
experiments
is an
established exercise.
Cross

validation and
resampling
method
s
are genera
lly
used to
arrive
point
evaluates the efficienci
es which are compared
to recognize
methods
with acceptable feature
s.
Predictiv
e accuracy
was evaluated
us
ing the root mean squared error
(RMSE)
and
Square of usual
correlation between predictors and observed mean inhibitory concentration of antitubercular activity
(R
square)
.
We found that all five regularized regression models were
able to produce feasible models
and
efficient capturing the linearity in the data
.
The elastic net and lars had similar accuracies
as well as lasso
and relaxed lasso
had similar accuracies
but outperformed ridge regression in terms of the RMSE and R
squ
are
metrics.
Be the first to like this
Be the first to comment