<div class="csl-bib-body">
<div class="csl-entry">Filzmoser, P., & Nordhausen, K. (2021). Robust linear regression for high-dimensional data: an overview. <i>Wiley Interdisciplinary Reviews: Computational Statistics</i>. https://doi.org/10.1002/wics.1524</div>
</div>
-
dc.identifier.issn
1939-0068
-
dc.identifier.uri
http://hdl.handle.net/20.500.12708/137149
-
dc.description.abstract
Digitization as the process of converting information into numbers leads to bigger and more complex data sets, bigger also with respect to the number of measured variables. This makes it harder or impossible for the practitioner to identify outliers or observations that are inconsistent with an underlying model. Classical least‐squares based procedures can be affected by those outliers. In the regression context, this means that the parameter estimates are biased, with consequences on the validity of the statistical inference, on regression diagnostics, and on the prediction accuracy. Robust regression methods aim at assigning appropriate weights to observations that deviate from the model. While robust regression techniques are widely known in the low‐dimensional case, researchers and practitioners might still not be very familiar with developments in this direction for high‐dimensional data. Recently, different strategies have been proposed for robust regression in the high‐dimensional case, typically based on dimension reduction, on shrinkage, including sparsity, and on combinations of such techniques. A very recent concept is downweighting single cells of the data matrix rather than complete observations, with the goal to make better use of the model‐consistent information, and thus to achieve higher efficiency of the parameter estimates.