Characterization of the equivalence of robustification and regularization in linear and matrix regression

•We characterize the connection between robust optimization and regularization.•We extend the characterization to new settings such as matrix completion.•Robust optimization and regularization are not always equivalent. The notion of developing statistical methods in machine learning which are robus...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:European journal of operational research Jg. 270; H. 3; S. 931 - 942
Hauptverfasser: Bertsimas, Dimitris, Copenhaver, Martin S.
Format: Journal Article
Sprache:Englisch
Veröffentlicht: Elsevier B.V 01.11.2018
Schlagworte:
ISSN:0377-2217, 1872-6860
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•We characterize the connection between robust optimization and regularization.•We extend the characterization to new settings such as matrix completion.•Robust optimization and regularization are not always equivalent. The notion of developing statistical methods in machine learning which are robust to adversarial perturbations in the underlying data has been the subject of increasing interest in recent years. A common feature of this work is that the adversarial robustification often corresponds exactly to regularization methods which appear as a loss function plus a penalty. In this paper we deepen and extend the understanding of the connection between robustification and regularization (as achieved by penalization) in regression problems. Specifically,(a)In the context of linear regression, we characterize precisely under which conditions on the model of uncertainty used and on the loss function penalties robustification and regularization are equivalent.(b)We extend the characterization of robustification and regularization to matrix regression problems (matrix completion and Principal Component Analysis).
ISSN:0377-2217
1872-6860
DOI:10.1016/j.ejor.2017.03.051