A majorization-minimization scheme for L 2 support vector regression
In a support vector regression (SVR) model, using the squared ϵ-insensitive loss function makes the optimization problem strictly convex and yields a more concise solution. However, the formulation of -SVR leads to a quadratic programming which is expensive to solve. This paper reformulates the opti...
Uloženo v:
| Vydáno v: | Journal of statistical computation and simulation Ročník 91; číslo 15; s. 3087 - 3107 |
|---|---|
| Hlavní autor: | |
| Médium: | Journal Article |
| Jazyk: | angličtina |
| Vydáno: |
Taylor & Francis
13.10.2021
|
| Témata: | |
| ISSN: | 0094-9655, 1563-5163 |
| On-line přístup: | Získat plný text |
| Tagy: |
Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
|
| Shrnutí: | In a support vector regression (SVR) model, using the squared ϵ-insensitive loss function makes the optimization problem strictly convex and yields a more concise solution. However, the formulation of
-SVR leads to a quadratic programming which is expensive to solve. This paper reformulates the optimization problem of
-SVR by absorbing the constraints in the objective function, which can be solved efficiently by a majorization-minimization approach, in which an upper bound for the objective function is derived in each iteration which is easier to be minimized. The proposed approach is easy to implement, without requiring any additional computing package other than basic linear algebra operations. Numerical studies on real-world datasets show that, compared to the alternatives, the proposed approach can achieve similar prediction accuracy with substantially higher time efficiency in training. |
|---|---|
| ISSN: | 0094-9655 1563-5163 |
| DOI: | 10.1080/00949655.2021.1918691 |