Search Results - "prox-linear algorithm"

  • Showing 1 - 18 results of 18
Refine Results
  1. 1

    Hybrid SGD algorithms to solve stochastic composite optimization problems with application in sparse portfolio selection problems by Yang, Zhen-Ping, Zhao, Yong

    ISSN: 0377-0427, 1879-1778
    Published: Elsevier B.V 15.01.2024
    “…In this paper, we study stochastic composite problems where the objective can be the composition of an outer single-valued function and an inner vector-valued…”
    Get full text
    Journal Article
  2. 2

    Stochastic variance-reduced prox-linear algorithms for nonconvex composite optimization by Zhang, Junyu, Xiao, Lin

    ISSN: 0025-5610, 1436-4646
    Published: Berlin/Heidelberg Springer Berlin Heidelberg 01.09.2022
    Published in Mathematical programming (01.09.2022)
    “… We propose a class of stochastic variance-reduced prox-linear algorithms for solving such problems and bound their sample complexities for finding an ϵ…”
    Get full text
    Journal Article
  3. 3

    REAL-TIME POWER SYSTEM STATE ESTIMATION VIA DEEP UNROLLED NEURAL NETWORKS by Zhang, Liang, Wang, Gang, Giannakis, Georgios B.

    Published: IEEE 01.11.2018
    “…Contemporary smart power grids are being challenged by rapid voltage fluctuations, due to large-scale deployment of electric vehicles, demand response…”
    Get full text
    Conference Proceeding
  4. 4

    Low-Rank Matrix Recovery with Composite Optimization: Good Conditioning and Rapid Convergence by Charisopoulos, Vasileios, Chen, Yudong, Davis, Damek, Díaz, Mateo, Ding, Lijun, Drusvyatskiy, Dmitriy

    ISSN: 1615-3375, 1615-3383
    Published: New York Springer US 01.12.2021
    Published in Foundations of computational mathematics (01.12.2021)
    “…The task of recovering a low-rank matrix from its noisy linear measurements plays a central role in computational science. Smooth formulations of the problem…”
    Get full text
    Journal Article
  5. 5

    Robust and Scalable Power System State Estimation via Composite Optimization by Wang, Gang, Giannakis, Georgios B., Chen, Jie

    ISSN: 1949-3053, 1949-3061
    Published: Piscataway IEEE 01.11.2019
    Published in IEEE transactions on smart grid (01.11.2019)
    “…In today's cyber-enabled smart grids, high penetration of uncertain renewables, purposeful manipulation of meter readings, and the need for wide-area…”
    Get full text
    Journal Article
  6. 6

    Stochastic Variance-Reduced Prox-Linear Algorithms for Nonconvex Composite Optimization by Zhang, Junyu, Lin, Xiao

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 14.05.2021
    Published in arXiv.org (14.05.2021)
    “… We propose a class of stochastic variance-reduced prox-linear algorithms for solving such problems and bound their sample complexities for finding an \(\epsilon…”
    Get full text
    Paper
  7. 7

    A Barzilai–Borwein-Like Iterative Half Thresholding Algorithm for the L1/2 Regularized Problem by Wu, Lei, Sun, Zhe, Li, Dong-Hui

    ISSN: 0885-7474, 1573-7691
    Published: New York Springer US 01.05.2016
    Published in Journal of scientific computing (01.05.2016)
    “…In this paper, we propose a Barzilai–Borwein-like iterative half thresholding algorithm for the L 1 / 2 regularized problem. The algorithm is closely related…”
    Get full text
    Journal Article
  8. 8

    Large-Scale Distributed Sparse Class-Imbalance Learning by Maurya, Chandresh Kumar, Toshniwal, Durga

    ISSN: 0020-0255, 1872-6291
    Published: Elsevier Inc 01.08.2018
    Published in Information sciences (01.08.2018)
    “…Class-imbalance learning is a classic problem in data mining and machine learning community. In class-imbalance learning, the idea is to learn the model so…”
    Get full text
    Journal Article
  9. 9

    The proximal point method revisited by Drusvyatskiy, Dmitriy

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 17.12.2017
    Published in arXiv.org (17.12.2017)
    “… I focus on three recent examples: a proximally guided subgradient method for weakly convex stochastic approximation, the prox-linear algorithm for minimizing compositions of convex functions and smooth maps, and Catalyst generic…”
    Get full text
    Paper
  10. 10

    Stochastic Methods for Composite and Weakly Convex Optimization Problems by Duchi, John, Ruan, Feng

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 21.09.2018
    Published in arXiv.org (21.09.2018)
    “… We develop a family of stochastic methods---including a stochastic prox-linear algorithm and a stochastic (generalized…”
    Get full text
    Paper
  11. 11

    Solving (most) of a set of quadratic equalities: Composite optimization for robust phase retrieval by Duchi, John C, Ruan, Feng

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 22.04.2018
    Published in arXiv.org (22.04.2018)
    “… We show that the prox-linear algorithm we develop can solve phase retrieval problems---even with adversarially faulty measurements---with high probability as soon as the number of measurements \(m…”
    Get full text
    Paper
  12. 12

    Stochastic algorithms with geometric step decay converge linearly on sharp functions by Davis, Damek, Drusvyatskiy, Dmitriy, Charisopoulos, Vasileios

    ISSN: 0025-5610, 1436-4646
    Published: Berlin/Heidelberg Springer Berlin Heidelberg 01.09.2024
    Published in Mathematical programming (01.09.2024)
    “…Stochastic (sub)gradient methods require step size schedule tuning to perform well in practice. Classical tuning strategies decay the step size polynomially…”
    Get full text
    Journal Article
  13. 13

    Parallel and distributed sparse optimization by Peng, Zhimin, Yan, Ming, Yin, Wotao

    ISSN: 1058-6393
    Published: IEEE 01.11.2013
    “… (i) distributed implementations of prox-linear algorithms and (ii) GRock, a parallel greedy block coordinate descent method…”
    Get full text
    Conference Proceeding
  14. 14

    Modified Gauss-Newton Algorithms under Noise by Pillutla, Krishna, Roulet, Vincent, Kakade, Sham M., Harchaoui, Zaid

    ISSN: 2693-3551
    Published: IEEE 02.07.2023
    “… Their nonsmooth counterparts, modified Gauss-Newton or prox-linear algorithms, can lead to contrasting outcomes when compared to gradient descent in large-scale statistical settings…”
    Get full text
    Conference Proceeding
  15. 15

    A Barzilai-Borwein-Like Iterative Half Thresholding Algorithm for the \(L_{1/2}\) Regularized Problem by Wu, Lei, Sun, Zhe, Li, Dong-Hui

    ISSN: 0885-7474, 1573-7691
    Published: 01.05.2016
    Published in Journal of scientific computing (01.05.2016)
    “…In this paper, we propose a Barzilai-Borwein-like iterative half thresholding algorithm for the \(L_{1/2}\) regularized problem. The algorithm is closely…”
    Get full text
    Journal Article
  16. 16

    Modified Gauss-Newton Algorithms under Noise by Pillutla, Krishna, Roulet, Vincent, Kakade, Sham, Harchaoui, Zaid

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 18.05.2023
    Published in arXiv.org (18.05.2023)
    “… Their nonsmooth counterparts, modified Gauss-Newton or prox-linear algorithms, can lead to contrasting outcomes when compared to gradient descent in large-scale statistical settings…”
    Get full text
    Paper
  17. 17

    Stochastic algorithms with geometric step decay converge linearly on sharp functions by Davis, Damek, Drusvyatskiy, Dmitriy, Charisopoulos, Vasileios

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 22.07.2019
    Published in arXiv.org (22.07.2019)
    “…Stochastic (sub)gradient methods require step size schedule tuning to perform well in practice. Classical tuning strategies decay the step size polynomially…”
    Get full text
    Paper
  18. 18

    Run-and-Inspect Method for Nonconvex Optimization and Global Optimality Bounds for R-Local Minimizers by Chen, Yifan, Sun, Yuejiao, Yin, Wotao

    ISSN: 2331-8422
    Published: Ithaca Cornell University Library, arXiv.org 29.06.2018
    Published in arXiv.org (29.06.2018)
    “…Many optimization algorithms converge to stationary points. When the underlying problem is nonconvex, they may get trapped at local minimizers and occasionally…”
    Get full text
    Paper