Nam Le

Optimization Research Papers in JMLR Volume 22

Nam Le
Table of Contents

Optimization Research Papers in JMLR Volume 22 (2021) #

This document lists papers from JMLR Volume 22 (2021) that focus on optimization research, categorized by their primary themes. Each paper is numbered starting from 1 within its subsection, with a brief description of its key contributions to optimization theory, algorithms, or applications.

Convex Optimization #

Papers addressing convex optimization problems, including clustering, Wasserstein barycenters, sparse optimization, and bandits.

  1. Convex Clustering: Model, Theoretical Guarantee and Efficient Algorithm
    Authors: Defeng Sun, Kim-Chuan Toh, Yancheng Yuan
    Description: Proposes a convex clustering model with theoretical guarantees and an efficient algorithm.

  2. A Fast Globally Linearly Convergent Algorithm for the Computation of Wasserstein Barycenters
    Authors: Lei Yang, Jia Li, Defeng Sun, Kim-Chuan Toh
    Description: Develops a fast, globally linearly convergent algorithm for computing Wasserstein barycenters.

  3. Wasserstein Barycenters Can Be Computed in Polynomial Time in Fixed Dimension
    Authors: Jason M. Altschuler, Enric Boix-Adsera
    Description: Demonstrates that Wasserstein barycenters can be computed in polynomial time for fixed dimensions.

  4. From Low Probability to High Confidence in Stochastic Convex Optimization
    Authors: Damek Davis, Dmitriy Drusvyatskiy, Lin Xiao, Junyu Zhang
    Description: Analyzes methods to achieve high-confidence solutions in stochastic convex optimization.

  5. Sparse and Smooth Signal Estimation: Convexification of L0-Formulations
    Authors: Alper Atamturk, Andres Gomez, Shaoning Han
    Description: Proposes convexification techniques for L0-formulations in sparse and smooth signal estimation.

  6. Stochastic Proximal AUC Maximization
    Authors: Yunwen Lei, Yiming Ying
    Description: Develops stochastic proximal methods for maximizing the area under the ROC curve (AUC) in convex settings.

  7. Sparse Convex Optimization via Adaptively Regularized Hard Thresholding
    Authors: Kyriakos Axiotis, Maxim Sviridenko
    Description: Introduces adaptively regularized hard thresholding for sparse convex optimization.

  8. Learning Sparse Classifiers: Continuous and Mixed Integer Optimization Perspectives
    Authors: Antoine Dedieu, Hussein Hazimeh, Rahul Mazumder
    Description: Explores continuous and mixed-integer optimization approaches for learning sparse classifiers.

  9. First-Order Convergence Theory for Weakly-Convex-Weakly-Concave Min-max Problems
    Authors: Mingrui Liu, Hassan Rafique, Qihang Lin, Tianbao Yang
    Description: Provides first-order convergence theory for weakly convex-weakly concave min-max problems.

  10. Convex Geometry and Duality of Over-parameterized Neural Networks
    Authors: Tolga Ergen, Mert Pilanci
    Description: Analyzes convex geometry and duality in over-parameterized neural networks.

  11. Linear Bandits on Uniformly Convex Sets
    Authors: Thomas Kerdreux, Christophe Roux, Alexandre d’Aspremont, Sebastian Pokutta
    Description: Studies linear bandits on uniformly convex sets, focusing on convex optimization techniques.

Nonconvex Optimization #

Papers tackling nonconvex optimization, including stochastic gradient descent, neural network training, and stability properties.

  1. Online Stochastic Gradient Descent on Non-Convex Losses from High-Dimensional Inference
    Authors: Gerard Ben Arous, Reza Gheissari, Aukosh Jagannath
    Description: Analyzes online stochastic gradient descent for nonconvex losses in high-dimensional inference.

  2. Non-attracting Regions of Local Minima in Deep and Wide Neural Networks
    Authors: Henning Petzka, Cristian Sminchisescu
    Description: Investigates non-attracting regions of local minima in deep and wide neural networks.

  3. When Does Gradient Descent with Logistic Loss Find Interpolating Two-Layer Networks?
    Authors: Niladri S. Chatterji, Philip M. Long, Peter L. Bartlett
    Description: Examines conditions under which gradient descent with logistic loss finds interpolating two-layer networks.

  4. Replica Exchange for Non-Convex Optimization
    Authors: Jing Dong, Xin T. Tong
    Description: Proposes replica exchange methods for nonconvex optimization problems.

  5. Failures of Model-Dependent Generalization Bounds for Least-Norm Interpolation
    Authors: Peter L. Bartlett, Philip M. Long
    Description: Analyzes limitations of model-dependent generalization bounds in least-norm interpolation.

  6. On the Stability Properties and the Optimization Landscape of Training Problems with Squared Loss for Neural Networks and General Nonlinear Conic Approximation Schemes
    Authors: Constantin Christof
    Description: Studies stability and optimization landscapes for neural network training with squared loss.

Stochastic Optimization #

Papers focusing on stochastic optimization methods, including momentum, Langevin dynamics, and communication-efficient algorithms.

  1. Continuous Time Analysis of Momentum Methods
    Authors: Nikola B. Kovachki, Andrew M. Stuart
    Description: Provides a continuous-time analysis of momentum methods in stochastic optimization.

  2. Generalization Performance of Multi-pass Stochastic Gradient Descent with Convex Loss Functions
    Authors: Yunwen Lei, Ting Hu, Ke Tang
    Description: Analyzes generalization performance of multi-pass stochastic gradient descent for convex losses.

  3. High-Order Langevin Diffusion Yields an Accelerated MCMC Algorithm
    Authors: Wenlong Mou, Yi-An Ma, Martin J. Wainwright, Peter L. Bartlett, Michael I. Jordan
    Description: Develops an accelerated MCMC algorithm using high-order Langevin diffusion.

  4. Path Length Bounds for Gradient Descent and Flow
    Authors: Chirag Gupta, Sivaraman Balakrishnan, Aaditya Ramdas
    Description: Establishes path length bounds for gradient descent and flow in stochastic optimization.

  5. Optimization with Momentum: Dynamical, Control-Theoretic, and Symplectic Perspectives
    Authors: Michael Muehlebach, Michael I. Jordan
    Description: Analyzes momentum-based optimization from dynamical, control-theoretic, and symplectic perspectives.

  6. L-SVRG and L-Katyusha with Arbitrary Sampling
    Authors: Xun Qian, Zheng Qu, Peter Richtárik
    Description: Introduces L-SVRG and L-Katyusha algorithms with arbitrary sampling for stochastic optimization.

  7. A Lyapunov Analysis of Accelerated Methods in Optimization
    Authors: Ashia C. Wilson, Ben Recht, Michael I. Jordan
    Description: Provides a Lyapunov analysis for accelerated optimization methods.

  8. NUQSGD: Provably Communication-Efficient Data-Parallel SGD via Nonuniform Quantization
    Authors: Ali Ramezani-Kebrya, Fartash Faghri, Ilya Markov, Vitalii Aksenov, Dan Alistarh, Daniel M. Roy
    Description: Proposes NUQSGD, a communication-efficient stochastic gradient descent method using nonuniform quantization.

  9. An Inertial Newton Algorithm for Deep Learning
    Authors: Camille Castera, Jérôme Bolte, Cédric Févotte, Edouard Pauwels
    Description: Develops an inertial Newton algorithm for deep learning optimization.

  10. Accelerating Ill-Conditioned Low-Rank Matrix Estimation via Scaled Gradient Descent
    Authors: Tian Tong, Cong Ma, Yuejie Chi
    Description: Proposes scaled gradient descent for accelerating ill-conditioned low-rank matrix estimation.

  11. On ADMM in Deep Learning: Convergence and Saturation-Avoidance
    Authors: Jinshan Zeng, Shao-Bo Lin, Yuan Yao, Ding-Xuan Zhou
    Description: Analyzes convergence and saturation-avoidance properties of ADMM in deep learning.

  12. A Unified Convergence Analysis for Shuffling-Type Gradient Methods
    Authors: Lam M. Nguyen, Quoc Tran-Dinh, Dzung T. Phan, Phuong Ha Nguyen, Marten van Dijk
    Description: Provides a unified convergence analysis for shuffling-type gradient methods.

  13. Stochastic Online Optimization Using Kalman Recursion
    Authors: Joseph de Vilmarest, Olivier Wintenberger
    Description: Applies Kalman recursion to stochastic online optimization.

  14. Expanding Boundaries of Gap Safe Screening
    Authors: Cassio F. Dantas, Emmanuel Soubies, Cédric Févotte
    Description: Expands gap safe screening techniques for stochastic optimization.

  15. Consensus-Based Optimization on the Sphere: Convergence to Global Minimizers and Machine Learning
    Authors: Massimo Fornasier, Lorenzo Pareschi, Hui Huang, Philippe Sünnen
    Description: Develops consensus-based optimization on the sphere with applications to machine learning.

  16. Decentralized Stochastic Gradient Langevin Dynamics and Hamiltonian Monte Carlo
    Authors: Mert Gürbüzbalaban, Xuefeng Gao, Yuanhan Hu, Lingjiong Zhu
    Description: Proposes decentralized stochastic gradient Langevin dynamics and Hamiltonian Monte Carlo methods.

Distributed/Decentralized Optimization #

Papers addressing distributed or decentralized optimization algorithms, focusing on communication efficiency and scalability.

  1. Projection-Free Decentralized Online Learning for Submodular Maximization over Time-Varying Networks
    Authors: Junlong Zhu, Qingtao Wu, Mingchuan Zhang, Ruijuan Zheng, Keqin Li
    Description: Develops projection-free decentralized online learning for submodular maximization over time-varying networks.

  2. Communication-Efficient Distributed Covariance Sketch, with Application to Distributed PCA
    Authors: Zengfeng Huang, Xuemin Lin, Wenjie Zhang, Ying Zhang
    Description: Proposes a communication-efficient distributed covariance sketch for distributed PCA.

  3. Optimal Rates of Distributed Regression with Imperfect Kernels
    Authors: Hongwei Sun, Qiang Wu
    Description: Establishes optimal rates for distributed regression with imperfect kernels.

  4. One-Shot Federated Learning: Theoretical Limits and Algorithms to Achieve Them
    Authors: Saber Salehkaleybar, Arsalan Sharifnassab, S. Jamaloddin Golestani
    Description: Analyzes theoretical limits and algorithms for one-shot federated learning.

  5. Cooperative SGD: A Unified Framework for the Design and Analysis of Local-Update SGD Algorithms
    Authors: Jianyu Wang, Gauri Joshi
    Description: Introduces a unified framework for designing and analyzing local-update SGD algorithms.

  6. DeEPCA: Decentralized Exact PCA with Linear Convergence Rate
    Authors: Haishan Ye, Tong Zhang
    Description: Develops DeEPCA, a decentralized exact PCA method with linear convergence.

Submodular Optimization #

Papers focusing on submodular optimization, particularly in experimental design.

  1. Batch Greedy Maximization of Non-Submodular Functions: Guarantees and Applications to Experimental Design
    Authors: Jayanth Jagalur-Mohan, Youssef Marzouk
    Description: Provides guarantees for batch greedy maximization of non-submodular functions with applications to experimental design.

Bandits and Online Learning #

Papers addressing multi-armed bandits, online optimization, and regret minimization.

  1. Regulating Greed Over Time in Multi-Armed Bandits
    Authors: Stefano Tracà, Cynthia Rudin, Weiyu Yan
    Description: Studies methods to regulate greed over time in multi-armed bandits.

  2. Preference-Based Online Learning with Dueling Bandits: A Survey
    Authors: Viktor Bengs, Róbert Busa-Fekete, Adil El Mesaoudi-Paul, Eyke Hüllermeier
    Description: Surveys preference-based online learning with dueling bandits.

  3. On Multi-Armed Bandit Designs for Dose-Finding Trials
    Authors: Maryam Aziz, Emilie Kaufmann, Marie-Karelle Riviere
    Description: Explores multi-armed bandit designs for dose-finding trials.

  4. Tsallis-INF: An Optimal Algorithm for Stochastic and Adversarial Bandits
    Authors: Julian Zimmert, Yevgeny Seldin
    Description: Proposes Tsallis-INF, an optimal algorithm for stochastic and adversarial bandits.

  5. Bandit Convex Optimization in Non-Stationary Environments
    Authors: Peng Zhao, Guanghui Wang, Lijun Zhang, Zhi-Hua Zhou
    Description: Addresses bandit convex optimization in non-stationary environments.

  6. A Contextual Bandit Bake-off
    Authors: Alberto Bietti, Alekh Agarwal, John Langford
    Description: Compares contextual bandit algorithms in a comprehensive evaluation.

  7. MetaGrad: Adaptation Using Multiple Learning Rates in Online Learning
    Authors: Tim van Erven, Wouter M. Koolen, Dirk van der Hoeven
    Description: Introduces MetaGrad, an adaptive online learning algorithm with multiple learning rates.

  8. Achieving Fairness in the Stochastic Multi-Armed Bandit Problem
    Authors: Vishakha Patil, Ganesh Ghalme, Vineet Nair, Y. Narahari
    Description: Develops methods for achieving fairness in stochastic multi-armed bandits.

  9. Refined Approachability Algorithms and Application to Regret Minimization with Global Costs
    Authors: Joon Kwon
    Description: Proposes refined approachability algorithms for regret minimization with global costs.

  10. Bandit Learning in Decentralized Matching Markets
    Authors: Lydia T. Liu, Feng Ruan, Horia Mania, Michael I. Jordan
    Description: Applies bandit learning to decentralized matching markets.

  11. Thompson Sampling Algorithms for Cascading Bandits
    Authors: Zixin Zhong, Wang Chi Chueng, Vincent Y. F. Tan
    Description: Develops Thompson sampling algorithms for cascading bandits.

  12. Fast Learning for Renewal Optimization in Online Task Scheduling
    Authors: Michael J. Neely
    Description: Proposes fast learning methods for renewal optimization in online task scheduling.

Bayesian and Hyperparameter Optimization #

Papers addressing Bayesian optimization and hyperparameter tuning for scalable and robust optimization.

  1. An Empirical Study of Bayesian Optimization: Acquisition Versus Partition
    Authors: Erich Merrill, Alan Fern, Xiaoli Fern, Nima Dolatnia
    Description: Conducts an empirical study comparing acquisition and partition strategies in Bayesian optimization.

  2. Hyperparameter Optimization via Sequential Uniform Designs
    Authors: Zebin Yang, Aijun Zhang
    Description: Proposes sequential uniform designs for hyperparameter optimization.

  3. Are We Forgetting about Compositional Optimisers in Bayesian Optimisation?
    Authors: Antoine Grosnit, Alexander I. Cowen-Rivers, Rasul Tutunov, Ryan-Rhys Griffiths, Jun Wang, Haitham Bou-Ammar
    Description: Explores the role of compositional optimizers in Bayesian optimization.

  4. GIBBON: General-Purpose Information-Based Bayesian Optimisation
    Authors: Henry B. Moss, David S. Leslie, Javier Gonzalez, Paul Rayson
    Description: Introduces GIBBON, a general-purpose information-based Bayesian optimization framework.

  5. On lp-Hyperparameter Learning via Bilevel Nonsmooth Optimization
    Authors: Takayuki Okuno, Akiko Takeda, Akihiro Kawana, Motokazu Watanabe
    Description: Studies lp-hyperparameter learning using bilevel nonsmooth optimization.

Optimization in Reinforcement Learning #

Papers focusing on optimization techniques for reinforcement learning, including policy iteration and Q-learning.

  1. Safe Policy Iteration: A Monotonically Improving Approximate Policy Iteration Approach
    Authors: Alberto Maria Metelli, Matteo Pirotta, Daniele Calandriello, Marcello Restelli
    Description: Proposes a safe policy iteration method with monotonic improvement for reinforcement learning.

  2. On the Theory of Policy Gradient Methods: Optimality, Approximation, and Distribution Shift
    Authors: Alekh Agarwal, Sham M. Kakade, Jason D. Lee, Gaurav Mahajan
    Description: Analyzes the optimality, approximation, and distribution shift in policy gradient methods.

  3. Langevin Dynamics for Adaptive Inverse Reinforcement Learning of Stochastic Gradient Algorithms
    Authors: Vikram Krishnamurthy, George Yin
    Description: Applies Langevin dynamics to adaptive inverse reinforcement learning for stochastic gradient algorithms.

  4. Hamilton-Jacobi Deep Q-Learning for Deterministic Continuous-Time Systems with Lipschitz Continuous Controls
    Authors: Jeongho Kim, Jaeuk Shin, Insoon Yang
    Description: Develops Hamilton-Jacobi deep Q-learning for deterministic continuous-time systems.

  5. Partial Policy Iteration for L1-Robust Markov Decision Processes
    Authors: Chin Pang Ho, Marek Petrik, Wolfram Wiesemann
    Description: Introduces partial policy iteration for L1-robust Markov decision processes.

  6. Gaussian Approximation for Bias Reduction in Q-Learning
    Authors: Carlo D’Eramo, Andrea Cini, Alessandro Nuara, Matteo Pirotta, Cesare Alippi, Jan Peters, Marcello Restelli
    Description: Proposes Gaussian approximation techniques for bias reduction in Q-learning.

Other Optimization Topics #

Papers covering miscellaneous optimization topics, including Newton methods, SVM training, and eigenvector computation.

  1. Global and Quadratic Convergence of Newton Hard-Thresholding Pursuit
    Authors: Shenglong Zhou, Naihua Xiu, Hou-Duo Qi
    Description: Analyzes global and quadratic convergence of Newton hard-thresholding pursuit.

  2. A Two-Level Decomposition Framework Exploiting First and Second Order Information for SVM Training Problems
    Authors: Giulio Galvan, Matteo Lapucci, Chih-Jen Lin, Marco Sciandrone
    Description: Proposes a two-level decomposition framework for SVM training using first and second-order information.

  3. Approximate Newton Methods
    Authors: Haishan Ye, Luo Luo, Zhihua Zhang
    Description: Develops approximate Newton methods for optimization.

  4. A Unified Analysis of First-Order Methods for Smooth Games via Integral Quadratic Constraints
    Authors: Guodong Zhang, Xuchan Bao, Laurent Lessard, Roger Grosse
    Description: Provides a unified analysis of first-order methods for smooth games using integral quadratic constraints.

  5. LassoNet: A Neural Network with Feature Sparsity
    Authors: Ismael Lemhadri, Feng Ruan, Louis Abraham, Robert Tibshirani
    Description: Introduces LassoNet, a neural network architecture promoting feature sparsity.

  6. An Algorithmic View of L2 Regularization and Some Path-Following Algorithms
    Authors: Yunzhang Zhu, Renxiong Liu
    Description: Explores L2 regularization from an algorithmic perspective with path-following algorithms.

  7. The Ensmallen Library for Flexible Numerical Optimization
    Authors: Ryan R. Curtin, Marcus Edel, Rahul Ganesh Prabhu, Suryoday Basak, Zhihao Lou, Conrad Sanderson
    Description: Introduces the ensmallen library for flexible numerical optimization.

  8. Black-Box Reductions for Zeroth-Order Gradient Algorithms to Achieve Lower Query Complexity
    Authors: Bin Gu, Xiyuan Wei, Shangqian Gao, Ziran Xiong, Cheng Deng, Heng Huang
    Description: Proposes black-box reductions for zeroth-order gradient algorithms to reduce query complexity.

  9. On the Riemannian Search for Eigenvector Computation
    Authors: Zhiqiang Xu, Ping Li
    Description: Develops Riemannian search methods for eigenvector computation.

Tags:
Categories: