Skip to main content
Article
Zeroth-Order Hard-Thresholding: Gradient Error vs. Expansivity
Advances in Neural Information Processing Systems
  • William de Vazelhes, Mohamed Bin Zayed University of Artificial Intelligence
  • Hualin Zhang, Nanjing University of Information Science & Technology
  • Huimin Wu, Nanjing University of Information Science & Technology
  • Xiao Tong Yuan, Nanjing University of Information Science & Technology
  • Bin Gu, Mohamed Bin Zayed University of Artificial Intelligence
Document Type
Conference Proceeding
Abstract

ℓ0 constrained optimization is prevalent in machine learning, particularly for high-dimensional problems, because it is a fundamental approach to achieve sparse learning. Hard-thresholding gradient descent is a dominant technique to solve this problem. However, first-order gradients of the objective function may be either unavailable or expensive to calculate in a lot of real-world problems, where zeroth-order (ZO) gradients could be a good surrogate. Unfortunately, whether ZO gradients can work with the hard-thresholding operator is still an unsolved problem. To solve this puzzle, in this paper, we focus on the ℓ0 constrained black-box stochastic optimization problems, and propose a new stochastic zeroth-order gradient hard-thresholding (SZOHT) algorithm with a general ZO gradient estimator powered by a novel random support sampling. We provide the convergence analysis of SZOHT under standard assumptions. Importantly, we reveal a conflict between the deviation of ZO estimators and the expansivity of the hard-thresholding operator, and provide a theoretical minimal value of the number of random directions in ZO gradients. In addition, we find that the query complexity of SZOHT is independent or weakly dependent on the dimensionality under different settings. Finally, we illustrate the utility of our method on a portfolio optimization problem as well as black-box adversarial attacks.

Publication Date
11-1-2022
Comments

Access available at NeurIPS Proceedings

Citation Information
W. de Vazelhes, H. Zhang, H. Wu, X. Yuan, and B. Gu, "Zeroth-Order Hard-Thresholding: Gradient Error vs. Expansivity", in 36th Conference on Neural Information Processing Systems, NeurIPS 2022, Advances in Neural Information Processing Systems, vol 35, Nov 2022. https://proceedings.neurips.cc/paper_files/paper/2022/file/8de5384f522efff26884559599c09312-Paper-Conference.pdf