Skip to main content
Article
Rare Gems: Finding Lottery Tickets at Initialization
Advances in Neural Information Processing Systems
  • Kartik Sreenivasan, University of Wisconsin-Madison
  • Jy Yong Sohn, University of Wisconsin-Madison
  • Liu Yang, University of Wisconsin-Madison
  • Matthew Grinde, University of Wisconsin-Madison
  • Alliot Nagle, University of Wisconsin-Madison
  • Hongyi Wang, Carnegie Mellon University
  • Eric Xing, Carnegie Mellon University & Mohamed bin Zayed University of Artificial Intelligence
  • Kangwook Lee, University of Wisconsin-Madison
  • Dimitris Papailiopoulos, University of Wisconsin-Madison
Document Type
Conference Proceeding
Abstract

Large neural networks can be pruned to a small fraction of their original size, with little loss in accuracy, by following a time-consuming “train, prune, re-train” approach. Frankle & Carbin [9] conjecture that we can avoid this by training lottery tickets, i.e., special sparse subnetworks found at initialization, that can be trained to high accuracy. However, a subsequent line of work [11, 41] presents concrete evidence that current algorithms for finding trainable networks at initialization, fail simple baseline comparisons, e.g., against training random sparse subnetworks. Finding lottery tickets that train to better accuracy compared to simple baselines remains an open problem. In this work, we resolve this open problem by proposing GEM-MINER which finds lottery tickets at initialization that beat current baselines. GEM-MINER finds lottery tickets trainable to accuracy competitive or better than Iterative Magnitude Pruning (IMP), and does so up to 19× faster.

Publication Date
12-1-2022
Keywords
  • Gems,
  • Neural-networks,
  • subnetworks,
  • Simple++
Comments

Conference proceedings available at NeurIPS

Archived, thanks to NeurIPS

Uploaded 18 January 2024

Citation Information
K. Sreenivasan, et al, "Rare Gems: Finding Lottery Tickets at Initialization", in 36th Conf. on Nueral Information Processing Systems (NeurIPS 2022), New Orleands, Dec 2022. available at: https://proceedings.neurips.cc/paper_files/paper/2022/file/5d52b102ebd672023628cac20e9da5ff-Paper-Conference.pdf