Efficiently Learning One Hidden Layer ReLU Networks From QueriesDownload PDF

21 May 2021, 20:51 (edited 26 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: PAC learning, polynomial-time algorithms, neural networks, query learning, model extraction
  • TL;DR: We give the first provable, polynomial-time algorithm for learning two-layer neural networks from queries.
  • Abstract: While the problem of PAC learning neural networks from samples has received considerable attention in recent years, in certain settings like model extraction attacks, it is reasonable to imagine having more than just the ability to observe random labeled examples. Motivated by this, we consider the following problem: given \emph{black-box query access} to a neural network $F$, recover $F$ up to some error. Formally, we show that if $F$ is an arbitrary one hidden layer neural network with ReLU activations, there is an algorithm with query complexity and runtime polynomial in all parameters which outputs a network $F’$ achieving low square loss relative to $F$ with respect to the Gaussian measure. While a number of works in the security literature have proposed and empirically demonstrated the effectiveness of certain algorithms for this problem, ours is to the best of our knowledge the first provable guarantee in this vein.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
12 Replies