Comparison of two gradient computation methods in PythonDownload PDF

Sri Hari Krishna Narayanan, Paul Hovland, Kshitij Kulshreshtha, Devashri Nagarkar, Kaitlyn MacIntyre, Riley Wagner, Deqing Fu

28 Oct 2017NIPS 2017 Workshop Autodiff SubmissionReaders: Everyone
  • Keywords: Automatic Differentiation, Python, autograd, adolc, performance comparison
  • TL;DR: Compares the performance and features of two automatic differentiation tools, autograd and ADOL-C
  • Abstract: Gradient based optimization and machine learning applications require the computation of derivatives. For example, artificial neural networks (ANNs), a widely used learning system, use backpropagation to calculate the error contribution of each neuron after a batch of data is processed. Languages such as Python and R are are popular for machine learning. Therefore, there has been interest in the last few years to build tools for Python and R to compute derivatives.
3 Replies

Loading