autograd 1.5-foss-2021bAutograd can automatically differentiate native Python and Numpy code. It can handle a large subset of Python's features, including loops, ifs, recursion and closures, and it can even take derivatives of derivatives of derivatives. It supports reverse-mode differentiation (a.k.a. backpropagation), which means it can efficiently take gradients of scalar-valued functions with respect to array-valued arguments, as well as forward-mode differentiation, and the two can be composed arbitrarily. The main intended application of Autograd is gradient-based optimization.
Accessing autograd 1.5-foss-2021b
To load the module for autograd 1.5-foss-2021b please use this command on the BEAR systems (BlueBEAR, BEARCloud VMs, and CaStLeS VMs):
module load bear-apps/2021b
module load autograd/1.5-foss-2021b
BEAR Apps Version
The listed architectures consist of two part: OS-CPU.
- BlueBEAR: The OS used on BlueBEAR is represented by EL and there are several different processor (CPU) types available on BlueBEAR. More information about the processor types on BlueBEAR is available on the BlueBEAR Job Submission page.
- BEAR and CaStLeS Cloud VMs: These VMs can have one of two OSes. Those with access to a BEAR Cloud or CaStLeS VM should check that the listed architectures for an application include the OS of VM being used. The VMs, irrespective of OS, will use the haswell CPU type.
For more information visit the autograd website.
Last modified on 30th November 2022