DiffSharp 0.6.0
dotnet add package DiffSharp --version 0.6.0
NuGet\Install-Package DiffSharp -Version 0.6.0
<PackageReference Include="DiffSharp" Version="0.6.0" />
paket add DiffSharp --version 0.6.0
#r "nuget: DiffSharp, 0.6.0"
// Install DiffSharp as a Cake Addin #addin nuget:?package=DiffSharp&version=0.6.0 // Install DiffSharp as a Cake Tool #tool nuget:?package=DiffSharp&version=0.6.0
DiffSharp is an automatic differentiation (AD) library implemented in the F# language.
AD allows exact and efficient calculation of derivatives, by systematically applying the chain rule of calculus at the elementary operator level. AD is different from numerical differentiation, which is prone to truncation and round-off errors, and symbolic differentiation, which is exact but not efficient for run-time calculations and can only handle closed-form mathematical expressions.
Using the DiffSharp library, derivative calculations (gradients, Hessians, Jacobians, directional derivatives, and matrix-free Hessian- and Jacobian-vector products) can be incorporated with minimal change into existing algorithms. Operations can be nested to any level, meaning that you can compute exact higher-order derivatives and differentiate functions that are internally making use of differentiation. Please see the API Overview page for a list of available operations.
The library is under active development by Atılım Güneş Baydin and Barak A. Pearlmutter mainly for research applications in machine learning, as part of their work at the Brain and Computation Lab, Hamilton Institute, National University of Ireland Maynooth.
Product | Versions Compatible and additional computed target framework versions. |
---|---|
.NET Framework | net is compatible. |
-
- FsAlg (>= 0.5.4)
- FSharp.Quotations.Evaluator (>= 1.0.6)
GitHub repositories
This package is not used by any popular GitHub repositories.
Please visit
https://github.com/gbaydin/DiffSharp/releases
for the latest release notes.