Search Results for "automatic differentiation"

Showing 74 open source projects for "automatic differentiation"

View related business solutions
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • Turn traffic into pipeline and prospects into customers Icon
    Turn traffic into pipeline and prospects into customers

    For account executives and sales engineers looking for a solution to manage their insights and sales data

    Docket is an AI-powered sales enablement platform designed to unify go-to-market (GTM) data through its proprietary Sales Knowledge Lake™ and activate it with intelligent AI agents. The platform helps marketing teams increase pipeline generation by 15% by engaging website visitors in human-like conversations and qualifying leads. For sales teams, Docket improves seller efficiency by 33% by providing instant product knowledge, retrieving collateral, and creating personalized documents. Built for GTM teams, Docket integrates with over 100 tools across the revenue tech stack and offers enterprise-grade security with SOC 2 Type II, GDPR, and ISO 27001 compliance. Customers report improved win rates, shorter sales cycles, and dramatically reduced response times. Docket’s scalable, accurate, and fast AI agents deliver reliable answers with confidence scores, empowering teams to close deals faster.
    Learn More
  • 1
    ImplicitDifferentiation.jl

    ImplicitDifferentiation.jl

    Automatic differentiation of implicit functions

    ImplicitDifferentiation.jl is a package for automatic differentiation of functions defined implicitly, i.e., forward mappings. Those for which automatic differentiation fails. Reasons can vary depending on your backend, but the most common include calls to external solvers, mutating operations or type restrictions. Those for which automatic differentiation is very slow. A common example is iterative procedures like fixed point equations or optimization algorithms.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    ForwardDiff.jl

    ForwardDiff.jl

    Forward Mode Automatic Differentiation for Julia

    ForwardDiff implements methods to take derivatives, gradients, Jacobians, Hessians, and higher-order derivatives of native Julia functions (or any callable object, really) using forward mode automatic differentiation (AD). While performance can vary depending on the functions you evaluate, the algorithms implemented by ForwardDiff generally outperform non-AD algorithms (such as finite-differencing) in both speed and accuracy. Functions like f which map a vector to a scalar are the best case for reverse-mode automatic differentiation, but ForwardDiff may still be a good choice if x is not too large, as it is much simpler. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    StructuralEquationModels.jl

    StructuralEquationModels.jl

    A fast and flexible Structural Equation Modelling Framework

    ...We provide fast objective functions, gradients, and for some cases hessians as well as approximations thereof. As a user, you can easily define custom loss functions. For those, you can decide to provide analytical gradients or use finite difference approximation / automatic differentiation. You can choose to mix loss functions natively found in this package and those you provide. In such cases, you optimize over a sum of different objectives (e.g. ML + Ridge). This strategy also applies to gradients, where you may supply analytic gradients or opt for automatic differentiation or mixed analytical and automatic differentiation. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    VoronoiFVM.jl

    VoronoiFVM.jl

    Solution of nonlinear multiphysics partial differential equations

    Solver for coupled nonlinear partial differential equations (elliptic-parabolic conservation laws) based on the Voronoi finite volume method. It uses automatic differentiation via ForwardDiff.jl and DiffResults.jl to evaluate user functions along with their jacobians and calculate derivatives of solutions with respect to their parameters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • The AI workplace management platform Icon
    The AI workplace management platform

    Plan smart spaces, connect teams, manage assets, and get insights with the leading AI-powered operating system for the built world.

    By combining AI workflows, predictive intelligence, and automated insights, OfficeSpace gives leaders a complete view of how their spaces are used and how people work. Facilities, IT, HR, and Real Estate teams use OfficeSpace to optimize space utilization, enhance employee experience, and reduce portfolio costs with precision.
    Learn More
  • 5
    ReverseDiff

    ReverseDiff

    Reverse Mode Automatic Differentiation for Julia

    ReverseDiff is a fast and compile-able tape-based reverse mode automatic differentiation (AD) that implements methods to take gradients, Jacobians, Hessians, and higher-order derivatives of native Julia functions (or any callable object, really). While performance can vary depending on the functions you evaluate, the algorithms implemented by ReverseDiff generally outperform non-AD algorithms in both speed and accuracy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Enzyme.jl

    Enzyme.jl

    Julia bindings for the Enzyme automatic differentiator

    This is a package containing the Julia bindings for Enzyme. This is very much a work in progress and bug reports/discussion is greatly appreciated. Enzyme is a plugin that performs automatic differentiation (AD) of statically analyzable LLVM. It is highly-efficient and its ability perform AD on optimized code allows Enzyme to meet or exceed the performance of state-of-the-art AD tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Transformers.jl

    Transformers.jl

    Julia Implementation of Transformer models

    ...Inspired by architectures like BERT, GPT, and T5, the library offers a modular and flexible interface for building, training, and using transformer-based deep learning models. It supports training from scratch and fine-tuning pretrained models, and integrates with Flux.jl for automatic differentiation and optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Taichi

    Taichi

    Productive, portable, and performant GPU programming in Python

    ...It uses JIT compilation (via LLVM and its runtime TiRT) to offload compute-heavy code to CPUs, GPUs, mobile devices, and embedded systems. With built-in support for sparse data structures (SNode), automatic differentiation, AOT deployment, and compatibility with CUDA, Vulkan, Metal, and OpenGL ES, it empowers disciplines like simulation, graphics, AI, and robotics
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    ChainRulesCore

    ChainRulesCore

    AD-backend agnostic system defining custom forward and reverse rules

    ...The ChainRulesCore package provides a light-weight dependency for defining sensitivities for functions in your packages, without you needing to depend on ChainRules itself. This will allow your package to be used with ChainRules.jl, which aims to provide a variety of common utilities that can be used by downstream automatic differentiation (AD) tools to define and execute forward-, reverse-, and mixed-mode primitives.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Award-Winning Medical Office Software Designed for Your Specialty Icon
    Award-Winning Medical Office Software Designed for Your Specialty

    Succeed and scale your practice with cloud-based, data-backed, AI-powered healthcare software.

    RXNT is an ambulatory healthcare technology pioneer that empowers medical practices and healthcare organizations to succeed and scale through innovative, data-backed, AI-powered software.
    Learn More
  • 10
    Yao

    Yao

    Extensible, Efficient Quantum Algorithm Design for Humans

    An intermediate representation to construct and manipulate your quantum circuit and let you make own abstractions on the quantum circuit in native Julia. Yao supports both forward-mode (faithful gradient) and reverse-mode automatic differentiation with its builtin engine optimized specifically for quantum circuits. Top performance for quantum circuit simulations. Its CUDA backend and batched quantum register support can make typical quantum circuits even faster. Yao is designed to be extensible. Its hierarchical architecture allows you to extend the framework to support and share your new algorithm and hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Hasktorch

    Hasktorch

    Tensors and neural networks in Haskell

    Hasktorch is a powerful Haskell library for tensor computation and neural network modeling, built on top of libtorch (the backend of PyTorch). It brings differentiable programming, automatic differentiation, and efficient tensor operations into Haskell’s strongly typed functional paradigm. This project is in active development, so expect changes to the library API as it evolves. We would like to invite new users to join our Hasktorch discord space for questions and discussions. Contributions/PR are encouraged. Hasktorch is a library for tensors and neural networks in Haskell. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    DynamicHMC

    DynamicHMC

    Implementation of robust dynamic Hamiltonian Monte Carlo methods

    ...In contrast to frameworks that utilize a directed acyclic graph to build a posterior for a Bayesian model from small components, this package requires that you code a log-density function of the posterior in Julia. Derivatives can be provided manually, or using automatic differentiation. Consequently, this package requires that the user is comfortable with the basics of the theory of Bayesian inference, to the extent of coding a (log) posterior density in Julia. This approach allows the use of standard tools like profiling and benchmarking to optimize its performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ChainRules.jl

    ChainRules.jl

    Forward and reverse mode automatic differentiation primitives

    The ChainRules package provides a variety of common utilities that can be used by downstream automatic differentiation (AD) tools to define and execute forward-, reverse--, and mixed-mode primitives. The core logic of ChainRules is implemented in ChainRulesCore.jl. To add ChainRules support to your package, by defining new rules or frules, you only need to depend on the very light-weight package ChainRulesCore.jl. This repository contains ChainRules.jl, which is what people actually use directly. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ggml

    ggml

    Tensor library for machine learning

    ggml is an open-source tensor library designed for efficient machine learning computation with a focus on running models locally and with minimal dependencies. Written primarily in C and C++, the library provides low-level tensor operations and automatic differentiation that allow developers to implement machine learning algorithms and neural networks efficiently. The project emphasizes portability and performance, enabling machine learning inference across a wide range of hardware environments including CPUs and specialized accelerators. It is widely used as a foundational component in projects that run large language models locally, including tools that perform inference for transformer-based models. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Diffrax

    Diffrax

    Numerical differential equation solvers in JAX

    ...The project provides tools for solving ordinary differential equations, stochastic differential equations, controlled differential equations, and related systems in a way that fits naturally into modern machine learning and differentiable programming workflows. Because it is written to work closely with JAX, it supports just-in-time compilation, automatic differentiation, vectorization, and accelerator-backed execution on hardware such as GPUs and TPUs. This makes it especially appealing for researchers who need equation solvers that can be embedded inside trainable models or simulation-heavy learning systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    PyTorch

    PyTorch

    Open source machine learning framework

    PyTorch is a Python package that offers Tensor computation (like NumPy) with strong GPU acceleration and deep neural networks built on tape-based autograd system. This project allows for fast, flexible experimentation and efficient production. PyTorch consists of torch (Tensor library), torch.autograd (tape-based automatic differentiation library), torch.jit (a compilation stack [TorchScript]), torch.nn (neural networks library), torch.multiprocessing (Python multiprocessing), and torch.utils (DataLoader and other utility functions). PyTorch can be used as a replacement for Numpy, or as a deep learning research platform that provides optimum flexibility and speed.
    Downloads: 106 This Week
    Last Update:
    See Project
  • 17
    JAX

    JAX

    Composable transformations of Python+NumPy programs

    ...But JAX also lets you just-in-time compile your own Python functions into XLA-optimized kernels using a one-function API, jit. Compilation and automatic differentiation can be composed arbitrarily, so you can express sophisticated algorithms and get maximal performance without leaving Python.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    NNlib.jl

    NNlib.jl

    Neural Network primitives with multiple backends

    This package provides a library of functions useful for neural networks, such as softmax, sigmoid, batched multiplication, convolutions and pooling. Many of these are used by Flux.jl, which loads this package, but they may be used independently.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    ...TensorFlow Quantum integrates with the Cirq quantum computing framework to define and manipulate quantum circuits, while leveraging TensorFlow’s infrastructure for optimization, automatic differentiation, and large-scale computation. The library also supports high-performance simulation of quantum circuits, enabling researchers to test and evaluate quantum models even without direct access to quantum hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Shumai

    Shumai

    Fast Differentiable Tensor Library in JavaScript & TypeScript with Bun

    ...It provides a high-performance framework for numerical computing and machine learning within modern JavaScript runtimes. Built on Bun and Flashlight, with ArrayFire as its numerical backend, Shumai brings GPU-accelerated tensor operations, automatic differentiation, and scientific computing tools directly to JavaScript developers. It allows seamless integration of machine learning, deep learning, and custom differentiable programs into web-based or server-side environments without relying on Python frameworks. The library supports matrix operations, gradient computation, and tensor conversions with intuitive APIs and near-native speed, thanks to Bun’s low-overhead FFI bindings. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    CasADi

    CasADi

    CasADi is a symbolic framework for numeric optimization

    CasADi is a symbolic framework for numeric optimization implementing automatic differentiation in forward and reverse modes on sparse matrix-valued computational graphs. It supports self-contained C-code generation and interfaces state-of-the-art codes such as SUNDIALS, IPOPT, etc. It can be used in C++, Python, or Matlab/Octave. CasADi's backbone is a symbolic framework implementing forward and reverse modes of AD on expression graphs to construct gradients, large-and-sparse Jacobians, and Hessians. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    PETSc.jl

    PETSc.jl

    Julia wrappers for the PETSc library

    This package provides a low level interface for PETSc and allows combining julia features (such as automatic differentiation) with the PETSc infrastructure and nonlinear solvers.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Tequila

    Tequila

    A High-Level Abstraction Framework for Quantum Algorithms

    Tequila is an abstraction framework for (variational) quantum algorithms. It operates on abstract data structures allowing the formulation, combination, automatic differentiation and optimization of generalized objectives. Tequila can execute the underlying quantum expectation values on state-of-the-art simulators as well as on real quantum devices.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    NonlinearSolve.jl

    NonlinearSolve.jl

    High-performance and differentiation-enabled nonlinear solvers

    ...The package includes its own high-performance nonlinear solvers which include the ability to swap out to fast direct and iterative linear solvers, along with the ability to use sparse automatic differentiation for Jacobian construction and Jacobian-vector products. NonlinearSolve.jl interfaces with other packages of the Julia ecosystem to make it easy to test alternative solver packages and pass small types to control algorithm swapping. It also interfaces with the ModelingToolkit.jl world of symbolic modeling to allow for automatically generating high-performance code.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    CoordinateTransformations.jl

    CoordinateTransformations.jl

    A fresh approach to coordinate transformations

    ...Transformations can be easily applied, inverted, composed, and differentiated (both with respect to the input coordinates and with respect to transformation parameters such as rotation angle). Transformations are designed to be light-weight and efficient enough for, e.g., real-time graphical applications, while support for both explicit and automatic differentiation makes it easy to perform optimization and therefore ideal for computer vision applications such as SLAM (simultaneous localization and mapping).
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB