Open Source Python Neural Network Libraries - Page 2

Python Neural Network Libraries

View 50 business solutions

Browse free open source Python Neural Network Libraries and projects below. Use the toggles on the left to filter open source Python Neural Network Libraries by OS, license, language, programming language, and project status.

  • Your go-to FinOps platform Icon
    Your go-to FinOps platform

    Analyze, optimize, and govern your multi-cloud environment effortlessly with AI Agentic FinOps.

    Unlike reporting-only FinOps tools, FinOpsly unifies cloud (AWS, Azure, GCP), data (Snowflake, Databricks, BigQuery), and AI costs into a single system of action — enabling teams to plan spend before it happens, automate optimization safely, and prove value in weeks, not quarters.
    Learn More
  • Hightouch is a data and AI platform for marketing and personalization. Icon
    Hightouch is a data and AI platform for marketing and personalization.

    Marketing needs data and AI. Give them Hightouch.

    Find insights, run real-time campaigns, and build AI agents with all your data.
    Learn More
  • 1
    CoreNet

    CoreNet

    CoreNet: A library for training deep neural networks

    CoreNet is Apple’s internal deep learning framework for distributed neural network training, designed for high scalability, low-latency communication, and strong hardware efficiency. It focuses on enabling large-scale model training across clusters of GPUs and accelerators by optimizing data flow and parallelism strategies. CoreNet provides abstractions for data, tensor, and pipeline parallelism, allowing models to scale without code duplication or heavy manual configuration. Its distributed runtime manages synchronization, load balancing, and mixed-precision computation to maximize throughput while minimizing communication bottlenecks. CoreNet integrates tightly with Apple’s proprietary ML stack and hardware, serving as the foundation for research in computer vision, language models, and multimodal systems within Apple AI. The framework includes monitoring tools, fault tolerance mechanisms, and efficient checkpointing for massive training runs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    DeepXDE

    DeepXDE

    A library for scientific machine learning & physics-informed learning

    DeepXDE is a library for scientific machine learning and physics-informed learning. DeepXDE includes the following algorithms. Physics-informed neural network (PINN). Solving different problems. Solving forward/inverse ordinary/partial differential equations (ODEs/PDEs) [SIAM Rev.] Solving forward/inverse integro-differential equations (IDEs) [SIAM Rev.] fPINN: solving forward/inverse fractional PDEs (fPDEs) [SIAM J. Sci. Comput.] NN-arbitrary polynomial chaos (NN-aPC): solving forward/inverse stochastic PDEs (sPDEs) [J. Comput. Phys.] PINN with hard constraints (hPINN): solving inverse design/topology optimization [SIAM J. Sci. Comput.] Residual-based adaptive sampling [SIAM Rev., arXiv] Gradient-enhanced PINN (gPINN) [Comput. Methods Appl. Mech. Eng.] PINN with multi-scale Fourier features [Comput. Methods Appl. Mech. Eng.]
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Differentiable Neural Computer

    Differentiable Neural Computer

    A TensorFlow implementation of the Differentiable Neural Computer

    The Differentiable Neural Computer (DNC), developed by Google DeepMind, is a neural network architecture augmented with dynamic external memory, enabling it to learn algorithms and solve complex reasoning tasks. Published in Nature in 2016 under the paper “Hybrid computing using a neural network with dynamic external memory,” the DNC combines the pattern recognition power of neural networks with a memory module that can be written to and read from in a differentiable way. This allows the model to learn how to store and retrieve information across long time horizons, much like a traditional computer. The architecture consists of modular components including an access module for managing memory operations, a controller (often an LSTM or feedforward network) for issuing read/write commands, and submodules for temporal linkage and memory allocation tracking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the new FullyShardedDataParallel (FSDP) wrapper provided by fairscale. Fairseq can be extended through user-supplied plug-ins. Models define the neural network architecture and encapsulate all of the learnable parameters. Criterions compute the loss function given the model outputs and targets. Tasks store dictionaries and provide helpers for loading/iterating over Datasets, initializing the Model/Criterion and calculating the loss.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Automate Proposals with AI in Microsoft Word. Icon
    Automate Proposals with AI in Microsoft Word.

    Streamline proposal creation with the smartest AI, the best content, seamless integration with Microsoft Word, and unmatched efficiency.

    Automate your best practices, processes, and standards to guide your proposal writers, sales teams, and subject experts. And don’t worry, it’s so easy to use they will use it. We would love the opportunity to help you quantify the impact your business can expect from investing in Expedience Software. Click here to request a Return on Investment (ROI) calculation. In this 15-minute session, we will ask 20 simple questions to assess and grade your current proposal quality and scalability. Manual proposal processes are likely costing you far more than you realize. These models waste time and kill the productivity of proposal writers, sales team members, senior staff, and subject experts.
    Learn More
  • 5
    Flax

    Flax

    Flax is a neural network library for JAX

    Flax is a flexible neural-network library for JAX that embraces functional programming while offering ergonomic module abstractions. Its design separates pure computation from state by threading parameter collections and RNGs explicitly, enabling reproducibility, transformation, and easy experimentation with JAX transforms like jit, pmap, and vmap. Modules define parameterized computations, but initialization and application remain side-effect free, which pairs naturally with JAX’s staging and compilation model. Flax emphasizes composability: optimizers, training loops, and checkpointing are provided as examples or utilities rather than monolithic frameworks, encouraging research-friendly customization. The library is widely used in vision, language, and reinforcement learning, often serving as a thin layer atop NumPy-like JAX primitives. Tutorials and examples show patterns for multi-host training, mixed precision, and advanced input pipelines that scale from laptops to TPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Forecasting Best Practices

    Forecasting Best Practices

    Time Series Forecasting Best Practices & Examples

    Time series forecasting is one of the most important topics in data science. Almost every business needs to predict the future in order to make better decisions and allocate resources more effectively. This repository provides examples and best practice guidelines for building forecasting solutions. The goal of this repository is to build a comprehensive set of tools and examples that leverage recent advances in forecasting algorithms to build solutions and operationalize them. Rather than creating implementations from scratch, we draw from existing state-of-the-art libraries and build additional utilities around processing and featuring the data, optimizing and evaluating models, and scaling up to the cloud. The examples and best practices are provided as Python Jupyter notebooks and R markdown files and a library of utility functions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Haiku

    Haiku

    JAX-based neural network library

    Haiku is a library built on top of JAX designed to provide simple, composable abstractions for machine learning research. Haiku is a simple neural network library for JAX that enables users to use familiar object-oriented programming models while allowing full access to JAX’s pure function transformations. Haiku is designed to make the common things we do such as managing model parameters and other model state simpler and similar in spirit to the Sonnet library that has been widely used across DeepMind. It preserves Sonnet’s module-based programming model for state management while retaining access to JAX’s function transformations. Haiku can be expected to compose with other libraries and work well with the rest of JAX. Similar to Sonnet modules, Haiku modules are Python objects that hold references to their own parameters, other modules, and methods that apply functions on user inputs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Jraph

    Jraph

    A Graph Neural Network Library in Jax

    Jraph (pronounced “giraffe”) is a lightweight JAX library developed by Google DeepMind for building and experimenting with graph neural networks (GNNs). It provides an efficient and flexible framework for representing, manipulating, and training models on graph-structured data. The core of Jraph is the GraphsTuple data structure, which enables users to define graphs with arbitrary node, edge, and global attributes, and to batch variable-sized graphs efficiently for JAX’s just-in-time compilation. The library includes a comprehensive set of utilities for batching, padding, masking, and partitioning graph data, making it ideal for distributed and large-scale GNN experiments. Jraph also comes with a model zoo—a collection of forkable reference implementations of common message-passing GNN architectures, such as Graph Networks, Graph Convolutional Networks, and Graph Attention Networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    MMDeploy

    MMDeploy

    OpenMMLab Model Deployment Framework

    MMDeploy is an open-source deep learning model deployment toolset. It is a part of the OpenMMLab project. Models can be exported and run in several backends, and more will be compatible. All kinds of modules in the SDK can be extended, such as Transform for image processing, Net for Neural Network inference, Module for postprocessing and so on. Install and build your target backend. ONNX Runtime is a cross-platform inference and training accelerator compatible with many popular ML/DNN frameworks. Please read getting_started for the basic usage of MMDeploy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • The CRM you will want to use every day Icon
    The CRM you will want to use every day

    With CRM, Sales, and Marketing Automation in one, Act! gives you everything you need for happier clients, more revenue, and less stress.

    Act! Premium is perfect for small and midsize businesses looking to market better, sell more, and create customers for life. With unparalleled flexibility and freedom of choice, Act! Premium accommodates the unique ways you do business. Whether it’s customizations to fit your specific business or industry processes or your preferences for deployment and access, the possibilities with Act! Premium are limitless.
    Learn More
  • 10
    MMdnn

    MMdnn

    Tools to help users inter-operate among deep learning frameworks

    MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g. model conversion and visualization. Convert models between Caffe, Keras, MXNet, Tensorflow, CNTK, PyTorch Onnx and CoreML. MMdnn is a comprehensive and cross-framework tool to convert, visualize and diagnose deep learning (DL) models. The "MM" stands for model management, and "dnn" is the acronym of deep neural network. We implement a universal converter to convert DL models between frameworks, which means you can train a model with one framework and deploy it with another. During the model conversion, we generate some code snippets to simplify later retraining or inference. We provide a model collection to help you find some popular models. We provide a model visualizer to display the network architecture more intuitively. We provide some guidelines to help you deploy DL models to another hardware platform.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MetaNet

    MetaNet

    Free portable library for meta neural network research

    MetaNet provides free library for meta neural network research. MetaNet library contain feed-forward neural net realisation and several integrated dataset (MNIST).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Minkowski Engine

    Minkowski Engine

    Auto-diff neural network library for high-dimensional sparse tensors

    The Minkowski Engine is an auto-differentiation library for sparse tensors. It supports all standard neural network layers such as convolution, pooling, unspooling, and broadcasting operations for sparse tensors. The Minkowski Engine supports various functions that can be built on a sparse tensor. We list a few popular network architectures and applications here. To run the examples, please install the package and run the command in the package root directory. Compressing a neural network to speed up inference and minimize memory footprint has been studied widely. One of the popular techniques for model compression is pruning the weights in convnets, is also known as sparse convolutional networks. Such parameter-space sparsity used for model compression compresses networks that operate on dense tensors and all intermediate activations of these networks are also dense tensors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    NLP Architect

    NLP Architect

    A model library for exploring state-of-the-art deep learning

    NLP Architect is an open-source Python library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing and Natural Language Understanding neural networks. The library includes our past and ongoing NLP research and development efforts as part of Intel AI Lab. NLP Architect is designed to be flexible for adding new models, neural network components, data handling methods, and for easy training and running models. NLP Architect is a model-oriented library designed to showcase novel and different neural network optimizations. The library contains NLP/NLU-related models per task, different neural network topologies (which are used in models), procedures for simplifying workflows in the library, pre-defined data processors and dataset loaders and misc utilities. The library is designed to be a tool for model development: data pre-processing, build model, train, validate, infer, save or load a model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Nerfies

    Nerfies

    This is the code for Deformable Neural Radiance Fields

    Nerfies demonstrates deformation-aware neural radiance fields that reconstruct and render dynamic, real-world scenes from casual video. Instead of assuming a static world, the method learns a canonical space plus a deformation field that maps changing poses or expressions back to that space during training. This lets the system generate photorealistic novel views of nonrigid subjects—faces, bodies, cloth—while preserving fine detail and consistent lighting. The training pipeline handles imperfect captures by modeling camera poses, exposure variations, and background segmentation, producing stable geometry and appearance. A set of utilities manages dataset preparation, pose estimation, and checkpoints so researchers can reproduce results on their own footage. The work sits at the intersection of graphics and vision, showing how learned volumetric rendering can handle human motion without dense markers or studio rigs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    NeuMan

    NeuMan

    Neural Human Radiance Field from a Single Video (ECCV 2022)

    NeuMan is a reference implementation that reconstructs both an animatable human and its background scene from a single monocular video using neural radiance fields. It supports novel view and novel pose synthesis, enabling compositional results like transferring reconstructed humans into new scenes. The pipeline separates human/body and environment, learning consistent geometry and appearance to support animation. Demos showcase sequences such as dance and handshake, and the code provides guidance for running evaluations and rendering. As a research release, it serves both as a baseline and as a starting point for work on human-centric NeRFs. The emphasis is on practical reconstruction quality from minimal capture setups. Compositional outputs to blend humans and backgrounds. Novel view and novel pose synthesis from learned radiance fields.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Neural Libs

    Neural Libs

    Neural network library for developers

    This project includes the implementation of a neural network MLP, RBF, SOM and Hopfield networks in several popular programming languages. The project also includes examples of the use of neural networks as function approximation and time series prediction. Includes a special program makes it easy to test neural network based on training data and the optimization of the network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Neural Network Visualization

    Neural Network Visualization

    Project for processing neural networks and rendering to gain insights

    nn_vis is a minimalist visualization tool for neural networks written in Python using OpenGL and Pygame. It provides an interactive, graphical representation of how data flows through neural network layers, offering a unique educational experience for those new to deep learning or looking to explain it visually. By animating input, weights, activations, and outputs, the tool demystifies neural network operations and helps users intuitively grasp complex concepts. Its lightweight codebase is great for customization and teaching purposes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. Open source, modular API for differential privacy research. Everyone is welcome to contribute. ML practitioners will find this to be a gentle introduction to training a model with differential privacy as it requires minimal code changes. Differential Privacy researchers will find this easy to experiment and tinker with, allowing them to focus on what matters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Penzai

    Penzai

    A JAX research toolkit to build, edit, & visualize neural networks

    Penzai, developed by Google DeepMind, is a JAX-based library for representing, visualizing, and manipulating neural network models as functional pytree data structures. It is designed to make machine learning research more interpretable and interactive, particularly for tasks like model surgery, ablation studies, architecture debugging, and interpretability research. Unlike conventional neural network libraries, Penzai exposes the full internal structure of models, enabling fine-grained inspection and modification after training. Its modular design includes tools for tree manipulation, named axes, and declarative neural network construction. The library integrates tightly with Treescope, an advanced pretty-printer for visualizing deeply nested JAX pytrees and NDArray structures. Penzai’s penzai.nn module provides a compositional, combinator-based API for building neural networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    PyTorch Book

    PyTorch Book

    PyTorch tutorials and fun projects including neural talk

    This is the corresponding code for the book "The Deep Learning Framework PyTorch: Getting Started and Practical", but it can also be used as a standalone PyTorch Getting Started Guide and Tutorial. The current version of the code is based on pytorch 1.0.1, if you want to use an older version please git checkout v0.4or git checkout v0.3. Legacy code has better python2/python3 compatibility, CPU/GPU compatibility test. The new version of the code has not been fully tested, it has been tested under GPU and python3. But in theory there shouldn't be too many problems on python2 and CPU. The basic part (the first five chapters) explains the content of PyTorch. This part introduces the main modules in PyTorch and some tools commonly used in deep learning. For this part of the content, Jupyter Notebook is used as a teaching tool here, and readers can modify and run with notebooks and repeat experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    A neural net module written in python. The aim of the project is to provide a large set of neural network types accessed by an API that is easy to use and powerful.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    What is QPF 2.6 ? QPF 2.6 (or Quantum Programming framework 2.6) is a free simple and easy to use framework dedicated to supporting programmers who are developing software for the D-wave one series of quantum computers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    RNNLIB is a recurrent neural network library for sequence learning problems. Applicable to most types of spatiotemporal data, it has proven particularly effective for speech and handwriting recognition. full installation and usage instructions given at http://sourceforge.net/p/rnnl/wiki/Home/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    SVoice (Speech Voice Separation)

    SVoice (Speech Voice Separation)

    We provide a PyTorch implementation of the paper Voice Separation

    SVoice is a PyTorch-based implementation of Facebook Research’s study on speaker voice separation as described in the paper “Voice Separation with an Unknown Number of Multiple Speakers.” This project presents a deep learning framework capable of separating mixed audio sequences where several people speak simultaneously, without prior knowledge of how many speakers are present. The model employs gated neural networks with recurrent processing blocks that disentangle voices over multiple computational steps, while maintaining speaker consistency across output channels. Separate models are trained for different speaker counts, and the largest-capacity model dynamically determines the actual number of speakers in a mixture. The repository includes all necessary scripts for training, dataset preparation, distributed training, evaluation, and audio separation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    TFLearn

    TFLearn

    Deep learning library featuring a higher-level API for TensorFlow

    TFlearn is a modular and transparent deep learning library built on top of Tensorflow. It was designed to provide a higher-level API to TensorFlow in order to facilitate and speed up experimentations while remaining fully transparent and compatible with it. Easy-to-use and understand high-level API for implementing deep neural networks, with tutorials and examples. Fast prototyping through highly modular built-in neural network layers, regularizers, optimizers, and metrics. Full transparency over Tensorflow. All functions are built over tensors and can be used independently of TFLearn. Powerful helper functions to train any TensorFlow graph, with support of multiple inputs, outputs, and optimizers. Easy and beautiful graph visualization, with details about weights, gradients, activations, and more. Effortless device placement for using multiple CPU/GPU. The high-level API currently supports the most of the recent deep learning models, such as Convolutions, LSTM, BiRNN, BatchNorm, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB