Open Source Machine Learning Software - Page 5

Machine Learning Software

View 446 business solutions
  • Failed Payment Recovery for Subscription Businesses Icon
    Failed Payment Recovery for Subscription Businesses

    For subscription companies searching for a failed payment recovery solution to grow revenue, and retain customers.

    FlexPay’s innovative platform uses multiple technologies to achieve the highest number of retained customers, resulting in reduced involuntary churn, longer life span after recovery, and higher revenue. Leading brands like LegalZoom, Hooked on Phonics, and ClinicSense trust FlexPay to recover failed payments, reduce churn, and increase customer lifetime value.
    Learn More
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • 1
    Aim

    Aim

    An easy-to-use & supercharged open-source experiment tracker

    Aim logs all your AI metadata (experiments, prompts, etc) enabling a UI to compare & observe them and SDK to query them programmatically. The Aim standard package comes with all integrations. If you'd like to modify the integration and make it custom, create a new integration package and share with others. Aim is an open-source, self-hosted AI Metadata tracking tool designed to handle 100,000s of tracked metadata sequences. The two most famous AI metadata applications are: experiment tracking and prompt engineering. Aim provides a performant and beautiful UI for exploring and comparing training runs, and prompt sessions.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 2
    AtomAI

    AtomAI

    Deep and Machine Learning for Microscopy

    AtomAI is a Pytorch-based package for deep and machine-learning analysis of microscopy data that doesn't require any advanced knowledge of Python or machine learning. The intended audience is domain scientists with a basic understanding of how to use NumPy and Matplotlib. It was developed by Maxim Ziatdinov at Oak Ridge National Lab. The purpose of the AtomAI is to provide an environment that bridges the instrument-specific libraries and general physical analysis by enabling the seamless deployment of machine learning algorithms including deep convolutional neural networks, invariant variational autoencoders, and decomposition/unmixing techniques for image and hyperspectral data analysis. Ultimately, it aims to combine the power and flexibility of the PyTorch deep learning framework and the simplicity and intuitive nature of packages such as scikit-learn, with a focus on scientific data.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 3
    CVPR 2026

    CVPR 2026

    Collection of CVPR 2026 Papers and Open Source Projects

    CVPR2026-Papers-with-Code is a community-maintained repository that collects research papers and corresponding open-source implementations from the CVPR 2026 conference and related computer vision research. The repository acts as a continuously updated catalog of cutting-edge research across a wide range of topics including computer vision, multimodal AI, generative models, diffusion systems, autonomous driving, medical imaging, and remote sensing. Each entry typically links to the research paper as well as the public code repository associated with the work, allowing researchers and developers to quickly access reproducible implementations. The project serves as a centralized index that makes it easier for practitioners to explore the latest advances presented at major computer vision conferences. In addition to the current CVPR cycle, the repository also references related lists covering earlier conferences such as ECCV and ICCV, creating a broader archive of vision research.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 4
    DeepLabCut

    DeepLabCut

    Implementation of DeepLabCut

    DeepLabCut™ is an efficient method for 2D and 3D markerless pose estimation based on transfer learning with deep neural networks that achieves excellent results (i.e. you can match human labeling accuracy) with minimal training data (typically 50-200 frames). We demonstrate the versatility of this framework by tracking various body parts in multiple species across a broad collection of behaviors. The package is open source, fast, robust, and can be used to compute 3D pose estimates or for multi-animals. Please see the original paper and the latest work below! This package is collaboratively developed by the Mathis Group & Mathis Lab at EPFL (releases prior to 2.1.9 were developed at Harvard University). The code is freely available and easy to install in a few clicks with Anaconda (and pypi). DeepLabCut is an open-source Python package for animal pose estimation.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Skillfully - The future of skills based hiring Icon
    Skillfully - The future of skills based hiring

    Realistic Workplace Simulations that Show Applicant Skills in Action

    Skillfully transforms hiring through AI-powered skill simulations that show you how candidates actually perform before you hire them. Our platform helps companies cut through AI-generated resumes and rehearsed interviews by validating real capabilities in action. Through dynamic job specific simulations and skill-based assessments, companies like Bloomberg and McKinsey have cut screening time by 50% while dramatically improving hire quality.
    Learn More
  • 5
    Diffgram

    Diffgram

    Training data (data labeling, annotation, workflow) for all data types

    From ingesting data to exploring it, annotating it, and managing workflows. Diffgram is a single application that will improve your data labeling and bring all aspects of training data under a single roof. Diffgram is world’s first truly open source training data platform that focuses on giving its users an unlimited experience. This is aimed to reduce your data labeling bills and increase your Training Data Quality. Training Data is the art of supervising machines through data. This includes the activities of annotation, which produces structured data; ready to be consumed by a machine learning model. Annotation is required because raw media is considered to be unstructured and not usable without it. That’s why training data is required for many modern machine learning use cases including computer vision, natural language processing and speech recognition.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 6
    DoWhy

    DoWhy

    DoWhy is a Python library for causal inference

    DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks. Much like machine learning libraries have done for prediction, DoWhy is a Python library that aims to spark causal thinking and analysis. DoWhy provides a wide variety of algorithms for effect estimation, causal structure learning, diagnosis of causal structures, root cause analysis, interventions and counterfactuals. DoWhy builds on two of the most powerful frameworks for causal inference: graphical causal models and potential outcomes. For effect estimation, it uses graph-based criteria and do-calculus for modeling assumptions and identifying a non-parametric causal effect. For estimation, it switches to methods based primarily on potential outcomes.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 7
    DocTR

    DocTR

    Library for OCR-related tasks powered by Deep Learning

    DocTR provides an easy and powerful way to extract valuable information from your documents. Seemlessly process documents for Natural Language Understanding tasks: we provide OCR predictors to parse textual information (localize and identify each word) from your documents. Robust 2-stage (detection + recognition) OCR predictors with pretrained parameters. User-friendly, 3 lines of code to load a document and extract text with a predictor. State-of-the-art performances on public document datasets, comparable with GoogleVision/AWS Textract. Easy integration (available templates for browser demo & API deployment). End-to-End OCR is achieved in docTR using a two-stage approach: text detection (localizing words), then text recognition (identify all characters in the word). As such, you can select the architecture used for text detection, and the one for text recognition from the list of available implementations.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 8
    FSRS4Anki

    FSRS4Anki

    A modern Anki custom scheduling based on Free Spaced Repetition

    A modern spaced-repetition scheduler for Anki based on the Free Spaced Repetition Scheduler algorithm.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 9
    Machine Learning cheatsheets Stanford

    Machine Learning cheatsheets Stanford

    VIP cheatsheets for Stanford's CS 229 Machine Learning

    stanford-cs-229-machine-learning is an open-source educational repository that provides illustrated cheat sheets summarizing the key concepts taught in Stanford University’s CS229 machine learning course. The project compiles concise explanations of important topics in machine learning and presents them in an accessible format that helps learners review complex ideas quickly. The repository includes summaries covering areas such as supervised learning, unsupervised learning, deep learning, and optimization techniques. In addition to machine learning algorithms, it also contains refresher materials on mathematical prerequisites including probability theory, statistics, linear algebra, and calculus. These cheat sheets are designed to serve as quick reference guides that students can use while studying or reviewing machine learning material.
    Downloads: 9 This Week
    Last Update:
    See Project
  • The full-stack observability platform that protects your dataLayer, tags and conversion data Icon
    The full-stack observability platform that protects your dataLayer, tags and conversion data

    Stop losing revenue to bad data today. and protect your marketing data with Code-Cube.io.

    Code-Cube.io detects issues instantly, alerts you in real time and helps you resolve them fast. No manual QA. No unreliable data. Just data you can trust and act on.
    Learn More
  • 10
    Metaflow

    Metaflow

    A framework for real-life data science

    Metaflow is a human-friendly Python library that helps scientists and engineers build and manage real-life data science projects. Metaflow was originally developed at Netflix to boost productivity of data scientists who work on a wide variety of projects from classical statistics to state-of-the-art deep learning.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 11
    Stable Baselines3

    Stable Baselines3

    PyTorch version of Stable Baselines

    Stable Baselines3 (SB3) is a set of reliable implementations of reinforcement learning algorithms in PyTorch. It is the next major version of Stable Baselines. You can read a detailed presentation of Stable Baselines3 in the v1.0 blog post or our JMLR paper. These algorithms will make it easier for the research community and industry to replicate, refine, and identify new ideas, and will create good baselines to build projects on top of. We expect these tools will be used as a base around which new ideas can be added, and as a tool for comparing a new approach against existing ones. We also hope that the simplicity of these tools will allow beginners to experiment with a more advanced toolset, without being buried in implementation details.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 12
    Turing.jl

    Turing.jl

    Bayesian inference with probabilistic programming

    Bayesian inference with probabilistic programming.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    UnrealCV

    UnrealCV

    Connecting Computer Vision to Unreal Engine

    UnrealCV is a project to help computer vision researchers build virtual worlds using Unreal Engine (UE). It extends UE with a plugin. UnrealCV can be used in two ways. The first one is using a compiled game binary with UnrealCV embedded. This is as simple as running a game, no knowledge of Unreal Engine is required. The second is installing the UnrealCV plugin into Unreal Engine and using the editor to build a new virtual world.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 14
    Weights and Biases

    Weights and Biases

    Tool for visualizing and tracking your machine learning experiments

    Use W&B to build better models faster. Track and visualize all the pieces of your machine learning pipeline, from datasets to production models. Quickly identify model regressions. Use W&B to visualize results in real time, all in a central dashboard. Focus on the interesting ML. Spend less time manually tracking results in spreadsheets and text files. Capture dataset versions with W&B Artifacts to identify how changing data affects your resulting models. Reproduce any model, with saved code, hyperparameters, launch commands, input data, and resulting model weights. Set wandb.config once at the beginning of your script to save your hyperparameters, input settings (like dataset name or model type), and any other independent variables for your experiments. This is useful for analyzing your experiments and reproducing your work in the future. Setting configs also allows you to visualize the relationships between features of your model architecture or data pipeline and model performance.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 15
    gensim

    gensim

    Topic Modelling for Humans

    Gensim is a Python library for topic modeling, document indexing, and similarity retrieval with large corpora. The target audience is the natural language processing (NLP) and information retrieval (IR) community.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 16
    skfolio

    skfolio

    Python library for portfolio optimization built on top of scikit-learn

    skfolio is a Python library designed for portfolio optimization and financial risk management that integrates closely with the scikit-learn ecosystem. The project provides a unified machine learning-style framework for building, validating, and comparing portfolio allocation strategies using financial data. By following the familiar scikit-learn API design, the library allows quantitative researchers and developers to apply techniques such as model selection, cross-validation, and hyperparameter tuning to portfolio construction workflows. It supports a wide range of allocation methods, from classical mean-variance optimization to modern techniques that rely on clustering, factor models, and risk-based allocations. The framework also includes tools for evaluating portfolio performance under different market conditions, enabling users to test robustness and reduce the risk of overfitting.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 17
    Java Neural Network Framework Neuroph
    Neuroph is lightweight Java Neural Network Framework which can be used to develop common neural network architectures. Small number of basic classes which correspond to basic NN concepts, and GUI editor makes it easy to learn and use.
    Leader badge
    Downloads: 41 This Week
    Last Update:
    See Project
  • 18
    AI-Tutorials/Implementations Notebooks

    AI-Tutorials/Implementations Notebooks

    Codes/Notebooks for AI Projects

    AI-Tutorials/Implementations Notebooks repository is a comprehensive collection of artificial intelligence tutorials and implementation examples intended for developers, students, and researchers who want to learn by building practical AI projects. The repository contains numerous Jupyter notebooks and code samples that demonstrate modern techniques in machine learning, deep learning, data science, and large language model workflows. It includes implementations for a wide range of AI topics such as computer vision, agent systems, federated learning, distributed systems, adversarial attacks, and generative AI. Many of the tutorials focus on building AI agents, multi-agent systems, and workflows that integrate language models with external tools or APIs. The codebase acts as a hands-on learning resource, allowing users to experiment with new frameworks, architectures, and machine learning workflows through guided examples.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    Alibi Detect

    Alibi Detect

    Algorithms for outlier, adversarial and drift detection

    Alibi Detect is an open source Python library focused on outlier, adversarial and drift detection. The package aims to cover both online and offline detectors for tabular data, text, images and time series. Both TensorFlow and PyTorch backends are supported for drift detection.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    Alibi Explain

    Alibi Explain

    Algorithms for explaining machine learning models

    Alibi is a Python library aimed at machine learning model inspection and interpretation. The focus of the library is to provide high-quality implementations of black-box, white-box, local and global explanation methods for classification and regression models.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 21
    Apache Hamilton

    Apache Hamilton

    Helps data scientists define testable self-documenting dataflows

    Apache Hamilton is an open-source Python framework designed to simplify the creation and management of dataflows used in analytics, machine learning pipelines, and data engineering workflows. The framework enables developers to define data transformations as simple Python functions, where each function represents a node in a dataflow graph and its parameters define dependencies on other nodes. Hamilton automatically analyzes these functions and constructs a directed acyclic graph representing the pipeline, allowing the system to execute transformations in the correct order. This approach encourages modular, testable, and maintainable data pipelines because each transformation is isolated and easily unit tested. The framework also automatically tracks lineage and metadata about how data is produced, which improves debugging, reproducibility, and transparency in data workflows.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 22
    Axon

    Axon

    Nx-powered Neural Networks

    Nx-powered Neural Networks for Elixir. Axon consists of the following components. Functional API – A low-level API of numerical definitions (defn) of which all other APIs build on. Model Creation API – A high-level model creation API which manages model initialization and application. Optimization API – An API for creating and using first-order optimization techniques based on the Optax library. Training API – An API for quickly training models, inspired by PyTorch Ignite. Axon provides abstractions that enable easy integration while maintaining a level of separation between each component. You should be able to use any of the APIs without dependencies on others. By decoupling the APIs, Axon gives you full control over each aspect of creating and training a neural network. At the lowest-level, Axon consists of a number of modules with functional implementations of common methods in deep learning.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 23
    Bootstrap Your Own Latent (BYOL)

    Bootstrap Your Own Latent (BYOL)

    Usable Implementation of "Bootstrap Your Own Latent" self-supervised

    Practical implementation of an astoundingly simple method for self-supervised learning that achieves a new state-of-the-art (surpassing SimCLR) without contrastive learning and having to designate negative pairs. This repository offers a module that one can easily wrap any image-based neural network (residual network, discriminator, policy network) to immediately start benefitting from unlabelled image data. There is now new evidence that batch normalization is key to making this technique work well. A new paper has successfully replaced batch norm with group norm + weight standardization, refuting that batch statistics are needed for BYOL to work. Simply plugin your neural network, specifying (1) the image dimensions as well as (2) the name (or index) of the hidden layer, whose output is used as the latent representation used for self-supervised training.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 24
    Causal ML

    Causal ML

    Uplift modeling and causal inference with machine learning algorithms

    Causal ML is a Python package that provides a suite of uplift modeling and causal inference methods using machine learning algorithms based on recent research [1]. It provides a standard interface that allows users to estimate the Conditional Average Treatment Effect (CATE) or Individual Treatment Effect (ITE) from experimental or observational data. Essentially, it estimates the causal impact of intervention T on outcome Y for users with observed features X, without strong assumptions on the model form. An important lever to increase ROI in an advertising campaign is to target the ad to the set of customers who will have a favorable response in a given KPI such as engagement or sales. CATE identifies these customers by estimating the effect of the KPI from ad exposure at the individual level from A/B experiments or historical observational data.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 25
    Chemprop

    Chemprop

    Message Passing Neural Networks for Molecule Property Prediction

    Chemprop is a repository containing message-passing neural networks for molecular property prediction.
    Downloads: 8 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB