Open Source Machine Learning Software - Page 9

Machine Learning Software

View 446 business solutions
  • Workable Hiring Software - Hire The Best People, Fast Icon
    Workable Hiring Software - Hire The Best People, Fast

    Find the best candidates with the best recruitment software

    Workable is the preferred software for today's recruiting industry and HR teams, trusted by over 6,000 companies to streamline their hiring processes. Finding the right person for the job has never been easier—users now possess the ability to manage multiple hiring pipelines at once, from posting a job to sourcing candidates. Workable is also seamlessly integrated between desktop and mobile, allowing admins full control and flexibility all in the ATS without needing additional software.
    Learn More
  • Power through agendas and documents, make more informed decisions and conduct board meetings faster. Icon
    Power through agendas and documents, make more informed decisions and conduct board meetings faster.

    For team managers searching for a solution to manage their meetings

    iBabs not only captures the entire decision-making process – it takes all the paperwork out of meetings. iBabs empowers everyone who has ever organized or attended, a meeting. With a seemingly simple app that offers complete control and a comprehensive overview of all those fiddly details. With about 3000 organizations and over 300,000 users, iBabs gives you peace of mind. So you can quickly organize effective meetings, and good decisions can be made with confidence. iBabs didn’t just happen overnight. We started analyzing and simplifying board meeting processes many years ago. We understand all the work that goes into meetings, and how to streamline everything so it all flows smoothly. On any device, confidentially, securely and automatically. Make good decisions with confidence.
    Learn More
  • 1
    Chemprop

    Chemprop

    Message Passing Neural Networks for Molecule Property Prediction

    Chemprop is a repository containing message-passing neural networks for molecular property prediction.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Cloud Annotations

    Cloud Annotations

    A fast, easy and collaborative open source image annotation tool

    Learn computer vision & AI by building real-world applications. Learn to build and train computer vision models—then show off your skills in an interactive web application. Build impressive applications and learn coveted skills. The examples below were created by the Skills Network Team—right here in CV Studio. Create your own project dataset by uploading images and videos. Coming soon, you'll be able to use a pre-compiled dataset so you can hit the ground running. Creating image annotations for your project is easy inside CV Studio. For classification projects, just select and label your images. For object detection, use the integrated tool to highlight target elements in your images. Train your model using the image annotations from the previous step. Practice using cutting-edge tools like Jupyter Notebook, Watson Machine Learning, Elyra, and more.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder, Codeforces, CodeChef, Aizu, and HackerEarth. Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    The Transformer architecture has improved the performance of deep learning models in domains such as Computer Vision and Natural Language Processing. Together with better performance come larger model sizes. This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment. However, distributed training, especially model parallelism, often requires domain expertise in computer systems and architecture. It remains a challenge for AI researchers to implement complex distributed training solutions for their models. Colossal-AI provides a collection of parallel components for you. We aim to support you to write your distributed deep learning models just like how you write your model on your laptop.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Diagnose and Resolve IT Issues in Real Time Icon
    Diagnose and Resolve IT Issues in Real Time

    Engage your employees and agents more efficiently with ScreenMeet as a seamless extension of your existing IT Service Delivery Platform.

    ScreenMeet’s unique combination of video calling, screen share, and remote desktop functionality lets you quickly diagnose hardware and software issues with no frustration.
    Learn More
  • 5
    Computer Vision in Action

    Computer Vision in Action

    A computer vision closed-loop learning platform

    Computer Vision in Action is a practical, example-rich repository that demonstrates real-world applications of computer vision techniques and algorithms in Python, often using OpenCV, deep learning models, and related tooling. It serves as a hands-on companion for learners and engineers who want to understand not just the theory, but how computer vision is actually implemented for tasks like object detection, image classification, feature tracking, optical flow, and image segmentation. The repository includes structured code examples, scripts, and notebooks that cover pipeline construction, preprocessing, model inference, and visual output rendering, making it easy for newcomers or intermediate practitioners to adapt patterns to their own projects. It also explores how to combine classical computer vision techniques with modern neural network-based models, offering insight into when each approach is most effective.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    DALI

    DALI

    A GPU-accelerated library containing highly optimized building blocks

    The NVIDIA Data Loading Library (DALI) is a library for data loading and pre-processing to accelerate deep learning applications. It provides a collection of highly optimized building blocks for loading and processing image, video and audio data. It can be used as a portable drop-in replacement for built-in data loaders and data iterators in popular deep learning frameworks. Deep learning applications require complex, multi-stage data processing pipelines that include loading, decoding, cropping, resizing, and many other augmentations. These data processing pipelines, which are currently executed on the CPU, have become a bottleneck, limiting the performance and scalability of training and inference. DALI addresses the problem of the CPU bottleneck by offloading data preprocessing to the GPU. Additionally, DALI relies on its own execution engine, built to maximize the throughput of the input pipeline.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    DATA SCIENCE ROADMAP

    DATA SCIENCE ROADMAP

    Data Science Roadmap from A to Z

    DATA SCIENCE ROADMAP is an educational repository designed to guide learners through the process of becoming proficient in data science and machine learning. The project presents a structured roadmap that outlines the knowledge and skills required for different stages of a data science career. Topics typically include programming with Python, statistics, mathematics, machine learning algorithms, data visualization, and big data technologies. The roadmap also includes links to courses, tutorials, and external resources that help learners study each topic in more depth. By organizing these subjects into a logical sequence, the repository helps beginners understand how different technical skills connect within the broader data science workflow. The roadmap format makes it easy for learners to track their progress as they move from foundational concepts to more advanced techniques.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    DIGITS

    DIGITS

    Deep Learning GPU training system

    The NVIDIA Deep Learning GPU Training System (DIGITS) puts the power of deep learning into the hands of engineers and data scientists. DIGITS can be used to rapidly train the highly accurate deep neural network (DNNs) for image classification, segmentation and object detection tasks. DIGITS simplifies common deep learning tasks such as managing data, designing and training neural networks on multi-GPU systems, monitoring performance in real-time with advanced visualizations, and selecting the best performing model from the results browser for deployment. DIGITS is completely interactive so that data scientists can focus on designing and training networks rather than programming and debugging. DIGITS is available as a free download to the members of the NVIDIA Developer Program. DIGITS is available on NVIDIA GPU Cloud (NGC) as an optimized container for on-demand usage. Sign-up for an NGC account and get started with DIGITS in minutes.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Deep-Learning-for-Recommendation-Systems

    Deep-Learning-for-Recommendation-Systems

    This repository contains Deep Learning based articles

    Deep-Learning-for-Recommendation-Systems is a curated repository that aggregates research papers, articles, and code related to deep learning methods for recommender systems. The project organizes influential academic work covering topics such as collaborative filtering, neural recommendation models, and deep feature learning. It includes references to papers describing architectures like collaborative deep learning, neural autoregressive models, and convolutional approaches to recommendation. The repository also provides links to implementations and external code repositories that demonstrate how these algorithms can be applied in real systems. By compiling research literature and practical resources in one location, the project helps researchers and engineers explore the evolving landscape of recommendation technologies. It highlights both theoretical innovations and applied engineering work used in modern recommendation engines.
    Downloads: 1 This Week
    Last Update:
    See Project
  • DataHub is the leading open-source data catalog helping teams discover, understand, and govern their data assets. Icon
    DataHub is the leading open-source data catalog helping teams discover, understand, and govern their data assets.

    Modern Data Catalog and Metadata Platform

    Built on an open source foundation with a thriving community of 13,000+ members, DataHub gives you unmatched flexibility to customize and extend without vendor lock-in. DataHub Cloud is a modern metadata platform with REST and GraphQL APIs that optimize performance for complex queries, essential for AI-ready data management and ML lifecycle support.
    Learn More
  • 10
    DeepCTR

    DeepCTR

    Package of deep-learning based CTR models

    DeepCTR is a Easy-to-use,Modular and Extendible package of deep-learning based CTR models along with lots of core components layers which can be used to easily build custom models. You can use any complex model with model.fit(), and model.predict(). Provide tf.keras.Model like interface for quick experiment. Provide tensorflow estimator interface for large scale data and distributed training. It is compatible with both tf 1.x and tf 2.x. With the great success of deep learning,DNN-based techniques have been widely used in CTR prediction task. The data in CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Since DNN are good at handling dense numerical features,we usually map the sparse categorical features to dense numerical through embedding technique.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    DeepDanbooru

    DeepDanbooru

    AI based multi-label girl image classification system

    DeepDanbooru is a deep learning system designed to automatically tag anime-style images using neural networks trained on datasets derived from the Danbooru imageboard. The project focuses on multi-label image classification, where a model predicts multiple descriptive tags that represent visual elements in an image. These tags may include characters, styles, clothing, emotions, or other attributes associated with anime artwork. The system uses convolutional neural networks trained on large datasets of tagged images to learn relationships between visual features and textual labels. Because the Danbooru dataset contains millions of images with extensive annotations, it provides a valuable training resource for machine learning models specializing in illustration analysis. Such datasets have been widely used for tasks including automatic image tagging, anime face detection, and generative modeling research.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Deepnote

    Deepnote

    Deepnote is a drop-in replacement for Jupyter

    Deepnote is an open-source collaborative data science notebook platform designed as a modern alternative to traditional Jupyter notebooks. The project provides an AI-first computational environment where users can write, analyze, and share code, data, and visualizations in a single integrated workspace. Built on top of the Jupyter kernel ecosystem, it maintains compatibility with existing notebook workflows while introducing additional features focused on collaboration and automation. The system supports programming languages such as Python, R, and SQL and allows users to execute and analyze data directly within interactive notebooks. Deepnote emphasizes team-based data science by enabling real-time collaboration similar to shared document editors, allowing multiple users to work simultaneously on the same notebook environment.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Deepvoice3_pytorch

    Deepvoice3_pytorch

    PyTorch implementation of convolutional neural networks

    An open source implementation of Deep Voice 3: Scaling Text-to-Speech with Convolutional Sequence Learning.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    DocTR

    DocTR

    Library for OCR-related tasks powered by Deep Learning

    DocTR provides an easy and powerful way to extract valuable information from your documents. Seemlessly process documents for Natural Language Understanding tasks: we provide OCR predictors to parse textual information (localize and identify each word) from your documents. Robust 2-stage (detection + recognition) OCR predictors with pretrained parameters. User-friendly, 3 lines of code to load a document and extract text with a predictor. State-of-the-art performances on public document datasets, comparable with GoogleVision/AWS Textract. Easy integration (available templates for browser demo & API deployment). End-to-End OCR is achieved in docTR using a two-stage approach: text detection (localizing words), then text recognition (identify all characters in the word). As such, you can select the architecture used for text detection, and the one for text recognition from the list of available implementations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Evidently

    Evidently

    Evaluate and monitor ML models from validation to production

    Evidently is an open-source Python library for data scientists and ML engineers. It helps evaluate, test, and monitor ML models from validation to production. It works with tabular, text data and embeddings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Flux.jl

    Flux.jl

    Relax! Flux is the ML library that doesn't make you tensor

    Flux is an elegant approach to machine learning. It's a 100% pure Julia stack and provides lightweight abstractions on top of Julia's native GPU and AD support. Flux makes the easy things easy while remaining fully hackable. Flux provides a single, intuitive way to define models, just like mathematical notation. Julia transparently compiles your code, optimizing and fusing kernels for the GPU, for the best performance. Existing Julia libraries are differentiable and can be incorporated directly into Flux models. Cutting-edge models such as Neural ODEs are first class, and Zygote enables overhead-free gradients. GPU kernels can be written directly in Julia via CUDA.jl. Flux is uniquely hackable and any part can be tweaked, from GPU code to custom gradients and layers.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    GROBID

    GROBID

    A machine learning software for extracting information

    GROBID is a machine learning library for extracting, parsing, and re-structuring raw documents such as PDF into structured XML/TEI encoded documents with a particular focus on technical and scientific publications. First developments started in 2008 as a hobby. In 2011 the tool has been made available in open source. Work on GROBID has been steady as a side project since the beginning and is expected to continue as such. Header extraction and parsing from article in PDF format. The extraction here covers the usual bibliographical information (e.g. title, abstract, authors, affiliations, keywords, etc.). References extraction and parsing from articles in PDF format, around .87 F1-score against on an independent PubMed Central set of 1943 PDF containing 90,125 references, and around .89 on a similar bioRxiv set of 2000 PDF (using the Deep Learning citation model). All the usual publication metadata are covered (including DOI, PMID, etc.).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Hubot

    Hubot

    A customizable life embetterment robot

    Hubot is a framework to build a custom chat bot, first built by GitHub, Inc. to automate their company chat room. Hubot gives you a very nice base for building your very own robot friend. Hubot comes with a small group of core scripts, including things like posting images, translating languages, and integrating with Google Maps. It's extendable with many other scripts, which make Hubot all the more personalized to fit your organization's needs and culture. Hubot can work on many different chat services, including third-party adapters Gitter, IRC, Slack and more. Build your own personalized chat bot with Hubot!
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Hugging Face Transformer

    Hugging Face Transformer

    CPU/GPU inference server for Hugging Face transformer models

    Optimize and deploy in production Hugging Face Transformer models in a single command line. At Lefebvre Dalloz we run in-production semantic search engines in the legal domain, in the non-marketing language it's a re-ranker, and we based ours on Transformer. In that setup, latency is key to providing a good user experience, and relevancy inference is done online for hundreds of snippets per user query. Most tutorials on Transformer deployment in production are built over Pytorch and FastAPI. Both are great tools but not very performant in inference. Then, if you spend some time, you can build something over ONNX Runtime and Triton inference server. You will usually get from 2X to 4X faster inference compared to vanilla Pytorch. It's cool! However, if you want the best in class performances on GPU, there is only a single possible combination: Nvidia TensorRT and Triton. You will usually get 5X faster inference compared to vanilla Pytorch.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Hummingbird

    Hummingbird

    Hummingbird compiles trained ML models into tensor computation

    Hummingbird is a library for compiling trained traditional ML models into tensor computations. Hummingbird allows users to seamlessly leverage neural network frameworks (such as PyTorch) to accelerate traditional ML models. Thanks to Hummingbird, users can benefit from (1) all the current and future optimizations implemented in neural network frameworks; (2) native hardware acceleration; (3) having a unique platform to support both traditional and neural network models; and having all of this (4) without having to re-engineer their models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    IREE

    IREE

    A retargetable MLIR-based machine learning compiler runtime toolkit

    IREE (Intermediate Representation Execution Environment, pronounced as "eerie") is an MLIR-based end-to-end compiler and runtime that lowers Machine Learning (ML) models to a unified IR that scales up to meet the needs of the data center and down to satisfy the constraints and special considerations of mobile and edge deployments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    ISLR-python

    ISLR-python

    An Introduction to Statistical Learning

    ISLR-python is an educational repository that provides Python implementations and notebooks corresponding to examples and exercises from the book An Introduction to Statistical Learning. The project recreates tables, figures, and laboratory exercises originally presented in the book so that readers can explore the concepts using Python rather than the original R environment. The repository includes Jupyter notebooks demonstrating statistical learning methods such as linear regression, classification algorithms, resampling methods, and model evaluation techniques. These notebooks combine theoretical explanations with practical coding exercises that allow users to reproduce the analyses described in the book. The datasets used in the book are also included so that users can run experiments directly within the provided notebooks. By translating the statistical learning material into Python code, the repository makes the book’s concepts accessible to a wider community of Python users.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Intel Extension for PyTorch

    Intel Extension for PyTorch

    A Python package for extending the official PyTorch

    Intel® Extension for PyTorch* extends PyTorch* with up-to-date features optimizations for an extra performance boost on Intel hardware. Optimizations take advantage of Intel® Advanced Vector Extensions 512 (Intel® AVX-512) Vector Neural Network Instructions (VNNI) and Intel® Advanced Matrix Extensions (Intel® AMX) on Intel CPUs as well as Intel Xe Matrix Extensions (XMX) AI engines on Intel discrete GPUs. Moreover, Intel® Extension for PyTorch* provides easy GPU acceleration for Intel discrete GPUs through the PyTorch* xpu device.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Karpathy

    Karpathy

    An agentic Machine Learning Engineer

    karpathy is an experimental agentic machine learning engineer framework designed to automate many aspects of the ML development workflow. The project sets up a sandboxed environment where an AI agent can access datasets, run experiments, and generate machine learning artifacts through a web interface. Its startup script automatically prepares the environment by creating a sandbox directory, installing key ML libraries, and launching the agent interface. The system is tightly integrated with the Claude Scientific Skills ecosystem, enabling the agent to leverage specialized scientific and machine learning tools. It is intended primarily for research and experimentation with autonomous ML workflows rather than as a polished production platform. Overall, karpathy represents an early step toward fully automated machine learning engineering driven by agentic AI systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Koila

    Koila

    Prevent PyTorch's `CUDA error: out of memory` in just 1 line of code

    Koila is a lightweight Python library designed to help developers avoid memory errors when training deep learning models with PyTorch. The library introduces a lazy evaluation mechanism that delays computation until it is actually required, allowing the framework to better estimate the memory requirements of a model before execution. By building a computational graph first and executing operations only when necessary, koila reduces the risk of running out of GPU memory during the forward pass of neural network training. This approach enables developers to experiment with larger batch sizes and more complex architectures while maintaining stable training behavior. The system acts as a thin wrapper around PyTorch tensors and operations, meaning that it integrates easily into existing PyTorch code without requiring major changes to model implementations. It is particularly useful in environments where GPU resources are limited or where models frequently encounter CUDA memory errors.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB