Open Source C++ Machine Learning Software - Page 7

C++ Machine Learning Software

View 446 business solutions

Browse free open source C++ Machine Learning Software and projects below. Use the toggles on the left to filter open source C++ Machine Learning Software by OS, license, language, programming language, and project status.

  • Time tracking software for the global workforce Icon
    Time tracking software for the global workforce

    Teams of all sizes and in various industries that want the best time tracking and employee monitoring solution.

    It's easy with Hubstaff, a time-tracking and workforce management platform that automates almost every aspect of running or growing a business. Teams can track time to projects and to-dos using Hubstaff's desktop, web, or mobile applications. You'll be able to see how much time your team spends on different tasks, plus productivity metrics like activity rates and app usage through Hubstaff's online dashboard. Most of the available features are customizable on a per-user basis, so you can create the team management tool you need.
    Learn More
  • Melis Platform is an enterprise-grade Low Code Platform simplifying app creation, management, and delivery. Icon
    Melis Platform is an enterprise-grade Low Code Platform simplifying app creation, management, and delivery.

    Ideal for websites, apps, e-commerce, CRMs, and more

    Melis is a new generation of Content Management System and eCommerce platform to achieve and manage websites from a single web interface easy to use while offering the best of open source technology.
    Learn More
  • 1
    Reinforcement learning and simulation-based search in computer Go
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    RobGP is a genetic programming system written from the ground up in C++. It's primary goals are efficiency, ease of use, and extensibility. It's distinguishing feature is that it has a modified version of Koza's architecture altering operations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    SINGA

    SINGA

    A distributed deep learning platform

    Apache SINGA is an Apache Top Level Project, focusing on distributed training of deep learning and machine learning models. Various example deep learning models are provided in SINGA repo on Github and on Google Colab. SINGA supports data parallel training across multiple GPUs (on a single node or across different nodes). SINGA supports various popular optimizers including stochastic gradient descent with momentum, Adam, RMSProp, and AdaGrad, etc. SINGA records the computation graph and applies the backward propagation automatically after forward propagation. The optimization of memory are implemented in the Device class. SINGA supports loading ONNX format models and saving models defined using SINGA APIs into ONNX format, which enables AI developers to use models across different libraries and tools. SINGA supports the time profiling of each of the operators buffered in the graph. Half precision is supported to bring benefits.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    SLING

    SLING

    A natural language frame semantics parser

    The aim of the SLING project is to learn to read and understand Wikipedia articles in many languages for the purpose of knowledge base completion, e.g. adding facts mentioned in Wikipedia (and other sources) to the Wikidata knowledge base. We use frame semantics as a common representation for both knowledge representation and document annotation. The SLING parser can be trained to produce frame semantic representations of text directly without any explicit intervening linguistic representation. The SLING project is still work in progress. We do not yet have a full system that can extract facts from arbitrary text, but we have built a number of the subsystems needed for such a system. The SLING frame store is our basic framework for building and manipulating frame semantic graph structures. The Wiki flow pipeline can take a raw dump of Wikidata and convert this into one big frame graph.
    Downloads: 0 This Week
    Last Update:
    See Project
  • All Things Performance and Partner Marketing, All in One Place Icon
    All Things Performance and Partner Marketing, All in One Place

    Track calls, leads, and clicks without the manual work

    Automatically tie revenue back to campaigns, channels, publishers, and networks through marketing attribution. Spend less time juggling reports, and more time optimizing for growth by using a single operating solution for partner and performance marketing.
    Learn More
  • 5

    Savant

    Python Computer Vision & Video Analytics Framework With Batteries Incl

    Savant is an open-source, high-level framework for building real-time, streaming, highly efficient multimedia AI applications on the Nvidia stack. It helps to develop dynamic, fault-tolerant inference pipelines that utilize the best Nvidia approaches for data center and edge accelerators. Savant is built on DeepStream and provides a high-level abstraction layer for building inference pipelines. It is designed to be easy to use, flexible, and scalable. It is a great choice for building smart CV and video analytics applications for cities, retail, manufacturing, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Simd Library

    Simd Library

    C++ image processing and machine learning library with using of SIMD

    The Simd Library is a free open-source image processing and machine learning library, designed for C and C++ programmers. It provides many useful high-performance algorithms for image processing such as pixel format conversion, image scaling and filtration, extraction of statistical information from images, motion detection, object detection and classification, neural networks. The algorithms are optimized with using of different SIMD CPU extensions. In particular, the library supports the following CPU extensions: SSE, AVX, AVX-512, and AMX for x86/x64, and NEON for ARM. The Simd Library has C API and also contains useful C++ classes and functions to facilitate access to C API. The library supports dynamic and static linking, 32-bit and 64-bit Windows and Linux, MSVS, G++ and Clang compilers, MSVS projects, and CMake build systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Computer Vision Application.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    An implementation of a new proposed model of smoothly spiking neural networks + a fully analytical gradient descent algorithm.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9

    StabLe

    An algorithm for learning stable graphical models from data

    Stable Graphical Model Learning (StabLe) is an algorithm for learning the structure and parameters of stable graphical (SG) models from data. Stable random variables are motivated by the central limit theorem for densities with (potentially) unbounded variance and can be thought of as natural generalizations of the Gaussian distribution to skewed and heavy-tailed phenomenon. SG models are multi-variate stable distributions that represent Bayesian networks whose edges encode linear dependencies amongst random variables. A preprint version of the manuscript describing stable graphical models is available at http://arxiv.org/abs/1404.4351.
    Downloads: 0 This Week
    Last Update:
    See Project
  • SIEM | API Security | Log Management Software Icon
    SIEM | API Security | Log Management Software

    AI-Powered Security and IT Operations Without Compromise.

    Built on the Graylog Platform, Graylog Security is the industry’s best-of-breed threat detection, investigation, and response (TDIR) solution. It simplifies analysts’ day-to-day cybersecurity activities with an unmatched workflow and user experience while simultaneously providing short- and long-term budget flexibility in the form of low total cost of ownership (TCO) that CISOs covet. With Graylog Security, security analysts can:
    Learn More
  • 10
    Star Ships Learning Framework
    The Star Ships Learning Framework (SSLF) provides basic routines and methods in order to interact with the space shooter Star Ships V2.1. It provides the necessary interfaces for controlling the program's ships and its logic from outside the game.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Stochastico is an implementation of stochastic discrimination for pattern recognition, predictive modeling and data mining applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    TNN

    TNN

    Uniform deep learning inference framework for mobile

    TNN, a high-performance, lightweight neural network inference framework open sourced by Tencent Youtu Lab. It also has many outstanding advantages such as cross-platform, high performance, model compression, and code tailoring. The TNN framework further strengthens the support and performance optimization of mobile devices on the basis of the original Rapidnet and ncnn frameworks. At the same time, it refers to the high performance and good scalability characteristics of the industry's mainstream open source frameworks, and expands the support for X86 and NV GPUs. On the mobile phone, TNN has been used by many applications such as mobile QQ, weishi, and Pitu. As a basic acceleration framework for Tencent Cloud AI, TNN has provided acceleration support for the implementation of many businesses. Everyone is welcome to participate in the collaborative construction to promote the further improvement of the TNN inference framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Tensor Comprehensions

    Tensor Comprehensions

    A domain specific language to express machine learning workloads

    Tensor Comprehensions (TC) is a fully functional C++ library that automatically synthesizes high-performance machine learning kernels using Halide, ISL, and NVRTC or LLVM. TC additionally provides basic integration with Caffe2 and PyTorch. We provide more details in our paper on arXiv. This library is designed to be highly portable, machine-learning-framework agnostic and only requires a simple tensor library with memory allocation, offloading, and synchronization capabilities.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    TensorFlow Serving

    TensorFlow Serving

    Serving system for machine learning models

    TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. It deals with the inference aspect of machine learning, taking models after training and managing their lifetimes, providing clients with versioned access via a high-performance, reference-counted lookup table. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data. The easiest and most straight-forward way of using TensorFlow Serving is with Docker images. We highly recommend this route unless you have specific needs that are not addressed by running in a container. In order to serve a Tensorflow model, simply export a SavedModel from your Tensorflow program. SavedModel is a language-neutral, recoverable, hermetic serialization format that enables higher-level systems and tools to produce, consume, and transform TensorFlow models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    TensorRT Backend For ONNX

    TensorRT Backend For ONNX

    ONNX-TensorRT: TensorRT backend for ONNX

    Parses ONNX models for execution with TensorRT. Development on the main branch is for the latest version of TensorRT 8.4.1.5 with full dimensions and dynamic shape support. For previous versions of TensorRT, refer to their respective branches. Building INetwork objects in full dimensions mode with dynamic shape support requires calling the C++ and Python API. Current supported ONNX operators are found in the operator support matrix. For building within docker, we recommend using and setting up the docker containers as instructed in the main (TensorRT repository). Note that this project has a dependency on CUDA. By default the build will look in /usr/local/cuda for the CUDA toolkit installation. If your CUDA path is different, overwrite the default path. ONNX models can be converted to serialized TensorRT engines using the onnx2trt executable.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    TensorRT Pro

    TensorRT Pro

    C++ library based on tensorrt integration

    High-level interface for C++/Python. Simplify the implementation of the custom plugin. And serialization and deserialization have been encapsulated for easier usage. Simplify the compilation of fp32, fp16 and int8 for facilitating the deployment with C++/Python in server or embedded device. Models ready for use also with examples are RetinaFace, Scrfd, YoloV5, YoloX, Arcface, AlphaPose, CenterNet and DeepSORT(C++).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    The Edge Machine Learning library

    The Edge Machine Learning library

    Machine learning algorithms for edge devices

    Machine learning models for edge devices need to have a small footprint in terms of storage, prediction latency, and energy. One instance of where such models are desirable is resource-scarce devices and sensors in the Internet of Things (IoT) setting. Making real-time predictions locally on IoT devices without connecting to the cloud requires models that fit in a few kilobytes.These algorithms can train models for classical supervised learning problems with memory requirements that are orders of magnitude lower than other modern ML algorithms. The trained models can be loaded onto edge devices such as IoT devices/sensors, and used to make fast and accurate predictions completely offline. A tool that adapts models trained by above algorithms to be inferred by fixed point arithmetic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    The STAIR Vision Library (SVL), originally developed to support the STanford AI Robot, provides software infrastructure for computer vision, machine learning, and probabilistic graphical models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    The goal of the project it to learn how to use Neural Networks, and teach one how to play tic tac toe, and get to the never losing ability that most people can play at.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Tiny CUDA Neural Networks

    Tiny CUDA Neural Networks

    Lightning fast C++/CUDA neural network framework

    This is a small, self-contained framework for training and querying neural networks. Most notably, it contains a lightning-fast "fully fused" multi-layer perceptron (technical paper), a versatile multiresolution hash encoding (technical paper), as well as support for various other input encodings, losses, and optimizers. We provide a sample application where an image function (x,y) -> (R,G,B) is learned. The fully fused MLP component of this framework requires a very large amount of shared memory in its default configuration. It will likely only work on an RTX 3090, an RTX 2080 Ti, or high-end enterprise GPUs. Lower-end cards must reduce the n_neurons parameter or use the CutlassMLP (better compatibility but slower) instead. tiny-cuda-nn comes with a PyTorch extension that allows using the fast MLPs and input encodings from within a Python context. These bindings can be significantly faster than full Python implementations; in particular for the multiresolution hash encoding.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    TorchCraft

    TorchCraft

    Connecting Torch to StarCraft

    We present TorchCraft, a library that enables deep learning research on Real-Time Strategy (RTS) games such as StarCraft: Brood War, by making it easier to control these games from a machine learning framework, here Torch. This white paper argues for using RTS games as a benchmark for AI research, and describes the design and components of TorchCraft. TorchCraft is a BWAPI module that sends StarCraft data out over a ZMQ connection. This lets you parse StarCraft data and interact with BWAPI from anywhere. The TorchCraft client should be installed from C++, Python, or Lua. We provide off-the-shelf solutions for Python and Lua.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    TrackMania Challenge Generator

    Challenge generator for racing game 'TrackMania Nations Forever'

    A small tool which generates infinite number of different challenges for racing game 'TrackMania Nations Forever'. Currently only "Stadium" environment is supported. Latest release: 1.0.1.0, both English and Chinese version available.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    TurboPilot

    TurboPilot

    Open source large-language-model based code completion engine

    TurboPilot is a self-hosted copilot clone that uses the library behind llama.cpp to run the 6 Billion Parameter Salesforce Codegen model in 4GiB of RAM. It is heavily based and inspired by on the fauxpilot project. This is a proof of concept right now rather than a stable tool. Autocompletion is quite slow in this version of the project. Feel free to play with it, but your mileage may vary.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Turi Create

    Turi Create

    Simplifies the development of custom machine learning models

    Turi Create simplifies the development of custom machine learning models. You don't have to be a machine learning expert to add recommendations, object detection, image classification, image similarity or activity classification to your app. If you want your app to recognize specific objects in images, you can build your own model with just a few lines of code. Turi Create supports macOS 10.12+, Linux (with glibc 2.10+), Windows 10 (via WSL). Turi Create requires Python 2.7, 3.5, 3.6, 3.7, 3.8. Also, x86_64 architecture, and at least 4 GB of RAM. We recommend using virtualenv to use, install, or build Turi Create. The package User Guide and API Docs contain more details on how to use Turi Create. If you want to build Turi Create from source, see BUILD.md. Turi Create does not require a GPU, but certain models can be accelerated 9-13x by utilizing a GPU.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Python framework for video processing and content analysis using CUDA for acceleration.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB