Showing 3 open source projects for "machine learning"

View related business solutions
  • Ecwid is a hosted cloud commerce platform used by over 1.5 million merchants and offers the easiest way to add an online store to any website, social site or multiple sites simultaneously. Icon
    Ecwid is a hosted cloud commerce platform used by over 1.5 million merchants and offers the easiest way to add an online store to any website, social site or multiple sites simultaneously.

    Your free online store is just a few clicks away.

    Set up your Ecwid store once to easily sync and sell across a website, social media, marketplaces like Amazon, and live in-person. Get started with one, or try them all.
    Start Selling
  • OpenMetal is an automated bare metal and on-demand private cloud provider. Icon
    OpenMetal is an automated bare metal and on-demand private cloud provider.

    Large Scale. Cloud Native. Fixed Costs.

    OpenMetal is an automated bare metal and on-demand private cloud provider. Our mission is to empower your team with cost effective private infrastructure that outperforms traditional public cloud.
    Learn More
  • 1
    ONNX Runtime

    ONNX Runtime

    ONNX Runtime: cross-platform, high performance ML inferencing

    ONNX Runtime is a cross-platform inference and training machine-learning accelerator. ONNX Runtime inference can enable faster customer experiences and lower costs, supporting models from deep learning frameworks such as PyTorch and TensorFlow/Keras as well as classical machine learning libraries such as scikit-learn, LightGBM, XGBoost, etc. ONNX Runtime is compatible with different hardware, drivers, and operating systems, and provides optimal performance by leveraging hardware accelerators where applicable alongside graph optimizations and transforms. ...
    Downloads: 53 This Week
    Last Update:
    See Project
  • 2
    AWS Neuron

    AWS Neuron

    Powering Amazon custom machine learning chips

    AWS Neuron is a software development kit (SDK) for running machine learning inference using AWS Inferentia chips. It consists of a compiler, run-time, and profiling tools that enable developers to run high-performance and low latency inference using AWS Inferentia-based Amazon EC2 Inf1 instances. Using Neuron developers can easily train their machine learning models on any popular framework such as TensorFlow, PyTorch, and MXNet, and run it optimally on Amazon EC2 Inf1 instances. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for deep learning inference applications. TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers,...
    Downloads: 10 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB