Showing 202 open source projects for "runtime"

View related business solutions
  • Find out just how much your login box can do for your customer | Auth0 Icon
    Find out just how much your login box can do for your customer | Auth0

    With over 53 social login options, you can fast-track the signup and login experience for users.

    From improving customer experience through seamless sign-on to making MFA as easy as a click of a button – your login box must find the right balance between user convenience, privacy and security.
    Sign up
  • Job Evaluation and Talent Management Software Icon
    Job Evaluation and Talent Management Software

    For human resources departments in search of a tool to manage time, expenses, leave, documents, recruitment, and onboarding

    Encompassing Visions (ENCV), industry-leading job evaluation and pay equity software, is the best choice for organizations requiring transparent, comprehensive, and objective Job Evaluation software designed to help them ensure equal pay for work of equal value.
    Learn More
  • 1
    ONNX Runtime

    ONNX Runtime

    ONNX Runtime: cross-platform, high performance ML inferencing

    ONNX Runtime is a cross-platform inference and training machine-learning accelerator. ONNX Runtime inference can enable faster customer experiences and lower costs, supporting models from deep learning frameworks such as PyTorch and TensorFlow/Keras as well as classical machine learning libraries such as scikit-learn, LightGBM, XGBoost, etc.
    Downloads: 47 This Week
    Last Update:
    See Project
  • 2
    kokoro-onnx

    kokoro-onnx

    TTS with kokoro and onnx runtime

    kokoro-onnx is a text-to-speech toolkit that wraps the Kokoro neural TTS model in an easy-to-use ONNX Runtime interface, so you can generate speech from Python with minimal setup. It focuses on running efficiently on commodity hardware, including macOS with Apple Silicon, while still delivering near real-time performance for many use cases. The project ships prebuilt model files and a simple example script, so you can go from installation to producing an audio.wav file in just a few steps. ...
    Downloads: 167 This Week
    Last Update:
    See Project
  • 3
    Osaurus

    Osaurus

    AI edge infrastructure for macOS. Run local or cloud models

    Osaurus is an open-source AI edge infrastructure platform designed specifically for macOS environments to run and manage AI models locally. The project provides a native runtime that allows applications to access large language models and AI tools directly on the user’s machine without relying entirely on cloud services. Osaurus supports running both local and remote models, enabling developers to build AI-powered applications that can operate offline or leverage external APIs when needed. The platform acts as an always-on runtime that coordinates AI tasks, tools, and workflows while enabling applications to communicate with models through standardized interfaces. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 4
    Monoio

    Monoio

    Rust async runtime based on io-uring

    Monoio is a Rust asynchronous runtime designed for high-performance I/O-bound servers and applications, built around native OS async I/O primitives (e.g. io_uring on Linux, epoll / kqueue on other Unix-like systems), rather than layering atop an existing runtime. Its design philosophy centers on a “thread-per-core” model where each core runs its own event loop, minimizing cross-thread synchronization needs, avoiding the overhead and complexity of task scheduling, and letting developers write efficient, low-overhead asynchronous networking or I/O code. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • QA Wolf | We Write, Run and Maintain Tests Icon
    QA Wolf | We Write, Run and Maintain Tests

    For developer teams searching for a testing software

    QA Wolf is an AI-native service that delivers 80% automated E2E test coverage for web & mobile apps in weeks not years.
    Learn More
  • 5
    LiteRT

    LiteRT

    LiteRT is the new name for TensorFlow Lite (TFLite)

    LiteRT is an experimental, real-time inference runtime built by Google AI Edge to run lightweight ML models on edge devices with ultra-low latency. It focuses on delivering predictable and consistent performance for models used in time-critical applications like robotics, AR/VR, and IoT. LiteRT is designed to be hardware-agnostic, with minimal dependencies and tight control over execution scheduling.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 6
    Claw Code

    Claw Code

    AI agent harness for AI coding agents

    ...It emphasizes harness engineering—how agents are structured, how they interact with tools, and how they maintain context during execution. The system is being actively expanded, with a Rust-based runtime in development to improve performance and memory safety. Overall, Claw Code serves as a research-driven platform for advancing agent-based software development systems.
    Downloads: 54 This Week
    Last Update:
    See Project
  • 7
    SmythOS

    SmythOS

    Cloud-native runtime for agentic AI

    SmythOS SRE (Smyth Runtime Environment) is an open-source runtime and development platform designed for building and operating production-grade AI agents. It provides a foundational infrastructure layer that functions similarly to an operating system for agentic AI systems, managing resources such as language models, storage, vector databases, and caching through a unified interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    ByteHook

    ByteHook

    ByteHook is an Android PLT hook library

    ...As such, Bhook would serve developers needing fine-grained control over runtime execution, e.g. to intercept calls, log behaviors, protect processes, or adapt system behavior dynamically.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Docker Agent

    Docker Agent

    AI Agent Builder and Runtime by Docker Engineering

    Docker Agent is an open-source multi-agent runtime developed by Docker that enables developers to define, run, and orchestrate AI agents using simple declarative configuration files instead of traditional code-heavy approaches. It introduces a YAML-based configuration model where users describe agent behavior, tools, models, and interaction logic in a single file, significantly reducing complexity in building AI systems.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Assembled is the only unified platform for staffing and managing your human and AI support team. Icon
    Assembled is the only unified platform for staffing and managing your human and AI support team.

    AI for world-class support operations

    Assembled is the only platform that unifies AI agents and intelligent workforce management to power fast and flexible support operations. Built for scale, we help teams automate over 50% of customer interactions, forecast with 90%+ accuracy, and optimize staffing across in-house and BPO teams. Orchestrate every chat, email, or call, balancing workloads between human and AI agents in real time — without sacrificing quality or control. Trusted by Stripe, Canva, and Robinhood, Assembled transforms support from a cost center into a strategic advantage. Our Workforce and Vendor Management tools connect forecasting, scheduling, and performance for smarter staffing decisions. AI Agents automate conversations across channels with your workflows and brand voice. AI Copilot empowers agents with real-time guidance, suggested replies, and one-click actions for faster, higher-quality resolutions.
    Learn More
  • 10
    IREE

    IREE

    A retargetable MLIR-based machine learning compiler runtime toolkit

    IREE (Intermediate Representation Execution Environment, pronounced as "eerie") is an MLIR-based end-to-end compiler and runtime that lowers Machine Learning (ML) models to a unified IR that scales up to meet the needs of the data center and down to satisfy the constraints and special considerations of mobile and edge deployments.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    AsmJit

    AsmJit

    Low-latency machine code generation

    ...It is commonly used in applications such as emulators, compilers, and high-performance computing systems where runtime optimization is essential. asmjit emphasizes low latency and efficiency, ensuring that generated code executes quickly without significant overhead. Its modular design allows developers to integrate it into various systems with minimal friction. Overall, asmjit bridges the gap between high-level programming and low-level execution by enabling efficient runtime code generation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    AnyClaw

    AnyClaw

    Android app running two AI coding agents with a built-in Linux runtime

    ...Both agents run inside a self-contained Linux userland that is embedded within the APK, allowing command execution, coding tasks, and agent interactions without requiring root access or external tools. openclaw-android-assistant provides a control dashboard interface where users can manage agents, sessions, skills, and conversations from a unified interface. Through the embedded runtime, the agents can read codebases, write code, and execute shell commands within the packaged Linux environment. It also supports parallel conversations with separate working contexts, enabling multi-threaded development workflows on a mobile device.
    Downloads: 24 This Week
    Last Update:
    See Project
  • 13
    Ralph AI Agent

    Ralph AI Agent

    AI agent loop that runs repeatedly until all PRD items are complete

    Ralph is a Rust-based AI agent runtime that focuses on safe, modular, and programmable autonomous behavior. It provides a reactive loop where agents can repeatedly assess the current context, reason about the next best action using large language models, and execute actions across integrated tools and services. The runtime emphasizes safety boundaries by sandboxing operations, enforcing time and token limits, and isolating execution layers to prevent unpredictable side effects. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 358 This Week
    Last Update:
    See Project
  • 15
    OpenFang

    OpenFang

    Open-source Agent Operating System

    OpenFang is an open-source agent operating system designed to orchestrate autonomous AI agents and workflows in a structured, production-oriented environment. Written primarily in Rust, the project focuses on building a high-performance runtime where multiple specialized agents can collaborate to complete complex computational or development tasks. It aims to move beyond simple chat-based agents by providing infrastructure for persistent agent memory, task coordination, and scalable execution. The system is positioned as a foundation for building advanced AI tooling, particularly in environments that require tight integration with GPU workflows and modern AI pipelines. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 16
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ...One of its key strengths is its flexibility, as it supports multiple backends and allows developers to configure execution providers depending on available hardware. ort also includes advanced capabilities such as model compilation and optimization, reducing startup time and improving runtime performance in production systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    OpenShell

    OpenShell

    OpenShell is the safe, private runtime for autonomous AI agents.

    OpenShell is an open-source runtime designed to safely run autonomous AI agents in isolated environments. Developed by NVIDIA, it provides sandboxed execution spaces that protect system resources, credentials, and data from unauthorized access. Each agent runs inside a containerized sandbox governed by declarative YAML security policies that control network access, file permissions, and process behavior.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 18
    langrocks

    langrocks

    Tools like web browser, computer access and code runner for LLMs

    Langrocks is a programming language experimentation toolkit that enables developers to create, test, and optimize custom programming languages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    NullClaw

    NullClaw

    Fastest, smallest, and fully autonomous AI assistant infrastructure

    NullClaw is the smallest fully autonomous AI assistant infrastructure, built entirely in Zig as a single static binary with zero runtime dependencies. At just 678 KB with ~1 MB peak RAM usage, it boots in under 2 milliseconds and runs on virtually any hardware, including low-cost ARM boards. Despite its size, it delivers a complete AI stack with 22+ model providers, 18+ communication channels, integrated tools, hybrid memory, and sandboxed runtime support. Its architecture is fully modular, using vtable interfaces that allow providers, channels, tools, memory backends, and runtimes to be swapped without code changes. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 20
    Monty

    Monty

    A minimal, secure Python interpreter written in Rust for use by AI

    ...Because it’s written in Rust, it’s positioned to deliver a compact, portable runtime that can be embedded into larger systems that need dependable isolation.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 21
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 152 This Week
    Last Update:
    See Project
  • 22
    Office Agents

    Office Agents

    Agent plugins for Microsoft Office but BYOK for any model and provider

    ...It enables users to interact with large language models inside tools like Excel, Word, and PowerPoint, allowing real-time automation, content generation, and data manipulation within familiar productivity environments. The system is built as a monorepo with multiple packages, including a core SDK for agent runtime, a React-based chat interface, and specialized add-ins tailored to each Office application. It supports a bring-your-own-key model, allowing users to connect to a wide range of AI providers, including OpenAI, Anthropic, and other compatible APIs. The framework also includes advanced capabilities such as a sandboxed execution environment, virtual file systems, and extensible skill modules that expand agent functionality.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Torch-TensorRT

    Torch-TensorRT

    PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

    Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate into the JIT runtime seamlessly. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 24
    Godot MCP

    Godot MCP

    MCP server for interfacing with Godot game engine

    Godot MCP is a Model Context Protocol server that enables AI assistants to directly interact with the Godot game engine, allowing programmatic control over game development workflows through natural language or agent-driven commands. It acts as a bridge between AI systems and the Godot editor, providing capabilities such as launching projects, running games in debug mode, and capturing runtime output for analysis. The tool is particularly valuable for AI-assisted game development, as it creates a feedback loop where agents can execute code, observe results, and iteratively improve their outputs. It also includes advanced features for manipulating scenes, managing assets, and editing project structures, making it possible to automate large portions of the development process. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    OpenSandbox

    OpenSandbox

    OpenSandbox is a general-purpose sandbox platform for AI applications

    OpenSandbox is a general purpose sandbox platform designed to securely run and isolate AI applications and untrusted workloads in controlled environments. The project focuses on providing a unified sandbox API that simplifies the process of executing code safely across different runtime backends. It supports multiple programming languages through SDKs, allowing developers to integrate sandbox capabilities into their systems without building custom isolation layers. The platform is built to work with container technologies such as Docker and Kubernetes, enabling scalable and production ready deployments. OpenSandbox is particularly useful for AI agents, code execution services, and any scenario where untrusted code must be executed safely. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB