Open Source Rust Artificial Intelligence Software - Page 3

Rust Artificial Intelligence Software

View 13574 business solutions

Browse free open source Rust Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Rust Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Simplify Purchasing For Your Business Icon
    Simplify Purchasing For Your Business

    Manage what you buy and how you buy it with Order.co, so you have control over your time and money spent.

    Simplify every aspect of buying for your business in Order.co. From sourcing products to scaling purchasing across locations to automating your AP and approvals workstreams, Order.co is the platform of choice for growing businesses.
    Learn More
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • 1
    BasedHardware

    BasedHardware

    Open source AI wearable platform for recording and summarizing speech

    Omi is an open source AI wearable platform designed to capture spoken conversations and convert them into useful digital information such as transcripts, summaries, and action items. It combines hardware, firmware, mobile applications, and backend services to create a complete ecosystem for voice-driven interaction. Users can connect the wearable device to a mobile phone and automatically record and transcribe meetings, conversations, and voice memos. Omi includes firmware for wearable hardware, a Flutter-based mobile companion application, backend services built with Python and FastAPI, and various SDKs for developers. These components work together to process audio, perform speech recognition, and integrate AI features such as summaries and automated actions. Developers can extend the platform by building plugins, integrations, and custom applications using provided SDKs and APIs. The repository also supports experimental hardware implementations.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 2
    Burn

    Burn

    Burn is a new comprehensive dynamic Deep Learning Framework

    Burn is a new comprehensive dynamic Deep Learning Framework from Tracel AI built using Rust with extreme flexibility, compute efficiency and portability as its primary goals. Burn emphasizes performance, flexibility, and portability for both training and inference. Developed in Rust, it is designed to empower machine learning engineers and researchers across industry and academia.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 3
    Crabtalk

    Crabtalk

    Agents daemon that hides nothing

    Crabtalk is a composable AI agent runtime designed to provide a minimal yet powerful foundation for building and orchestrating intelligent agents within a single lightweight binary. It is implemented in Rust and focuses on delivering high performance, reliability, and low overhead compared to more complex agent frameworks. The system is built around a small set of core primitives, including skills, memory, context isolation, and extensions, which together enable flexible and modular agent behavior. CrabTalk emphasizes simplicity by avoiding unnecessary abstractions, allowing developers to maintain full control over how agents operate and interact with their environment. One of its key design goals is to address common issues in multi-agent systems, such as context fragmentation and coordination inefficiencies, by providing clearer structure and tighter control over execution.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 4
    Lance

    Lance

    Modern columnar data format for ML and LLMs implemented in Rust

    Lance is a columnar data format that is easy and fast to version, query and train on. It’s designed to be used with images, videos, 3D point clouds, audio and of course tabular data. It supports any POSIX file systems, and cloud storage like AWS S3 and Google Cloud Storage.
    Downloads: 7 This Week
    Last Update:
    See Project
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • 5
    Rust Telegram Bot Library

    Rust Telegram Bot Library

    Rust Library for creating a Telegram Bot

    A library for writing your own Telegram bots.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    gptcommit

    gptcommit

    A git prepare-commit-msg hook for authoring commit messages with GPT-3

    A git prepare-commit-msg hook for authoring commit messages with GPT-3. With this tool, you can easily generate clear, comprehensive and descriptive commit messages letting you focus on writing code. To use gptcommit, simply run git commit as you normally would. The hook will automatically generate a commit message for you using a large language model like GPT. If you're not satisfied with the generated message, you can always edit it before committing. By default, gptcommit uses the GPT-3 model. Please ensure you have sufficient credits in your OpenAI account to use it. Commit messages are a key channel for developers to communicate their work with others, especially in code reviews. When making complex code changes, it can be tedious to thoroughly document the contents of each change. I often felt the impulse to just title my commit “fix bug” and move on. Surfacing these changes with gptcommit helps the author and reviewer by bringing attention to these additional changes.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 7
    Edgee

    Edgee

    AI gateway with token compression for Claude Code, Codex, and more

    Edgee is an edge-native execution platform designed to run AI-driven logic and data processing directly at the network edge, reducing latency and improving responsiveness for modern applications. It enables developers to deploy functions and workflows closer to users, allowing real-time processing without relying heavily on centralized cloud infrastructure. The platform is built to support event-driven architectures, where actions are triggered by incoming requests, user behavior, or external signals. It integrates AI capabilities into edge environments, making it possible to perform inference, personalization, and decision-making at the point of interaction. Edgee is optimized for performance and scalability, leveraging distributed execution to handle high volumes of requests efficiently. It also emphasizes developer experience, providing tools and abstractions that simplify deployment and orchestration across edge nodes.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    El Monitorro

    El Monitorro

    El Monitorro is a high-performance feed reader as a Telegram bot

    El Monitorro is a high-performance RSS/ATOM/JSON feed reader as a Telegram bot.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    Kheish

    Kheish

    Kheish: A multi-role LLM agent for tasks like code auditing

    Kheish is a framework designed for cybersecurity professionals to automate penetration testing tasks, providing tools to streamline security assessments.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Loan management software that makes it easy. Icon
    Loan management software that makes it easy.

    Ideal for lending professionals who are looking for a feature rich loan management system

    Bryt Software is ideal for lending professionals who are looking for a feature rich loan management system that is intuitive and easy to use. We are 100% cloud-based, software as a service. We believe in providing our customers with fair and honest pricing. Our monthly fees are based on your number of users and we have a minimal implementation charge.
    Learn More
  • 10
    Monoio

    Monoio

    Rust async runtime based on io-uring

    Monoio is a Rust asynchronous runtime designed for high-performance I/O-bound servers and applications, built around native OS async I/O primitives (e.g. io_uring on Linux, epoll / kqueue on other Unix-like systems), rather than layering atop an existing runtime. Its design philosophy centers on a “thread-per-core” model where each core runs its own event loop, minimizing cross-thread synchronization needs, avoiding the overhead and complexity of task scheduling, and letting developers write efficient, low-overhead asynchronous networking or I/O code. Because tasks do not need to be Send or Sync and can make use of thread-local data safely, Monoio simplifies certain concurrency paradigms while delivering performance benefits for workloads like high-throughput network servers, proxies, or real-time services. The runtime includes abstractions for async sockets, readers/writers, TCP/UDP networking, and compatibility layers (macros, crates) to ease adoption.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 11
    OWS - Open Wallet Standard

    OWS - Open Wallet Standard

    Local, policy-gated signing and wallet management for every chain

    Open Wallet Standard Core is a foundational library that defines a set of interfaces and conventions for integrating cryptocurrency wallets with decentralized applications in a consistent and interoperable way. It establishes a standardized method for wallets to expose functionality such as account management, transaction signing, and connection handling, enabling seamless interaction across different blockchain ecosystems. The project focuses on improving both developer experience and user experience by reducing fragmentation in how wallets communicate with applications. It includes mechanisms for registering wallets globally, detecting available wallets in browser environments, and handling events related to wallet connection or disconnection. The standard is designed to be extensible, allowing blockchain-specific features to be added through extensions without breaking compatibility.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 12
    Rewriting Project Claw Code

    Rewriting Project Claw Code

    Ensure consistency and alignment between different codebases

    Rewriting Project Claw Code is a development tool or framework designed to ensure consistency and alignment between different codebases, environments, or implementations. It focuses on maintaining parity across systems, which is particularly important in distributed architectures or multi-platform applications. The project provides mechanisms to compare, validate, and synchronize code or behavior, helping teams avoid discrepancies that can lead to bugs or inconsistencies. It may include automation tools that detect differences and enforce standards across repositories. The tool is useful in scenarios such as maintaining parity between frontend and backend logic, ensuring API consistency, or synchronizing multiple deployments. Its design promotes reliability and reduces the risk of divergence in complex systems. Overall, Claw Code Parity helps teams maintain cohesion across their codebases while improving development efficiency and quality assurance.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    Rust Port

    Rust Port

    The Rust workspace under rust/ is the current systems-language port

    Rust Port is an open-source reconstruction and experimentation framework derived from leaked or reverse-engineered versions of advanced AI coding agents, designed to replicate and extend the capabilities of agentic development systems. It functions as a programmable coding assistant that operates through autonomous workflows, enabling users to generate, modify, and analyze code with minimal manual intervention. The project emphasizes agent-based execution, where tasks are broken down into steps and handled iteratively, simulating how modern AI coding tools operate in production environments. It is often used as a sandbox for exploring how large-scale coding agents behave, including their decision-making processes, tool usage, and workflow orchestration. The system likely includes abstractions for handling file systems, executing commands, and maintaining context across sessions, allowing for more persistent and intelligent coding interactions.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 14
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to integrate embedding capabilities into applications without managing model internals directly. Text Embeddings Inference is optimized for throughput and low latency, enabling it to handle large volumes of requests reliably. It also emphasizes ease of deployment, often using containerization and configurable runtime options to adapt to different infrastructure setups.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 15
    dfdx

    dfdx

    Deep learning in Rust, with shape checked tensors and neural networks

    Deep learning in Rust, with shape-checked tensors and neural networks. Ergonomics & safety focused deep learning in Rust.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and efficiency, leveraging hardware acceleration across CPUs, GPUs, and specialized accelerators to deliver low-latency inference both on-device and in server environments. One of its key strengths is its flexibility, as it supports multiple backends and allows developers to configure execution providers depending on available hardware. ort also includes advanced capabilities such as model compilation and optimization, reducing startup time and improving runtime performance in production systems.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 17
    Monty

    Monty

    A minimal, secure Python interpreter written in Rust for use by AI

    Monty is an experimental, security-focused Python interpreter implemented in Rust and intended for running AI-generated Python safely under strict constraints. The project’s core goal is to enable code execution in environments where untrusted or model-produced code must be tightly sandboxed to reduce risk. Rather than offering a full “general-purpose Python runtime with everything enabled,” Monty is designed to be minimal and controlled, making it easier to reason about what code can do and what it cannot. It prioritizes guardrails like resource limits and restricted capabilities, which is especially useful for agentic workflows that need to execute small pieces of Python for data transforms, validation, or tool-like computations. Because it’s written in Rust, it’s positioned to deliver a compact, portable runtime that can be embedded into larger systems that need dependable isolation.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    Phantasm

    Phantasm

    Toolkits to create a human-in-the-loop approval layer

    Phantasm offers toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents' workflows in real-time, ensuring safety and reliability in AI operations.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 19
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    Fast State-of-the-art tokenizers, optimized for both research and production. Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also extremely versatile. Designed for both research and production. Full alignment tracking. Even with destructive normalization, it’s always possible to get the part of the original sentence that corresponds to any token. Does all the pre-processing: Truncation, Padding, add the special tokens your model needs.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 20
    claurst

    claurst

    Your favorite Terminal Coding Agent, now in Rust

    claurst is an experimental AI agent framework that appears to focus on structured reasoning and task execution within coding or automation environments. The project likely explores how agents can be designed to handle complex workflows through modular components and clearly defined execution steps. It may include abstractions for managing context, decision-making, and interaction with external tools, enabling agents to perform multi-step tasks efficiently. The architecture suggests a focus on flexibility, allowing developers to adapt the system to different use cases or domains. It is likely intended as a lightweight but extensible platform for experimenting with agent behavior and orchestration. The project may also emphasize simplicity, making it accessible for developers who want to prototype agent systems quickly.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    pgvecto.rs

    pgvecto.rs

    Vector database plugin for Postgres, written in Rust

    pgvecto.rs is a Postgres extension that provides vector similarity search functions. It is written in Rust and based on pgrx. It is currently under heavy development, please take care when using it in production. pgvecto.rs is a Postgres extension, which means that you can use it directly within your existing database. This makes it easy to integrate into your existing workflows and applications. pgvecto.rs supports filtering. You can set conditions when searching or retrieving points. This is the missing feature of other postgres extensions.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    Arch

    Arch

    Arch is an intelligent prompt gateway. Engineered with (fast) LLMs

    Arch is an intelligent Layer 7 gateway designed to protect, observe, and personalize LLM applications (agents, assistants, co-pilots) with your APIs. Engineered with purpose-built LLMs, Arch handles the critical but undifferentiated tasks related to the handling and processing of prompts, including detecting and rejecting jailbreak attempts, intelligently calling "backend" APIs to fulfill the user's request represented in a prompt, routing to and offering disaster recovery between upstream LLMs, and managing the observability of prompts and LLM interactions in a centralized way.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and device-mapping strategies, helping users get strong performance without manual tuning. It also supports serving multiple models from the same server process, enabling routing or quick switching between models depending on workload needs. For user-facing testing, mistral.rs can provide a built-in web UI, and it also offers a dedicated lightweight web chat interface that supports richer interaction patterns.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. Shimmy focuses on performance and simplicity, using efficient runtime components to minimize memory usage and startup time compared to heavier inference frameworks. It supports modern model formats such as GGUF and SafeTensors and can automatically discover models stored locally or in common directories used by other AI tools. Advanced capabilities include CPU offloading for Mixture-of-Experts models and GPU acceleration, enabling large models to run on consumer hardware with limited VRAM.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    Arch Gateway

    Arch Gateway

    The AI-native (edge and LLM) proxy for agents

    Arch is an AI-native proxy designed to facilitate the development of agentic applications by handling complex tasks such as input clarification, agent routing, and seamless integration of prompts with tools for common tasks. It provides unified access and observability of Large Language Models (LLMs), enabling developers to build applications more efficiently. Arch supports both edge and LLM deployments, offering flexibility in various environments.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB