Open Source Rust Artificial Intelligence Software - Page 4

Rust Artificial Intelligence Software

View 13574 business solutions

Browse free open source Rust Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Rust Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Skillfully - The future of skills based hiring Icon
    Skillfully - The future of skills based hiring

    Realistic Workplace Simulations that Show Applicant Skills in Action

    Skillfully transforms hiring through AI-powered skill simulations that show you how candidates actually perform before you hire them. Our platform helps companies cut through AI-generated resumes and rehearsed interviews by validating real capabilities in action. Through dynamic job specific simulations and skill-based assessments, companies like Bloomberg and McKinsey have cut screening time by 50% while dramatically improving hire quality.
    Learn More
  • AestheticsPro Medical Spa Software Icon
    AestheticsPro Medical Spa Software

    Our new software release will dramatically improve your medspa business performance while enhancing the customer experience

    AestheticsPro is the most complete Aesthetics Software on the market today. HIPAA Cloud Compliant with electronic charting, integrated POS, targeted marketing and results driven reporting; AestheticsPro delivers the tools you need to manage your medical spa business. It is our mission To Provide an All-in-One Cutting Edge Software to the Aesthetics Industry.
    Learn More
  • 1
    Code2Prompt

    Code2Prompt

    Convert codebases into structured prompts optimized for LLM analysis

    code2prompt is an open source command line tool designed to convert an entire codebase into a structured prompt that can be easily used with large language models. It analyzes a project directory, gathers relevant source files, and formats them into a single prompt that includes the source tree and code content. This approach helps developers quickly provide full project context to AI models without manually copying files or assembling prompts. code2prompt is built in Rust and focuses on performance, enabling fast traversal of large repositories while maintaining low resource usage. It also respects common project conventions such as .gitignore, ensuring that unnecessary files are automatically excluded from the generated prompt. The generated output can be saved to a file, printed to standard output, or copied to the clipboard for immediate use. In addition to the core command line interface, the project also includes a library, Python bindings, and an MCP server.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Floneum

    Floneum

    Instant, controllable, local pre-trained AI models in Rust

    Floneum is an open-source platform for building AI-powered workflows using large language models through a visual and extensible interface. The system allows users to design complex AI pipelines using a drag-and-drop workflow builder rather than writing extensive code. It focuses on enabling developers and researchers to create language model applications that combine different tools, data sources, and AI capabilities into automated workflows. Floneum supports a plugin architecture that allows external components to extend the platform while maintaining isolation and security. Many plugins can be written in different programming languages and compiled to WebAssembly modules, allowing them to run safely within the system. The platform is implemented primarily in Rust and emphasizes performance, modularity, and local execution.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    HumanLayer

    HumanLayer

    Open source IDE for orchestrating AI coding agents in large codebases

    HumanLayer is an open source development environment designed to help developers orchestrate and manage AI coding agents working within complex software projects. It provides a framework and tooling that allow AI agents to research, plan, and implement changes in large codebases while maintaining structured workflows. It focuses on enabling AI-assisted development through coordinated agent workflows rather than isolated code generation tasks. HumanLayer integrates with modern AI models and coding assistants to automate tasks such as code research, planning, and implementation while maintaining a structured development process. HumanLayer introduces advanced context management techniques that help AI agents understand large repositories and operate effectively across multiple tasks. It also supports collaborative workflows where developers and AI agents can work together with human oversight and control.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    LangChain Rust

    LangChain Rust

    LangChain for Rust, the easiest way to write LLM-based programs

    LangChain Rust is an open-source Rust implementation inspired by the LangChain ecosystem for building applications powered by large language models. The library aims to provide Rust developers with a structured framework for orchestrating prompts, chains, agents, and external tools within LLM-driven workflows. By adapting LangChain concepts to the Rust programming language, the project emphasizes performance, safety, and efficient memory management. Developers can use the framework to build chatbots, autonomous agents, and knowledge-augmented AI systems that interact with external data sources. The library provides abstractions for model providers, prompt templates, conversation memory, and vector search integrations. It also enables the construction of multi-step pipelines where LLM outputs feed into subsequent actions or tool calls.
    Downloads: 3 This Week
    Last Update:
    See Project
  • SoftCo: Enterprise Invoice and P2P Automation Software Icon
    SoftCo: Enterprise Invoice and P2P Automation Software

    For companies that process over 20,000 invoices per year

    SoftCo Accounts Payable Automation processes all PO and non-PO supplier invoices electronically from capture and matching through to invoice approval and query management. SoftCoAP delivers unparalleled touchless automation by embedding AI across matching, coding, routing, and exception handling to minimize the number of supplier invoices requiring manual intervention. The result is 89% processing savings, supported by a context-aware AI Assistant that helps users understand exceptions, answer questions, and take the right action faster.
    Learn More
  • 5
    Luminal

    Luminal

    Deep learning at the speed of light

    Luminal is a framework designed to accelerate and simplify the development of systems-level data applications by using a typed, functional, and streaming-first approach. Instead of treating data processing as a series of ad-hoc scripts, Luminal models transformations as strongly typed building blocks that can be composed into reliable, scalable pipelines. The project emphasizes correctness and performance by requiring explicit types for the data flowing through transformations, reducing runtime surprises and allowing for highly optimized execution. It is particularly well-suited for data engineering workflows where large datasets must be processed incrementally, efficiently, and deterministically. The framework also includes a runtime capable of executing pipelines across multiple backends, making it flexible in cloud and local environments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    MagicAPI AI Gateway

    MagicAPI AI Gateway

    Built for demanding AI workflows

    The world's fastest AI Gateway proxy, written in Rust and optimized for maximum performance. This high-performance API gateway routes requests to various AI providers (OpenAI, GROQ) with streaming support, making it perfect for developers who need reliable and blazing-fast AI API access.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Refact Agent

    Refact Agent

    WebUI for Fine-Tuning and Self-hosting of Open-Source LLMs

    Refact is an AI-powered code assistant designed to enhance software development workflows. It integrates with code editors and provides suggestions, refactoring assistance, and debugging insights.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    DeepReasoning

    DeepReasoning

    High-performance API combining reasoning and creative AI models

    DeepReasoning is a high-performance large language model inference API designed to unify advanced reasoning and creative generation capabilities into a single system. It combines DeepSeek R1’s chain-of-thought reasoning with Claude’s strengths in code generation and conversational output, enabling more capable and balanced responses. DeepReasoning provides both an API and a chat interface, allowing developers and users to interact with the combined models in a streamlined way. It emphasizes real-time streaming responses, where reasoning and final answers can be delivered efficiently in a single flow. Built with Rust, the backend focuses on performance, low latency, and efficient resource usage for demanding workloads. It follows a bring-your-own-key approach, giving users full control over their API credentials and data privacy. Additionally, it supports self-hosting and extensive configuration, making it suitable for developers who want flexibility and control.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    HASH

    HASH

    The best way to use and work with blocks

    This is HASH's public monorepo which contains our public code, docs, and other key resources. HASH is a platform for decision-making, which helps you integrate, understand and use data in a variety of different ways. HASH does this by combining various different powerful tools together into one simple interface. These range from data pipelines and a graph database, through to an all-in-one workspace, no-code tool builder, and agent-based simulation engine. These exist at varying stages of maturity, and while some are polished, not all are ready for real-world production use. You can read more about our big-picture vision at hash.dev
    Downloads: 2 This Week
    Last Update:
    See Project
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • 10
    Sprite Fusion Pixel Snapper

    Sprite Fusion Pixel Snapper

    A tool to snap pixels to a perfect grid

    Sprite Fusion Pixel Snapper is a utility designed to eliminate sub-pixel rendering issues that often arise in pixel art, UI icons, and 2D sprite graphics when displayed on screens with high DPI or during motion animations. The tool works by adjusting sprite rendering coordinates and texture sampling so that every pixel aligns cleanly to the screen’s pixel grid, avoiding blurring, distortion, or unintended smoothing artifacts. This is especially important in pixel art games, retro-styled interactive media, or precise UI designs where crisp edges and predictable alignment are essential. SpriteFusion Pixel Snapper integrates with popular game engines and rendering pipelines to ensure that assets remain sharp across a broad range of resolutions and aspect ratios without requiring manual fiddling from artists or developers. It includes options for snapping modes, filtering overrides, and automatic correction detection that can be applied on export or at runtime.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Tokscale

    Tokscale

    A CLI tool for tracking token usage from OpenCode, Claude Code

    Tokscale is a CLI and terminal UI tool that tracks token usage and estimated cost across multiple AI coding assistants and development workflows. It treats tokens like a measurable resource, helping developers understand how much “AI energy” they are consuming over time and where it is being spent. The tool aggregates usage across supported platforms and presents it through interactive views that let users filter, sort, and explore trends without leaving the terminal. Tokscale also includes rich visualizations such as contribution-graph style summaries and daily breakdowns, making it easy to spot spikes, habits, and long-term patterns. For cost estimation, it can map usage to pricing data so developers can see what their activity implies financially across different models.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    gptee

    gptee

    LLMs done the UNIX-y way

    Output from a language model using standard input as the prompt. Now supporting GPT3.5 chat completions! gptee was designed for use within shell scripts and other programs and also works in interactive shells. You can compose commands and execute them in a script. Proceed with caution before running arbitrary shell scripts. Using a chat completion model (like gpt-3.5-turbo), you can then inject a system message with -s or --system messages. For davinci and other non-chat models, the output is prefixed to the prompt. Compose shell commands like you would in a script. Try with a custom model. By default gptee uses gpt-3.5-turbo.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    AppFlowy

    AppFlowy

    Bring projects, wikis, and teams together with AI.

    AppFlowy is an AI collaborative workspace where you can achieve more without losing control of your data. It is the best open source alternative to Notion, offering a 100% offline mode and self-hosting with a cloud service of your choice. Build a centralized workspace for your wiki, projects, and notes with AppFlowy. It allows you to organize and visualize your data in tables, Kanban boards, calendars, and more. You can filter and sort your data in any way you want. AppFlowy comes with a beautiful rich-text editor that goes beyond just text and bullet points, offering 20+ content types, easy-to-use customized themes, keyboard shortcuts, and color options. It supports real-time team collaboration, enabling you to work with your friends and teammates on the same document in real time, similar to Google Docs. AppFlowy is powered by AppFlowy AI, which is accessible, collaborative, and contextual. Supercharge any type of work in a collaborative team workspace.
    Downloads: 51 This Week
    Last Update:
    See Project
  • 14
    Shinkai: Local AI Agents

    Shinkai: Local AI Agents

    Shinkai allows you to create advanced AI (local) agents effortlessly

    Shinkai is a free, open-source AI platform that lets anyone create powerful AI agents without coding. These agents can collaborate with each other, handle complex tasks, and operate in decentralized crypto environments. Key Features: - No-Code Agent Creation - Build specialized agents (trading bots, sentiment trackers, etc.) with simple descriptions - Multi-Agent Collaboration - Agents work together to solve complex problems - Crypto Integration - Built-in support for decentralized payments and transactions - Flexible AI Models - Choose from cloud models (GPT-4, Claude) or run locally - Universal Compatibility - Works with Model Context Protocol (MCP) for cross-platform integration - Local Security - Crypto keys and computations stay on your device Shinkai transforms AI from single-task tools into collaborative, autonomous systems that can operate in decentralized networks while maintaining privacy and security.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 15
    Cake

    Cake

    Distributed LLM and StableDiffusion inference

    Cake is a compact, powerful toolkit that combines a flexible TCP/UDP proxy, port forwarding system, and connection manager designed for both development and penetration testing scenarios. It enables users to create complex networking flows where traffic can be proxied, relayed, and manipulated between endpoints — useful for debugging networked applications, inspecting protocols, or tunneling traffic through different hops. The tool is designed to work with multiple protocols and supports dynamic rule definitions so that incoming and outgoing connections can be routed, rewritten, or logged according to user-defined policies. Unlike many simple proxies, Cake can act as a full connection broker: it can bind to arbitrary interfaces, handle simultaneous upstream/downstream sessions, and apply traffic rules on the fly. This makes it suitable for troubleshooting tricky network behavior, simulating network conditions, or chaining services in a modular test environment.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    ChatGPT Client

    ChatGPT Client

    A ChatGPT client written in Rust

    A ChatGPT client written in Rust. The ChatGPT model is a large language model trained by OpenAI that is capable of generating human-like text. By providing it with a prompt, it can generate responses that continue the conversation or expand on the given prompt.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Chidori

    Chidori

    A reactive runtime for building durable AI agents

    A reactive runtime for building durable AI agents. Chidori is an open-source orchestrator, runtime, and IDE for building software in symbiosis with modern AI tools. When using Chidori, you author code with python or javascript, we provide a layer for interfacing with the complexities of AI models in long-running workflows. We have avoided the need for declaring a new language or SDK in order to provide these capabilities so that you can leverage software patterns that you are already familiar with.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Extractous

    Extractous

    Fast and efficient unstructured data extraction

    Extractous is a Rust-based unstructured data extraction library focused on fast local parsing of documents and other content-heavy files. Its purpose is to extract text and metadata efficiently from formats such as PDF, Word, HTML, email archives, images, and more, without depending on external APIs or separate parsing servers. The project emphasizes performance and low memory usage, and its maintainers describe it as a local-first alternative to heavier extraction stacks. For broader format support, the system combines its Rust core with ahead-of-time compiled Apache Tika shared libraries, which allows it to extend parsing coverage while still avoiding traditional server-based overhead. It also supports OCR for images and scanned documents through Tesseract, making it useful for document ingestion pipelines that include image-based or scanned inputs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    OpenAI Harmony

    OpenAI Harmony

    Renderer for the harmony response format to be used with gpt-oss

    Harmony is a response format developed by OpenAI for use with the gpt-oss model series. It defines a structured way for language models to produce outputs, including regular text, reasoning traces, tool calls, and structured data. By mimicking the OpenAI Responses API, Harmony provides developers with a familiar interface while enabling more advanced capabilities such as multiple output channels, instruction hierarchies, and tool namespaces. The format is essential for ensuring gpt-oss models operate correctly, as they are trained to rely on this structure for generating and organizing their responses. For users accessing gpt-oss through third-party providers like HuggingFace, Ollama, or vLLM, Harmony formatting is handled automatically, but developers building custom inference setups must implement it directly. With its flexible design, Harmony serves as the foundation for creating more interpretable, controlled, and extensible interactions with open-weight language models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    Paddler is an open-source LLM infrastructure platform designed to deploy, manage, and scale large language models on private infrastructure. The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The architecture is designed with privacy and cost control in mind, making it suitable for organizations that handle sensitive data or require predictable operational costs. Paddler also includes tools for monitoring, request buffering, and autoscaling integration so that deployments can adapt dynamically to changing workloads. A built-in administrative interface allows developers and operations teams to manage models, observe system performance, and test inference endpoints.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Rig

    Rig

    Rust framework for building modular and scalable LLM-powered apps

    Rig is an open source Rust framework designed to help developers build modular and scalable applications powered by large language models. It provides a unified set of abstractions that allow applications to interact with many AI model providers and vector databases through a single interface. Its architecture emphasizes modularity, enabling developers to integrate only the components and integrations they need for a specific application. Rig includes built-in support for agent workflows, allowing systems to perform multi-turn reasoning, tool calling, and retrieval-based tasks within structured pipelines. It also supports capabilities such as text generation, embeddings, transcription, image generation, and audio generation depending on the provider used. Developers can integrate language models into their software with minimal boilerplate while maintaining flexibility for complex AI workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    aigc

    aigc

    An e-book about the real-world application of LLM

    "Building Large Language Model Applications: Application Development and Architecture Design" is an open source e-book about the real-world application of LLM. It introduces the basics and applications of large language models, as well as how to build your own models. These include writing, developing, and managing prompts, exploring what the best large language models can bring, and pattern and architecture design for LLM application development.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    llm

    llm

    An ecosystem of Rust libraries for working with large language models

    llm is an ecosystem of Rust libraries for working with large language models - it's built on top of the fast, efficient GGML library for machine learning. The primary entry point for developers is the llm crate, which wraps the llm-base and the supported model crates. Documentation for the released version is available on Docs.rs. For end-users, there is a CLI application, llm-cli, which provides a convenient interface for interacting with supported models. Text generation can be done as a one-off based on a prompt, or interactively, through REPL or chat modes. The CLI can also be used to serialize (print) decoded models, quantize GGML files, or compute the perplexity of a model. It can be downloaded from the latest GitHub release or by installing it from crates.io.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    pipeless

    pipeless

    A computer vision framework to create and deploy apps in minutes

    Pipeless is an open-source computer vision framework to create and deploy applications without the complexity of building and maintaining multimedia pipelines. It ships everything you need to create and deploy efficient computer vision applications that work in real-time in just minutes. Pipeless is inspired by modern serverless technologies. It provides the development experience of serverless frameworks applied to computer vision. You provide some functions that are executed for new video frames and Pipeless takes care of everything else. You can easily use industry-standard models, such as YOLO, or load your custom model in one of the supported inference runtimes. Pipeless ships some of the most popular inference runtimes, such as the ONNX Runtime, allowing you to run inference with high performance on CPU or GPU out-of-the-box. You can deploy your Pipeless application with a single command to edge and IoT devices or the cloud.
    Downloads: 21 This Week
    Last Update:
    See Project
  • 25
    OpenChat for Linux

    OpenChat for Linux

    OpenChat for Linux — a fast, lightweight desktop client for ChatGPT

    OpenChat for Linux is a desktop client for ChatGPT / OpenAI Chat designed specifically for Linux. It’s built with Tauri (Rust) for low resource usage and stability, and it uses a “message window” approach (keeps a small active slice of the conversation and loads more as you scroll) so long chats don’t bog down or crash the app. Downloads are available in common Linux formats (AppImage, Debian package, tarball), with additional packaging manifests for Flatpak, Snap, RPM, AUR, and Nix.
    Downloads: 5 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB