Showing 495 open source projects for "fast linux"

View related business solutions
  • AI Powered Global HCM for the Evolving World of Work Icon
    AI Powered Global HCM for the Evolving World of Work

    For Start-ups, SME's, Large Enterprise

    Darwinbox is a new-age & disruptive mobile-first, cloud-based HRMS platform built for the large enterprises to attract, engage and nurture their most critical resource - talent. It is an end-to-end integrated HR system that aids in streamlining activities across the employee lifecycle (Hire to Retire). Our powerful enterprise product features are built with a clear focus on intuitiveness and scalability, with standards of best in class consumer apps. Darwinbox’s motto is to engage, empower, and inspire employees on one side in addition to automating and simplifying all HR processes for the enterprise on the other. Over 350+ leading enterprises with 850k users manage their entire employee lifecycle on this unified platform.
    Learn More
  • More Bookings. Better Experience. Icon
    More Bookings. Better Experience.

    For tour and activity providers

    The all-in-one solution built to help you stay organised and get more bookings with thousands of connections to online travel agencies (OTAs), resellers and suppliers.
    Learn More
  • 1
    Fast MCP

    Fast MCP

    A Ruby Implementation of the Model Context Protocol

    Fast MCP is a lightweight framework designed to simplify the development and deployment of servers that implement the Model Context Protocol. The Model Context Protocol enables AI assistants and applications to connect with external tools, services, and data sources through a standardized interface. Fast-mcp provides developers with a streamlined toolkit for building MCP servers that expose application functionality to AI agents. The framework focuses on ease of use, allowing developers to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    fast-stable-diffusion

    fast-stable-diffusion

    Fast-stable-diffusion + DreamBooth

    fast-stable-diffusion is a community-curated GitHub repository that provides Colab notebooks and integration examples for running Stable Diffusion and associated UIs like AUTOMATIC1111, ComfyUI, and DreamBooth directly on Google Colab environments. Rather than being a standalone packaged application, this project offers ready-to-use interactive notebooks that install and launch full-feature Stable Diffusion web UIs inside Colab without requiring complex local setups or GPU installations....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    course.fast.ai

    course.fast.ai

    The fast.ai course notebooks

    course22 is the official repository containing the notebooks, slides, and supporting materials for the 2022 edition of the fast.ai course Practical Deep Learning for Coders. The repository serves as the core educational resource for the course, providing learners with hands-on exercises and coding tutorials that accompany each lecture. The project emphasizes learning deep learning through experimentation rather than purely theoretical study, encouraging students to build models and analyze...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    ZeroClaw

    ZeroClaw

    Fast, small, and fully autonomous AI assistant infrastructure

    ZeroClaw is a Rust-native autonomous AI agent framework engineered for teams and developers who need highly efficient, secure, and modular AI automation infrastructure that can run reliably in both production and self-hosted environments. It is designed around a trait-based architecture so that model providers, communication channels, memory systems, and tooling integrations can be swapped or extended without rewriting core components, giving engineers flexibility and long-term...
    Downloads: 46 This Week
    Last Update:
    See Project
  • Supercharge Your Manufacturing with Easy MRP and MES Software Icon
    Supercharge Your Manufacturing with Easy MRP and MES Software

    Designed for SME manufacturers who want to reduce wasteful manual processing, save time and increase profits.

    Flowlens eliminates stock-outs, shortage and overstocks, avoiding costly production delays. Stay in control of inventory levels and keep production running smoothly with real-time visibility and easy-to-use stock management. Import bulk data with ease.
    Learn More
  • 5
    Ninjabot

    Ninjabot

    A fast cryptocurrency platform for trading bot in Go

    A fast cryptocurrency trading bot framework implemented in Go. Ninjabot permits users to create and test custom strategies for spot markets. Ninjabot is an open-source platform that provides tools to implement custom strategies and backtests for trading cryptocurrencies in Go. Ninjabot CLI provides utilities commands to support backtesting and bot development. Currently, we only support Binance exchange. If you want to include support for other exchanges, you need to implement a new struct...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 6
    FastSD CPU

    FastSD CPU

    Fast stable diffusion on CPU and AI PC

    FastSD CPU is an optimized fork of Stable Diffusion designed to run efficiently on CPUs and devices without dedicated GPUs by leveraging Latent Consistency Models and Adversarial Diffusion Distillation techniques that accelerate inference. It focuses on bringing fast text-to-image generation to mainstream hardware like desktop CPUs, lower-end laptops, or edge devices without requiring high-end graphics processors. The repository contains multiple interfaces including a desktop GUI for simple...
    Downloads: 43 This Week
    Last Update:
    See Project
  • 7
    Zvec

    Zvec

    A lightweight, lightning-fast, in-process vector database

    Zvec is an open-source, lightweight, in-process vector database designed to embed directly into applications and serve fast similarity search workloads without the overhead of a separate server process. Developed by Alibaba’s Tongyi Lab, it positions itself as the “SQLite of vector databases” by being easy to integrate, minimal in dependencies, and capable of handling high throughput with low latency on edge devices or small systems. Zvec excels at approximate nearest neighbor search and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    PyTorch

    PyTorch

    Open source machine learning framework

    PyTorch is a Python package that offers Tensor computation (like NumPy) with strong GPU acceleration and deep neural networks built on tape-based autograd system. This project allows for fast, flexible experimentation and efficient production. PyTorch consists of torch (Tensor library), torch.autograd (tape-based automatic differentiation library), torch.jit (a compilation stack [TorchScript]), torch.nn (neural networks library), torch.multiprocessing (Python multiprocessing), and...
    Downloads: 106 This Week
    Last Update:
    See Project
  • 9
    EnvPool

    EnvPool

    C++-based high-performance parallel environment execution engine

    EnvPool is a fast, asynchronous, and parallel RL environment library designed for scaling reinforcement learning experiments. Developed by SAIL at Singapore, it leverages C++ backend and Python frontend for extremely high-speed environment interaction, supporting thousands of environments running in parallel on a single machine. It's compatible with Gymnasium API and RLlib, making it suitable for scalable training pipelines.
    Downloads: 11 This Week
    Last Update:
    See Project
  • The CI/CD Platform built for Mobile DevOps Icon
    The CI/CD Platform built for Mobile DevOps

    For mobile app developers interested in a powerful CI/CD platform for mobile app development and mobile DevOps

    Save time, money, and developer frustration with fast, flexible, and scalable mobile CI/CD that just works. Whether you swear by native or would rather go cross-platform, we have you covered. From Swift to Objective-C, Java to Kotlin, as well as Xamarin, Cordova, Ionic, React Native, and Flutter: Whatever you choose, we will automatically configure your initial workflows and have you building in minutes.
    Learn More
  • 10
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 152 This Week
    Last Update:
    See Project
  • 11
    PicoClaw

    PicoClaw

    Ultra-Efficient AI Assistant in Go

    ...The project’s goals include broad platform support (including Linux, macOS, and multiple CPU architectures), rapid startup times that make the assistant feel responsive, and integration with popular messaging platforms via gateways or bots.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 12
    MCP Everything Search

    MCP Everything Search

    An MCP server that provides fast file searching capabilities

    Everything Search MCP Server is an MCP server that provides fast file searching capabilities across Windows, macOS, and Linux. On Windows, it utilizes the Everything SDK; on macOS, it leverages the built-in mdfind command; and on Linux, it uses the locate or plocate command. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    vLLM

    vLLM

    A high-throughput and memory-efficient inference and serving engine

    vLLM is a fast and easy-to-use library for LLM inference and serving. High-throughput serving with various decoding algorithms, including parallel sampling, beam search, and more.
    Downloads: 43 This Week
    Last Update:
    See Project
  • 15
    CatBoost

    CatBoost

    High-performance library for gradient boosting on decision trees

    CatBoost is a fast, high-performance open source library for gradient boosting on decision trees. It is a machine learning method with plenty of applications, including ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. CatBoost offers superior performance over other GBDT libraries on many datasets, and has several superb features. It has best in class prediction speed, supports both numerical and categorical features, has a fast and scalable...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Keras

    Keras

    Python-based neural networks API

    Python Deep Learning library
    Downloads: 11 This Week
    Last Update:
    See Project
  • 17
    nanobot

    nanobot

    🐈 nanobot: The Ultra-Lightweight Clawdbot / OpenClaw

    nanobot is an ultra-lightweight personal AI assistant designed to deliver powerful agent capabilities without unnecessary complexity. Built in just ~4,000 lines of clean, readable code, it offers a minimalist alternative to heavyweight agent frameworks while retaining core intelligence and extensibility. nanobot is optimized for speed and efficiency, enabling fast startup times and low resource usage across environments. Its research-ready architecture makes it easy for developers to...
    Downloads: 24 This Week
    Last Update:
    See Project
  • 18
    FastMCP

    FastMCP

    The fast, Pythonic way to build Model Context Protocol servers

    FastMCP is a fast, Pythonic framework for building servers and clients using the Model Context Protocol (MCP). It abstracts away protocol complexity like serialization, validation, and error handling, letting developers focus entirely on their business logic. With simple decorators, you can expose Python functions as tools, resources, or prompts that AI agents can safely and efficiently use. FastMCP introduces clear abstractions—components, providers, and transforms—that make it easy to...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    Every Code (often referred to simply as Code) is a fast, local AI-powered coding agent designed to run directly in the terminal environment. It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 20
    Kitten TTS

    Kitten TTS

    State-of-the-art TTS model under 25MB

    KittenTTS is an open-source, ultra-lightweight, and high-quality text-to-speech model featuring just 15 million parameters and a binary size under 25 MB. It is designed for real-time CPU-based deployment across diverse platforms. Ultra-lightweight, model size less than 25MB. CPU-optimized, runs without GPU on any device. High-quality voices, several premium voice options available. Fast inference, optimized for real-time speech synthesis.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 21
    CTranslate2

    CTranslate2

    Fast inference engine for Transformer models

    CTranslate2 is a C++ and Python library for efficient inference with Transformer models. The project implements a custom runtime that applies many performance optimization techniques such as weights quantization, layers fusion, batch reordering, etc., to accelerate and reduce the memory usage of Transformer models on CPU and GPU. The execution is significantly faster and requires less resources than general-purpose deep learning frameworks on supported models and tasks thanks to many...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 22
    bloop

    bloop

    bloop is a fast code search engine written in Rust

    Bloop is an AI-powered code search tool designed to help developers quickly find relevant code snippets, documentation, and usage examples within large repositories. It provides natural language search capabilities and AI-enhanced recommendations for improving code discovery.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    Fast State-of-the-art tokenizers, optimized for both research and production. Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Beta9

    Beta9

    Run serverless GPU workloads with fast cold starts on bare-metal

    beta9 is a platform that enables running serverless GPU workloads with fast cold starts on bare-metal servers globally. It allows developers to deploy and scale GPU-accelerated applications without managing underlying infrastructure, offering flexibility and efficiency for AI and high-performance computing tasks. beta9 supports various frameworks and provides tools for monitoring and managing deployments effectively.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    spaCy

    spaCy

    Industrial-strength Natural Language Processing (NLP)

    spaCy is a library built on the very latest research for advanced Natural Language Processing (NLP) in Python and Cython. Since its inception it was designed to be used for real world applications-- for building real products and gathering real insights. It comes with pretrained statistical models and word vectors, convolutional neural network models, easy deep learning integration and so much more. spaCy is the fastest syntactic parser in the world according to independent benchmarks, with...
    Downloads: 11 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next