Showing 201 open source projects for "llama-cpp-static"

View related business solutions
  • Supercharge Your Manufacturing with Easy MRP and MES Software Icon
    Supercharge Your Manufacturing with Easy MRP and MES Software

    Designed for SME manufacturers who want to reduce wasteful manual processing, save time and increase profits.

    Flowlens eliminates stock-outs, shortage and overstocks, avoiding costly production delays. Stay in control of inventory levels and keep production running smoothly with real-time visibility and easy-to-use stock management. Import bulk data with ease.
    Learn More
  • Securely stream and govern industrial data to power intelligent operations with agentic insights. Icon
    Securely stream and govern industrial data to power intelligent operations with agentic insights.

    For IoT Developers, Solution Architects, Technical Architects, CTOs, OT/IT Engineers

    Trusted MQTT Platform — Fully-managed and cloud-native MQTT platform for bi-directional IoT data movement.
    Learn More
  • 1
    LLamaSharp

    LLamaSharp

    C#/.NET binding of llama.cpp, including LLaMa/GPT model inference

    The C#/.NET binding of llama.cpp. It provides APIs to infer the LLaMa Models and deploy it on the local environment. It works on both Windows, Linux and MAC without the requirement for compiling llama.cpp yourself. Its performance is close to llama.cpp. Furthermore, it provides integrations with other projects such as BotSharp to provide higher-level applications and UI.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    CodeLlama

    CodeLlama

    Inference code for CodeLlama models

    Code Llama is a family of Llama-based code models optimized for programming tasks such as code generation, completion, and repair, with variants specialized for base coding, Python, and instruction following. The repo documents the sizes and capabilities (e.g., 7B, 13B, 34B) and highlights features like infilling and large input context to support real IDE workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    LLM Foundry

    LLM Foundry

    LLM training code for MosaicML foundation models

    Introducing MPT-7B, the first entry in our MosaicML Foundation Series. MPT-7B is a transformer trained from scratch on 1T tokens of text and code. It is open source, available for commercial use, and matches the quality of LLaMA-7B. MPT-7B was trained on the MosaicML platform in 9.5 days with zero human intervention at a cost of ~$200k. Large language models (LLMs) are changing the world, but for those outside well-resourced industry labs, it can be extremely difficult to train and deploy these models. This has led to a flurry of activity centered on open-source LLMs, such as the LLaMA series from Meta, the Pythia series from EleutherAI, the StableLM series from StabilityAI, and the OpenLLaMA model from Berkeley AI Research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    bert4torch

    bert4torch

    An elegent pytorch implement of transformers

    An elegant PyTorch implement of transformers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Accounting practice management software Icon
    Accounting practice management software

    Accountants, accounting firms, tax attorneys, tax professionals

    Canopy is a cloud-based practice management software for accounting and tax firms, offering tools for client engagement, document management, workflow automation, and time & billing. Its Client Engagement platform centralizes interactions with a secure portal, customizable branding, and email integration, while the Document Management system enables organized, paperless file storage. The Workflow module enhances visibility into tasks and projects through templates, task assignments, and automation, reducing human error. Additionally, the Time & Billing feature tracks billable hours, generates invoices, and processes payments, ensuring accurate financial management. With its comprehensive features, Canopy streamlines operations, reduces stress, and enhances client experiences.
    Learn More
  • 5
    Text Generation Web UI

    Text Generation Web UI

    Oobabooga - The definitive Web UI for local AI, with powerful features

    ...Very efficient text streaming. Parameter presets, 8-bit mode. Layers splitting across GPU(s), CPU, and disk. CPU mode, FlexGen, DeepSpeed ZeRO-3, API with streaming and without streaming. LLaMA model, including 4-bit GPTQ. RWKV model, LoRA (loading and training), Softprompts, and extensions.
    Downloads: 44 This Week
    Last Update:
    See Project
  • 6
    Eliza

    Eliza

    Autonomous agents for everyone

    ...Full support for voice, text, and media interactions. Built-in RAG memory system, document processing, media analysis, and autonomous trading capabilities. Supports multiple AI models including Llama, GPT-4, and Claude. Create custom actions, add new platform integrations, and extend functionality through a modular plugin system. Full TypeScript support.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Jan.ai

    Jan.ai

    Open source alternative to ChatGPT that runs 100% offline

    ...It allows you to download and run LLMs (local language models) offline while also offering optional integration with cloud-based model providers—giving you full control over your data and AI interactions. Download and run LLMs (Llama, Gemma, Qwen, GPT-oss etc.) from HuggingFace. Connect to GPT models via OpenAI, Claude models via Anthropic, Mistral, Groq, and others. Create specialized AI assistants for your tasks. MCP integration for agentic capabilities.
    Downloads: 32 This Week
    Last Update:
    See Project
  • 8
    TRIBE v2

    TRIBE v2

    A multimodal model for brain response prediction

    ...It is designed for in-silico neuroscience, enabling researchers to model how the brain responds to complex real-world inputs. The system integrates state-of-the-art encoders—including LLaMA for text, V-JEPA for video, and Wav2Vec-BERT for audio—into a unified Transformer architecture. This combined representation is mapped onto the cortical surface to predict fMRI responses across thousands of brain regions. TRIBE v2 allows researchers to simulate and analyze brain activity without requiring direct human experiments. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    DeepSeek R1

    DeepSeek R1

    Open-source, high-performance AI model with advanced reasoning

    ...This approach has resulted in performance comparable to leading models like OpenAI's o1, while maintaining cost-efficiency. To further support the research community, DeepSeek has released distilled versions of the model based on architectures such as LLaMA and Qwen.
    Downloads: 58 This Week
    Last Update:
    See Project
  • DeskTime is a cloud-based time tracking software Icon
    DeskTime is a cloud-based time tracking software

    DeskTime is best for medium to large companies, as well as freelancers who want to boost productivity without overworking.

    DeskTime is a high-performance, automated time tracking and workforce management solution for teams and freelancers. It runs silently in the background, logging computer activity from the moment of boot-up to ensure 100% accurate data without the need for manual timers.
    Learn More
  • 10
    fullmoon

    fullmoon

    Chat with private and local large language models

    ...Users can personalize the app by adjusting themes, fonts, and system prompts, and it integrates with Apple's Shortcuts for enhanced functionality. Fullmoon supports models like Llama-3.2-1B-Instruct-4bit and Llama-3.2-3B-Instruct-4bit, facilitating efficient on-device AI interactions without the need for an internet connection.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    LLM-Finetuning

    LLM-Finetuning

    LLM Finetuning with peft

    ...Instead of requiring specialized hardware or complex training pipelines, many examples are designed to run in cloud notebook environments such as Google Colab. The repository includes step-by-step notebooks demonstrating how to fine-tune models such as LLaMA, Falcon, OPT, Vicuna, and GPT-NeoX. These tutorials show how developers can adapt pretrained models for tasks such as chatbots, classification, and instruction following. The project also illustrates how low-precision training techniques and adapter-based methods reduce memory requirements while maintaining strong model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Orpheus TTS

    Orpheus TTS

    Towards Human-Sounding Speech

    Orpheus TTS is a state-of-the-art open-source text-to-speech system built on a Llama-3B backbone, treating speech synthesis as a large language model problem instead of a traditional TTS pipeline. It is designed to produce human-like speech with natural intonation, emotion, and rhythm, targeting quality comparable to or better than many closed-source systems. The project ships both pretrained and finetuned English models, as well as a family of multilingual models released as a research preview, and includes data-processing scripts so users can train or finetune their own variants. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    llamafile

    llamafile

    Distribute and run LLMs with a single file

    ...We're doing that by combining llama.cpp with Cosmopolitan Libc into one framework that collapses all the complexity of LLMs down to a single-file executable (called a "llamafile") that runs locally on most computers, with no installation. The easiest way to try it for yourself is to download our example llamafile for the LLaVA model (license: LLaMA 2, OpenAI). LLaVA is a new LLM that can do more than just chat; you can also upload images and ask it questions about them. With llamafile, this all happens locally; no data ever leaves your computer.
    Downloads: 36 This Week
    Last Update:
    See Project
  • 14
    GPT4All

    GPT4All

    Run Local LLMs on Any Device. Open-source

    GPT4All is an open-source project that allows users to run large language models (LLMs) locally on their desktops or laptops, eliminating the need for API calls or GPUs. The software provides a simple, user-friendly application that can be downloaded and run on various platforms, including Windows, macOS, and Ubuntu, without requiring specialized hardware. It integrates with the llama.cpp implementation and supports multiple LLMs, allowing users to interact with AI models privately. This...
    Downloads: 118 This Week
    Last Update:
    See Project
  • 15
    RTP-LLM

    RTP-LLM

    Alibaba's high-performance LLM inference engine for diverse apps

    ...The framework is designed for large-scale AI services and is already used internally across several Alibaba platforms such as Taobao, Amap, and other business systems that rely on conversational or search-related AI services. RTP-LLM supports a wide variety of modern model architectures, including Qwen, DeepSeek, and Llama-based models, making it a flexible engine for deploying many different open-source LLMs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Jlama

    Jlama

    Jlama is a modern LLM inference engine for Java

    ...This allows organizations to integrate generative AI features into their systems while maintaining full control over data privacy and infrastructure. The engine supports a wide range of open-source model architectures and formats, including variants of Llama, Mistral, and other transformer-based models. It provides tools for running chat interactions, completing prompts, or exposing an OpenAI-compatible REST API for applications that expect standard LLM endpoints. The project focuses on performance and portability by using native Java optimizations and the Java Vector API to accelerate inference workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LightLLM

    LightLLM

    LightLLM is a Python-based LLM (Large Language Model) inference

    ...LightLLM is designed to handle large-scale model workloads in production environments, supporting efficient batching and GPU utilization for fast inference across multiple requests. Its architecture allows models to be deployed with minimal overhead while maintaining compatibility with popular transformer-based model families such as LLaMA and GPT-style architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    SillyTavern

    SillyTavern

    LLM Frontend for Power Users

    Mobile-friendly, Multi-API (KoboldAI/CPP, Horde, NovelAI, Ooba, OpenAI, OpenRouter, Claude, Scale), VN-like Waifu Mode, Horde SD, System TTS, WorldInfo (lorebooks), customizable UI, auto-translate, and more prompt options than you'd ever want or need. Optional Extras server for more SD/TTS options + ChromaDB/Summarize. SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters you or the community create. ...
    Downloads: 320 This Week
    Last Update:
    See Project
  • 19
    Clippy

    Clippy

    Clippy, now with some AI

    ...Users interact with the system through a simple animated assistant interface that can answer questions, generate text, and perform conversational tasks. The application includes one-click installation support for several popular models such as Meta’s Llama, Google’s Gemma, and other open models.
    Downloads: 40 This Week
    Last Update:
    See Project
  • 20
    Enchanted

    Enchanted

    Enchanted is iOS and macOS app for chatting with language models

    ...Designed to work seamlessly with servers like Ollama, it provides a privacy-focused alternative to traditional cloud AI UIs by connecting directly to your own LLM endpoints such as Llama, Mistral, Vicuna, and more. The interface resembles familiar commercial chat apps but emphasizes local control, offline capabilities, and multimodal support, making it ideal for users who want rich AI interaction without exposing sensitive prompts or conversations to third-party services. Enchanted enables features like voice prompts, image attachments, and markdown formatting within chats, giving flexibility for both casual and professional use. ...
    Downloads: 31 This Week
    Last Update:
    See Project
  • 21
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines. Additionally, it supports features such as...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 22
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Lepton AI

    Lepton AI

    A Pythonic framework to simplify AI service building

    A Pythonic framework to simplify AI service building. Cutting-edge AI inference and training, unmatched cloud-native experience, and top-tier GPU infrastructure. Ensure 99.9% uptime with comprehensive health checks and automatic repairs.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    Axolotl

    Axolotl

    Go ahead and axolotl questions

    Axolotl is a powerful and flexible framework for fine-tuning large language models on custom datasets. Built for researchers and developers, Axolotl simplifies the process of adapting LLMs for specific tasks, including chat, code generation, and instruction following. It supports a wide variety of model architectures and offers out-of-the-box optimization strategies for efficient training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    OpenLLM

    OpenLLM

    Operating LLMs in production

    ...With OpenLLM, you can run inference with any open-source large-language models, deploy to the cloud or on-premises, and build powerful AI apps. Built-in supports a wide range of open-source LLMs and model runtime, including Llama 2, StableLM, Falcon, Dolly, Flan-T5, ChatGLM, StarCoder, and more. Serve LLMs over RESTful API or gRPC with one command, query via WebUI, CLI, our Python/Javascript client, or any HTTP client.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB