Showing 12 open source projects for "linux android"

View related business solutions
  • Papirfly: Best user-friendly DAM and Content Creation Software Icon
    Papirfly: Best user-friendly DAM and Content Creation Software

    The #1 solution to create and manage content. On‑brand. At scale.

    Papirfly provides a single online destination for all your employees and other stakeholders who are engaging with your brand, ensuring consistency in all aspects of their communications. Teams can produce infinite studio-standard marketing materials from bespoke templates, store, share and adapt them for their own markets and stay firmly educated on the brand’s purpose, guidelines and evolution – with no specialist skills or agency help necessary.
    Learn More
  • Simplify your short-term rental business with our Channel Manager Icon
    Simplify your short-term rental business with our Channel Manager

    Boost your vacation rental revenue

    Manage your vacation rental like a pro with features that streamline, sync and simplify - all from one intuitive software.
    Learn More
  • 1
    JADX-AI-MCP

    JADX-AI-MCP

    Plugin for JADX to integrate MCP server

    JADX-AI-MCP is an open-source plugin that integrates large language models into the JADX Android decompiler to assist with reverse engineering and code analysis tasks. The project connects JADX with AI assistants through the Model Context Protocol, enabling language models to interact directly with decompiled Android application code. Through this integration, AI systems can inspect classes, analyze methods, retrieve application manifests, and examine other elements of Android packages in...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 2
    DeepSeek-V3

    DeepSeek-V3

    Powerful AI language model (MoE) optimized for efficiency/performance

    DeepSeek-V3 is a robust Mixture-of-Experts (MoE) language model developed by DeepSeek, featuring a total of 671 billion parameters, with 37 billion activated per token. It employs Multi-head Latent Attention (MLA) and the DeepSeekMoE architecture to enhance computational efficiency. The model introduces an auxiliary-loss-free load balancing strategy and a multi-token prediction training objective to boost performance. Trained on 14.8 trillion diverse, high-quality tokens, DeepSeek-V3...
    Downloads: 158 This Week
    Last Update:
    See Project
  • 3
    SillyTavern

    SillyTavern

    LLM Frontend for Power Users

    Mobile-friendly, Multi-API (KoboldAI/CPP, Horde, NovelAI, Ooba, OpenAI, OpenRouter, Claude, Scale), VN-like Waifu Mode, Horde SD, System TTS, WorldInfo (lorebooks), customizable UI, auto-translate, and more prompt options than you'd ever want or need. Optional Extras server for more SD/TTS options + ChromaDB/Summarize. SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters...
    Downloads: 373 This Week
    Last Update:
    See Project
  • 4
    DeepSeek R1

    DeepSeek R1

    Open-source, high-performance AI model with advanced reasoning

    DeepSeek-R1 is an open-source large language model developed by DeepSeek, designed to excel in complex reasoning tasks across domains such as mathematics, coding, and language. DeepSeek R1 offers unrestricted access for both commercial and academic use. The model employs a Mixture of Experts (MoE) architecture, comprising 671 billion total parameters with 37 billion active parameters per token, and supports a context length of up to 128,000 tokens. DeepSeek-R1's training regimen uniquely...
    Downloads: 73 This Week
    Last Update:
    See Project
  • Silverware is an enterprise-grade hospitality platform built for hotels, resorts, and complex multi-venue operations. Icon
    Silverware is an enterprise-grade hospitality platform built for hotels, resorts, and complex multi-venue operations.

    Silverware powers high-end hospitality environments

    Silverware is built for hotel, resort, and multi-venue hospitality operators who need enterprise-grade control, deep integrations, and always-on reliability to run complex operations at scale.
    Learn More
  • 5
    Aidea

    Aidea

    Flutter-based cross-platform app integrating major AI models

    AIdea is a comprehensive Flutter-based cross-platform app integrating major AI models—OpenAI GPT, Chinese models Tongyi Qianwen and Wenxin Yiyan, plus image models like Stable Diffusion for text-to-image, image-to-image, SDXL 1.0, super-resolution, and colorization. It includes a client app, server backend, and Docker deployment scripts for hosted setups.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    RWKV Runner

    RWKV Runner

    A RWKV management and startup tool, full automation, only 8MB

    RWKV (pronounced as RwaKuv) is an RNN with GPT-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, fast training, saves VRAM, "infinite" ctxlen, and free text embedding. Moreover it's 100% attention-free. Default configs has enabled custom CUDA kernel acceleration, which is much faster and consumes much less VRAM. If you encounter possible compatibility...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    nndeploy

    nndeploy

    An Easy-to-Use and High-Performance AI Deployment Framework

    nndeploy is an open-source framework designed to simplify the deployment of artificial intelligence models across multiple hardware platforms and devices. The framework focuses on making it easier to transform trained AI models into production-ready applications that can run efficiently on desktops, mobile devices, servers, and edge computing hardware. Developers can use visual workflows to design and configure AI processing pipelines by connecting modular nodes that represent different...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    ...By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. The system supports deployment on environments including Linux, macOS, Windows, iOS, Android, and web browsers while utilizing different acceleration technologies such as CUDA, Vulkan, Metal, and WebGPU. It also provides OpenAI-compatible APIs that allow developers to integrate locally deployed models into existing AI applications without major code changes.
    Downloads: 30 This Week
    Last Update:
    See Project
  • 9
    Cake

    Cake

    Distributed LLM and StableDiffusion inference

    Cake is a compact, powerful toolkit that combines a flexible TCP/UDP proxy, port forwarding system, and connection manager designed for both development and penetration testing scenarios. It enables users to create complex networking flows where traffic can be proxied, relayed, and manipulated between endpoints — useful for debugging networked applications, inspecting protocols, or tunneling traffic through different hops. The tool is designed to work with multiple protocols and supports...
    Downloads: 0 This Week
    Last Update:
    See Project
  • The #1 solution for profitable resource management Icon
    The #1 solution for profitable resource management

    Designed to give Operations and Finance leaders the insight and foresight they need to achieve profitable delivery at scale.

    Unlike spreadsheets or clunky PSAs, Float offers a clear, centralized view to schedule teams, plan capacity, estimate work, and track margins in real-time so that you can keep your people and profits on track.
    Learn More
  • 10
    mllm

    mllm

    Fast Multimodal LLM on Mobile Devices

    mllm is an open-source inference engine designed to run multimodal large language models efficiently on mobile devices and edge computing environments. The framework focuses on delivering high-performance AI inference in resource-constrained systems such as smartphones, embedded hardware, and lightweight computing platforms. Implemented primarily in C and C++, it is designed to operate with minimal external dependencies while taking advantage of hardware-specific acceleration technologies...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    MiniMax-M2.5

    MiniMax-M2.5

    State of the art LLM and coding model

    MiniMax-M2.5 is a state-of-the-art foundation model extensively trained with reinforcement learning across hundreds of thousands of real-world environments. It delivers leading performance in coding, agentic tool use, search, and complex office workflows, achieving top benchmark scores such as 80.2% on SWE-Bench Verified and 76.3% on BrowseComp. Designed to reason efficiently and decompose tasks like an experienced architect, M2.5 plans features, structure, and system design before...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Alpaca.cpp

    Alpaca.cpp

    Locally run an Instruction-Tuned Chat-Style LLM

    Run a fast ChatGPT-like model locally on your device. This combines the LLaMA foundation model with an open reproduction of Stanford Alpaca a fine-tuning of the base model to obey instructions (akin to the RLHF used to train ChatGPT) and a set of modifications to llama.cpp to add a chat interface. Download the zip file corresponding to your operating system from the latest release. The weights are based on the published fine-tunes from alpaca-lora, converted back into a PyTorch checkpoint...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB