Showing 9 open source projects for "mini project in cloud computing for source code"

View related business solutions
  • Stigg | SaaS Monetization and Entitlements API Icon
    Stigg | SaaS Monetization and Entitlements API

    For developers in need of a tool to launch pricing plans faster and build better buying experiences

    A monetization platform is a standalone middleware that sits between your application and your business applications, as part of the modern enterprise billing stack. Stigg unifies all the APIs and abstractions billing and platform engineers had to build and maintain in-house otherwise. Acting as your centralized source of truth, with a highly scalable and flexible entitlements management, rolling out any pricing and packaging change is now a self-service, risk-free, exercise.
    Learn More
  • SIEM | API Security | Log Management Software Icon
    SIEM | API Security | Log Management Software

    AI-Powered Security and IT Operations Without Compromise.

    Built on the Graylog Platform, Graylog Security is the industry’s best-of-breed threat detection, investigation, and response (TDIR) solution. It simplifies analysts’ day-to-day cybersecurity activities with an unmatched workflow and user experience while simultaneously providing short- and long-term budget flexibility in the form of low total cost of ownership (TCO) that CISOs covet. With Graylog Security, security analysts can:
    Learn More
  • 1
    HolmesGPT

    HolmesGPT

    CNCF Sandbox Project

    ...The project is developed by Robusta and has been accepted as a Cloud Native Computing Foundation Sandbox project, highlighting its relevance to the cloud-native ecosystem. It is designed to operate as an automated troubleshooting assistant that can analyze incidents continuously and support on-call engineers during outages.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Clippy

    Clippy

    Clippy, now with some AI

    Clippy is an open-source desktop assistant that allows users to run modern large language models locally while presenting them through a nostalgic interface inspired by Microsoft’s classic Clippy assistant from the 1990s. The project serves as both a playful homage to the early days of personal computing and a practical demonstration of local AI inference.
    Downloads: 43 This Week
    Last Update:
    See Project
  • 3
    PicoLM

    PicoLM

    Run a 1-billion parameter LLM on a $10 board with 256MB RAM

    PicoLM is an open-source inference framework designed to run large language models on extremely constrained hardware environments such as inexpensive single-board computers and embedded systems. The project focuses on enabling efficient local inference by optimizing memory usage, computation, and system dependencies so that relatively large models can operate on devices with minimal RAM.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    E2B Cookbook

    E2B Cookbook

    Examples of using E2B

    E2B Cookbook is an open-source collection of example projects, guides, and reference implementations demonstrating how to build applications using the E2B platform. The repository acts as a practical learning resource for developers who want to integrate AI agents with secure cloud execution environments that allow large language models to run code and interact with tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Share your screen instantly while on a phone call with CrankWheel for an engaging presentation. Icon
    Share your screen instantly while on a phone call with CrankWheel for an engaging presentation.

    For salespeople and customer service agents who want to compliment their phone calls with visual elements.

    Our 10x simpler screen sharing tool is designed for you if you spend your days on the phone with clients, and need to add a visual presentation to close sales. No more scheduling a follow-up meeting, or teaching them to use a complex tool. Send them a text message or email, and they see your screen in seconds.
    Learn More
  • 5
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Generative AI Use Cases (GenU)

    Generative AI Use Cases (GenU)

    Application implementation with business use cases

    AWS Generative AI Use Cases is an open-source repository developed by AWS that provides practical examples and reference implementations for building applications powered by generative artificial intelligence. The project collects a wide range of real-world scenarios that demonstrate how organizations can use large language models and generative AI services within cloud-based architectures.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    ...This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. It also detects the user’s shell environment automatically, allowing it to generate commands tailored to shells such as Bash, Zsh, or PowerShell.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    RAG from Scratch

    RAG from Scratch

    Demystify RAG by building it from scratch

    RAG From Scratch is an educational open-source project designed to teach developers how retrieval-augmented generation systems work by building them step by step. Instead of relying on complex frameworks or cloud services, the repository demonstrates the entire RAG pipeline using transparent and minimal implementations. The project walks through key concepts such as generating embeddings, building vector databases, retrieving relevant documents, and integrating the retrieved context into...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Mixtral offloading

    Mixtral offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

    Mixtral-Offloading is an open-source project designed to enable efficient inference of large Mixture-of-Experts language models such as Mixtral-8x7B on hardware with limited GPU memory. The project implements techniques that allow model components to be dynamically moved between CPU memory and GPU memory during inference, significantly reducing the amount of GPU VRAM required to run the model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Monitor production, track downtime and improve OEE. Icon
    Monitor production, track downtime and improve OEE.

    For manufacturing companies interested in OEE monitoring solutions

    Evocon is a visual and user-friendly OEE software that helps manufacturing companies improve productivity and remove waste as they become better.
    Learn More
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB