Showing 3 open source projects for "mini project in cloud computing for source code"

View related business solutions
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • Award-Winning Medical Office Software Designed for Your Specialty Icon
    Award-Winning Medical Office Software Designed for Your Specialty

    Succeed and scale your practice with cloud-based, data-backed, AI-powered healthcare software.

    RXNT is an ambulatory healthcare technology pioneer that empowers medical practices and healthcare organizations to succeed and scale through innovative, data-backed, AI-powered software.
    Learn More
  • 1
    HolmesGPT

    HolmesGPT

    CNCF Sandbox Project

    ...The project is developed by Robusta and has been accepted as a Cloud Native Computing Foundation Sandbox project, highlighting its relevance to the cloud-native ecosystem. It is designed to operate as an automated troubleshooting assistant that can analyze incidents continuously and support on-call engineers during outages.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 2
    E2B Cookbook

    E2B Cookbook

    Examples of using E2B

    E2B Cookbook is an open-source collection of example projects, guides, and reference implementations demonstrating how to build applications using the E2B platform. The repository acts as a practical learning resource for developers who want to integrate AI agents with secure cloud execution environments that allow large language models to run code and interact with tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Mixtral offloading

    Mixtral offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

    Mixtral-Offloading is an open-source project designed to enable efficient inference of large Mixture-of-Experts language models such as Mixtral-8x7B on hardware with limited GPU memory. The project implements techniques that allow model components to be dynamically moved between CPU memory and GPU memory during inference, significantly reducing the amount of GPU VRAM required to run the model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB