• Diagnose and Resolve IT Issues in Real Time Icon
    Diagnose and Resolve IT Issues in Real Time

    Engage your employees and agents more efficiently with ScreenMeet as a seamless extension of your existing IT Service Delivery Platform.

    ScreenMeet’s unique combination of video calling, screen share, and remote desktop functionality lets you quickly diagnose hardware and software issues with no frustration.
    Learn More
  • All-in-One Inspection Software Icon
    All-in-One Inspection Software

    flowdit is a connected worker platform tailored for industry needs in commissioning, quality, maintenance, and EHS management.

    Optimize Frontline Operations: Elevate Equipment Uptime, Operational Excellence, and Safety with Connected Teams and Data, Including Issue Capture and Corrective Action.
    Learn More
  • 1
    GPU Hot

    GPU Hot

    Real-time NVIDIA GPU dashboard

    GPU Hot is an open-source, lightweight monitoring dashboard designed to provide real-time visibility into NVIDIA GPU performance across single machines or entire clusters. The project offers a self-hosted web interface that streams hardware metrics directly from GPU servers, enabling developers, ML engineers, and system administrators to observe GPU utilization and system behavior in real time through a browser. The dashboard collects and displays a wide range of performance metrics including temperature, memory usage, power consumption, clock speeds, fan speed, and active processes. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Superagent

    Superagent

    Superagent protects your AI applications

    Superagent is an open-source AI safety platform built to protect applications from prompt injections, data leaks, and harmful outputs. It embeds real-time safety directly into AI workflows, helping teams secure models before threats cause damage. Superagent provides guardrails that block jailbreaks, prompt manipulation, and sensitive data exfiltration. It includes redaction tools to remove PII, PHI, and secrets automatically from text. The platform also scans code repositories to detect AI-specific attack vectors like repo poisoning. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Transformer Explainer

    Transformer Explainer

    Learn How LLM Transformer Models Work with Interactive Visualization

    Transformer Explainer is an interactive visualization tool created to help users understand how transformer-based language models operate internally. The platform runs a lightweight GPT-2 model directly in the user’s browser and allows users to experiment with text prompts while observing the model’s internal operations. Through visual diagrams and interactive interfaces, the tool reveals how tokens are processed through layers such as embeddings, attention mechanisms, and feed-forward...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    NLUX

    NLUX

    The powerful Conversational AI JavaScript Library

    NLUX, short for Natural Language User Experience, is an open-source JavaScript and React library designed to simplify the creation of conversational interfaces powered by large language models. The library provides developers with prebuilt components and utilities that make it easy to integrate chat-based AI functionality into web applications. By using NLUX, developers can connect their applications to models such as ChatGPT or other LLM providers and create interactive conversational...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Empower Your Contact Center with Human-Like AI Conversations Icon
    Empower Your Contact Center with Human-Like AI Conversations

    Deliver faster resolutions, lower costs, and better CX without hiring another agent.

    Enterprise Bot, based in Switzerland, is a pioneer in Conversational AI, Process Automation, and Generative AI. With the trust of esteemed enterprise giants across industries like Generali, SIX, SBB, DHL, and SWICA, Enterprise Bot is revolutionizing both customer and employee experiences. Through its advanced integration with Large Language Models (LLM) such as ChatGPT and Llama 2, and its unique patent-pending DocBrain technology, the company delivers unparalleled personalization, active engagement, and omnichannel solutions across platforms like email, voice, and chat. Furthermore, Enterprise Bot integrates with existing core systems, such as SAP, CRMs, Confluence and more, and with its proprietary middleware, Blitzico, enables the AI to not only respond to queries but also take action to resolve them. This dedication to innovation in four main use case areas, Customer Support, Sales and Marketing, Knowledge Management and Digital Coworker, elevates both CX and employee productivity.
    Learn More
  • 5
    Generative AI for beginners with JS

    Generative AI for beginners with JS

    Join a time-traveling adventure where you meet history’s legends

    Generative AI with JavaScript is an educational repository created by Microsoft that teaches developers how to build applications powered by large language models using the JavaScript ecosystem. The project is structured as a multi-lesson curriculum that introduces the concepts, tools, and practical techniques required to create generative AI applications. Each lesson includes written explanations, hands-on exercises, quizzes, and supporting videos to help developers learn the material...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    gpu_poor

    gpu_poor

    Calculate token/s & GPU memory requirement for any LLM

    gpu_poor is an open-source tool designed to help developers determine whether their hardware is capable of running a specific large language model and to estimate the performance they can expect from it. The project focuses on calculating GPU memory requirements and predicted inference speed for different models, hardware configurations, and quantization strategies. By analyzing factors such as model size, context length, batch size, and GPU specifications, the system estimates how much VRAM...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    ChatLLM Web

    ChatLLM Web

    Chat with LLM like Vicuna totally in your browser with WebGPU

    ...To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions ≤ 112 are not supported. You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower. The first time you use the app, you will need to download the model. For the Vicuna-7b model that we are currently using, the download size is about 4GB. After the initial download, the model will be loaded from the browser cache for faster usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB