Showing 10 open source projects for "npp-compare"

View related business solutions
  • Agentic AI SRE built for Engineering and DevOps teams. Icon
    Agentic AI SRE built for Engineering and DevOps teams.

    No More Time Lost to Troubleshooting

    NeuBird AI's agentic AI SRE delivers autonomous incident resolution, helping team cut MTTR up to 90% and reclaim engineering hours lost to troubleshooting.
    Learn More
  • Information Security Made Simple and Affordable | Carbide Icon
    Information Security Made Simple and Affordable | Carbide

    For companies requiring a solution to scale their business without incurring security debt

    Get expert guidance and smart tools to launch or level up your security and compliance efforts without the complexity.
    Learn More
  • 1
    promptfoo

    promptfoo

    Evaluate and compare LLM outputs, catch regressions, improve prompts

    Ensure high-quality LLM outputs with automatic evals. Use a representative sample of user inputs to reduce subjectivity when tuning prompts. Use built-in metrics, LLM-graded evals, or define your own custom metrics. Compare prompts and model outputs side-by-side, or integrate the library into your existing test/CI workflow. Use OpenAI, Anthropic, and open-source models like Llama and Vicuna, or integrate custom API providers for any LLM API.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    ...It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with various AI providers, enabling users to route tasks across different models depending on their needs. Every Code also includes browser integration and automation capabilities, extending its usefulness beyond simple code generation into more complex development tasks. ...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 3
    Ollama Grid Search

    Ollama Grid Search

    A multi-platform desktop application to evaluate and compare LLM

    ...The system integrates directly with local or remote Ollama servers, enabling seamless access to models already deployed in a user’s environment. It also includes experiment logging and A/B testing capabilities, which allow users to compare outputs side by side and track performance metrics such as latency or token usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Empirical

    Empirical

    Test and evaluate LLMs and model configurations

    Empirical is the fastest way to test different LLMs and model configurations, across all the scenarios that matter for your application.
    Downloads: 0 This Week
    Last Update:
    See Project
  • The ultimate digital workspace. Icon
    The ultimate digital workspace.

    Axero Intranet is an award-winning intranet and employee experience platform.

    Hundreds of companies and millions of employees use Axero’s intranet software to communicate, collaborate, manage tasks and events, organize content, and develop their company culture.
    Learn More
  • 5
    DVC Extension for Visual Studio Code

    DVC Extension for Visual Studio Code

    https://github.com/iterative/vscode-dvc

    A Visual Studio Code extension that integrates Data Version Control (DVC) into the development environment, enhancing reproducibility and collaboration for machine learning projects.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Learning Interpretability Tool

    Learning Interpretability Tool

    Interactively analyze ML models to understand their behavior

    The Learning Interpretability Tool (LIT, formerly known as the Language Interpretability Tool) is a visual, interactive ML model-understanding tool that supports text, image, and tabular data. It can be run as a standalone server, or inside of notebook environments such as Colab, Jupyter, and Google Cloud Vertex AI notebooks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    LangWatch

    LangWatch

    The platform for LLM evaluations and AI agent testing

    ...The platform includes dashboards that visualize model behavior, enabling teams to monitor trends in response quality and reliability over time. It also provides evaluation tools that allow developers to test prompts and compare outputs across different models or configurations. Through integration with popular AI development frameworks, LangWatch can be embedded directly into AI pipelines to provide continuous monitoring and evaluation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    ChainForge

    ChainForge

    An open-source visual programming environment

    ChainForge is an open-source visual programming environment designed to help developers systematically test, compare, and evaluate prompts and outputs across multiple large language models in a structured and scalable way. Instead of relying on isolated prompt experimentation, it introduces a dataflow-based interface that allows users to create complex prompt pipelines and evaluate them across different models, parameters, and datasets simultaneously.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    dLLM

    dLLM

    dLLM: Simple Diffusion Language Modeling

    ...The project provides an integrated pipeline that standardizes how diffusion language models are trained, evaluated, and deployed, helping researchers reproduce experiments and compare results more easily. The framework includes scalable training infrastructure inspired by modern deep learning toolkits and supports integrations with widely used libraries for distributed training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Intelligent Retail Management Icon
    Intelligent Retail Management

    Retail space, product categories, planograms, automatic ordering, and shelf labels management

    Quant offers a wide range of solutions for retail. Within one integrated software system, it allows you to efficiently combine the management of retail space, shelf labels and marketing materials with task management, reporting and automatic replenishment.
    Learn More
  • 10
    FastbuildAI

    FastbuildAI

    An open-source AI framework for developers and entrepreneurs

    ...It emphasizes fast iteration: you describe tasks declaratively, wire up tools with typed schemas, and let the runtime handle planning, retries, and result aggregation. The project leans into reproducibility with run records, seed control, and structured traces so you can compare behaviors across versions and inputs. Prompt and memory management are treated as first-class concerns, enabling short-lived scratchpads for reasoning as well as long-horizon state when an agent operates over multiple sessions. The codebase favors small, composable pieces—executors, routers, guards—so teams can adopt just what they need instead of buying into a monolith. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB