Showing 107 open source projects for "compare"

View related business solutions
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • Endpoint Protection Software for Businesses | HYPERSECURE Icon
    Endpoint Protection Software for Businesses | HYPERSECURE

    DriveLock protects systems, data, end devices from data loss and misuse.

    The HYPERSECURE endpoint protection platform is a comprehensive suite of products and services enhanced by European third-party solutions. It ensures our customers’ IT security, regulatory compliance, and digital sovereignty.
    Learn More
  • 1
    promptfoo

    promptfoo

    Evaluate and compare LLM outputs, catch regressions, improve prompts

    Ensure high-quality LLM outputs with automatic evals. Use a representative sample of user inputs to reduce subjectivity when tuning prompts. Use built-in metrics, LLM-graded evals, or define your own custom metrics. Compare prompts and model outputs side-by-side, or integrate the library into your existing test/CI workflow. Use OpenAI, Anthropic, and open-source models like Llama and Vicuna, or integrate custom API providers for any LLM API.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Google AI Edge Gallery

    Google AI Edge Gallery

    A gallery that showcases on-device ML/GenAI use cases

    Gallery is a curated collection of on-device machine learning examples, demo apps, and model artifacts designed to help developers experiment with and deploy ML at the edge. The project bundles runnable samples that show how to run TensorFlow Lite/Edge TPU models (and similar lightweight runtimes) on mobile and embedded platforms, demonstrating common tasks like image classification, object detection, audio recognition, and pose estimation. Each sample is intended to be both a learning aid...
    Downloads: 1,018 This Week
    Last Update:
    See Project
  • 3
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    ...It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with various AI providers, enabling users to route tasks across different models depending on their needs. Every Code also includes browser integration and automation capabilities, extending its usefulness beyond simple code generation into more complex development tasks. ...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 4
    TruLens

    TruLens

    Evaluation and Tracking for LLM Experiments

    ...An easy-to-use interface that allows developers to compare different versions of their applications, facilitating informed decision-making and optimization. TruLens supports various use cases, including question-answering, summarization, retrieval-augmented generation, and agent-based applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Contract Management Software | Concord Icon
    Contract Management Software | Concord

    AI-powered contract management that helps businesses track spending, negotiate smarter, and never miss deadlines.

    Concord serves small and mid-sized businesses and Fortune 500 companies. This robust, web-based platform is used by human resource, sales, procurement, and legal teams, and virtually anyone who deals with contracts.
    Learn More
  • 5
    pixelmatch

    pixelmatch

    The smallest, simplest JavaScript pixel-level image comparison library

    The smallest, simplest and fastest JavaScript pixel-level image comparison library, originally created to compare screenshots in tests. Features accurate anti-aliased pixels detection and perceptual color difference metrics. Inspired by Resemble.js and Blink-diff. Unlike these libraries, pixelmatch is around 150 lines of code, has no dependencies, and works on raw typed arrays of image data, so it's blazing fast and can be used in any environment (Node or browsers).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Ollama Grid Search

    Ollama Grid Search

    A multi-platform desktop application to evaluate and compare LLM

    ...The system integrates directly with local or remote Ollama servers, enabling seamless access to models already deployed in a user’s environment. It also includes experiment logging and A/B testing capabilities, which allow users to compare outputs side by side and track performance metrics such as latency or token usage.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Aim

    Aim

    An easy-to-use & supercharged open-source experiment tracker

    Aim logs all your AI metadata (experiments, prompts, etc) enabling a UI to compare & observe them and SDK to query them programmatically. The Aim standard package comes with all integrations. If you'd like to modify the integration and make it custom, create a new integration package and share with others. Aim is an open-source, self-hosted AI Metadata tracking tool designed to handle 100,000s of tracked metadata sequences.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    H2O LLM Studio

    H2O LLM Studio

    Framework and no-code GUI for fine-tuning LLMs

    ...With H2O LLM Studio, training your large language model is easy and intuitive. First, upload your dataset and then start training your model. Start by creating an experiment. You can then monitor and manage your experiment, compare experiments, or push the model to Hugging Face to share it with the community.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    SwanLab

    SwanLab

    An open-source, modern-design AI training tracking and visualization

    SwanLab is an open-source experiment tracking and visualization platform designed to help machine learning engineers monitor, compare, and analyze the training of artificial intelligence models. The tool records training metrics, hyperparameters, model outputs, and experiment configurations so that developers can easily understand how different experiments perform over time. It provides a modern user interface for visualizing results, enabling teams to compare runs, track model performance trends, and collaborate on machine learning research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Empower Your Workforce and Digitize Your Shop Floor Icon
    Empower Your Workforce and Digitize Your Shop Floor

    Benefits to Manufacturers

    Easily connect to most tools and equipment on the shop floor, enabling efficient data collection and boosting productivity with vital insights. Turn information into action to generate new ideas and better processes.
    Learn More
  • 10
    ReinforcementLearning.jl

    ReinforcementLearning.jl

    A reinforcement learning package for Julia

    A collection of tools for doing reinforcement learning research in Julia. Provide elaborately designed components and interfaces to help users implement new algorithms. Make it easy for new users to run benchmark experiments, compare different algorithms, and evaluate and diagnose agents. Facilitate reproducibility from traditional tabular methods to modern deep reinforcement learning algorithms. Make it easy for new users to run benchmark experiments, compare different algorithms, and evaluate and diagnose agents. Facilitate reproducibility from traditional tabular methods to modern deep reinforcement learning algorithms. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Prompt Master

    Prompt Master

    A Claude skill that writes the accurate prompts for any AI tool

    ...The project emphasizes clarity and organization, allowing users to categorize prompts by use case, domain, or functionality. It also supports experimentation, enabling users to refine prompts and compare results to achieve better outputs. The repository can be used as both a learning resource and a practical toolkit for developers working with AI systems. It reflects the growing importance of prompt engineering as a discipline in modern AI development.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 12
    Agentex

    Agentex

    Open source codebase for Scale Agentex

    ...It treats an “agent” as a composition of a policy (the LLM), tools, memory, and an execution runtime so you can test the whole loop, not just prompting. The repo focuses on structured experiments: standardized tasks, canonical tool interfaces, and logs that make it possible to compare models, prompts, and tool sets fairly. It also includes evaluation harnesses that capture success criteria and partial credit, plus traces you can inspect to understand where reasoning or tool use failed. The design encourages clean separation between experiment configuration and code, which makes sharing results or re-running baselines straightforward. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Empirical

    Empirical

    Test and evaluate LLMs and model configurations

    Empirical is the fastest way to test different LLMs and model configurations, across all the scenarios that matter for your application.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Norfair

    Norfair

    Lightweight Python library for adding real-time multi-object tracking

    ...Supports moving camera, re-identification with appearance embeddings, and n-dimensional object tracking. Norfair provides several predefined distance functions to compare tracked objects and detections. The distance functions can also be defined by the user, enabling the implementation of different tracking strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    DVC Extension for Visual Studio Code

    DVC Extension for Visual Studio Code

    https://github.com/iterative/vscode-dvc

    A Visual Studio Code extension that integrates Data Version Control (DVC) into the development environment, enhancing reproducibility and collaboration for machine learning projects.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MLflow

    MLflow

    Open source platform for the machine learning lifecycle

    MLflow is a platform to streamline machine learning development, including tracking experiments, packaging code into reproducible runs, and sharing and deploying models. MLflow offers a set of lightweight APIs that can be used with any existing machine learning application or library (TensorFlow, PyTorch, XGBoost, etc), wherever you currently run ML code (e.g. in notebooks, standalone applications or the cloud).
    Downloads: 7 This Week
    Last Update:
    See Project
  • 17
    Learning Interpretability Tool

    Learning Interpretability Tool

    Interactively analyze ML models to understand their behavior

    The Learning Interpretability Tool (LIT, formerly known as the Language Interpretability Tool) is a visual, interactive ML model-understanding tool that supports text, image, and tabular data. It can be run as a standalone server, or inside of notebook environments such as Colab, Jupyter, and Google Cloud Vertex AI notebooks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    PapersGPT

    PapersGPT

    A powerful Zotero AI and MCP plugin with ChatGPT, Gemini 3.1, Claude

    ...One of its most powerful features is its ability to process large volumes of academic content quickly, enabling tasks such as literature reviews, theoretical analysis, and research synthesis to be completed significantly faster. It also supports multi-document querying, allowing users to compare findings across multiple papers and generate comprehensive overviews of research topics.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    SAM 2

    SAM 2

    The repository provides code for running inference with SAM 2

    ...SAM2 comes with pretrained weights and easy-to-use APIs, enabling developers and researchers to integrate promptable segmentation into annotation tools, vision pipelines, or downstream tasks. The project also includes scripts and notebooks to compare SAM2 against SAM on edge cases, benchmarks showing improvements, and evaluation suites to measure mask quality metrics like IoU and boundary error.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    Opik

    Opik

    Debug, evaluate, and monitor your LLMapps, RAG systems, and agentic AI

    ...Opik is an open-source platform for evaluating, testing, and monitoring LLM applications. Built by Comet. Record, sort, search, and understand each step your LLM app takes to generate a response. Manually annotate, view, and compare LLM responses in a user-friendly table. Log traces during development and in production. Run experiments with different prompts and evaluate against a test set. Choose and run pre-configured evaluation metrics or define your own with our convenient SDK library. Consult built-in LLM judges for complex issues like hallucination detection, factuality, and moderation.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    Advanced + Agentic RAG Cookbooks

    Advanced + Agentic RAG Cookbooks

    Advanced RAG cookbooks for building accurate LLM applications

    ...Athina AI’s RAG Cookbooks covers the full RAG pipeline, including indexing, retrieval, augmentation, and generation, while also addressing evaluation to measure accuracy and relevance. It includes multiple approaches such as hybrid search, contextual compression, and agent-based retrieval strategies, allowing users to experiment and compare methods. It is designed to reduce development time by offering practical examples and references to research papers, making it useful for both learning and production use. Overall, it serves as a hands-on resource for improving LLM outputs using external data sources.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    Skills Janitor

    Skills Janitor

    Audit, track usage, and compare your Claude Code skills

    The Skills Janitor project is a lightweight plugin designed to manage, audit, and optimize AI agent skill ecosystems, particularly for environments like Claude Code and OpenAI Codex. It functions as a “maintenance layer” for AI skills by automatically scanning installed skill directories, identifying duplicates, and analyzing their structure and usage. One of its core purposes is to help developers maintain a clean and efficient skill environment, especially as the number of installed skills...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    Weights and Biases

    Weights and Biases

    Tool for visualizing and tracking your machine learning experiments

    Use W&B to build better models faster. Track and visualize all the pieces of your machine learning pipeline, from datasets to production models. Quickly identify model regressions. Use W&B to visualize results in real time, all in a central dashboard. Focus on the interesting ML. Spend less time manually tracking results in spreadsheets and text files. Capture dataset versions with W&B Artifacts to identify how changing data affects your resulting models. Reproduce any model, with saved...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    Agent Stack

    Agent Stack

    Deploy and share agents with open infrastructure

    ...The platform supports agents built in frameworks like LangChain, CrewAI, etc., enabling them to be hosted, managed and shared through a unified interface. It also offers multi-model, multi-provider support (OpenAI, Anthropic, Gemini, IBM WatsonX, Ollama etc.), letting users compare performance and cost across models. For developers and organizations building AI-agent products or automations, Agent Stack gives a scaffold that handles the “plumbing”, so they can focus on logic and domain.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 25
    Advanced RAG Techniques

    Advanced RAG Techniques

    Advanced techniques for RAG systems

    ...It includes hands-on Jupyter notebooks and runnable scripts that show how to implement ideas like optimizing chunk sizes, proposition chunking, HyDE/HyPE query transformations, fusion retrieval, reranking, and ensemble retrieval. There is also an evaluation section that demonstrates how to measure RAG performance and compare different configurations in a systematic way.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB