Showing 2 open source projects for "phpldapadmin-1.2.3-php5.5.patch"

View related business solutions
  • Effortlessly manage macOS, iOS, iPadOS and tvOS devices across multiple clients and locations. Icon
    Effortlessly manage macOS, iOS, iPadOS and tvOS devices across multiple clients and locations.

    The Most Powerful Apple Device Management Tool for MSPs and IT Teams

    Addigy solutions accelerate Apple adoption in any environment.
    Learn More
  • RentGuruz is an all-in-one vehicle rental software solution designed to streamline operations for car rental businesses worldwide. Icon
    RentGuruz is an all-in-one vehicle rental software solution designed to streamline operations for car rental businesses worldwide.

    Auto rental businesses seeking a solution to manage all their cloud business needs

    RentGuruz. The simple, intuitive, and powerful cloud application platform that manages all kinds of mobility for all kinds of rental businesses.
    Learn More
  • 1
    Agentless

    Agentless

    An agentless approach to automatically solve software development

    Agentless is an open-source framework that applies large language models to automatically resolve software development issues without relying on complex autonomous agent systems. The project proposes an alternative approach to AI-driven code repair that avoids the overhead of multi-agent orchestration by using a structured pipeline for identifying and fixing bugs. When solving a problem, the system first performs localization to determine which files, functions, or code segments are most...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    MiniMax-01

    MiniMax-01

    Large-language-model & vision-language-model based on Linear Attention

    MiniMax-01 is the official repository for two flagship models: MiniMax-Text-01, a long-context language model, and MiniMax-VL-01, a vision-language model built on top of it. MiniMax-Text-01 uses a hybrid attention architecture that blends Lightning Attention, standard softmax attention, and Mixture-of-Experts (MoE) routing to achieve both high throughput and long-context reasoning. It has 456 billion total parameters with 45.9 billion activated per token and is trained with advanced parallel...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB