Showing 92 open source projects for "jtdx-improved"

View related business solutions
  • The Most Powerful Software Platform for EHSQ and ESG Management Icon
    The Most Powerful Software Platform for EHSQ and ESG Management

    Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

    Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
    Learn More
  • Skillfully - The future of skills based hiring Icon
    Skillfully - The future of skills based hiring

    Realistic Workplace Simulations that Show Applicant Skills in Action

    Skillfully transforms hiring through AI-powered skill simulations that show you how candidates actually perform before you hire them. Our platform helps companies cut through AI-generated resumes and rehearsed interviews by validating real capabilities in action. Through dynamic job specific simulations and skill-based assessments, companies like Bloomberg and McKinsey have cut screening time by 50% while dramatically improving hire quality.
    Learn More
  • 1
    Improved Diffusion

    Improved Diffusion

    Release for Improved Denoising Diffusion Probabilistic Models

    improved-diffusion is an open source implementation of diffusion probabilistic models created by OpenAI. These models, also known as score-based generative models, are a class of generative models that have shown strong performance in producing high-quality synthetic data such as images. The repository provides code for training and sampling diffusion models with improved techniques that enhance stability, efficiency, and output fidelity.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    TurboQuant+

    TurboQuant+

    Implementation of TurboQuant (ICLR 2026)

    ...It builds upon the concept of reducing model precision to accelerate inference while attempting to maintain or recover accuracy through refined techniques. The project explores additional enhancements such as improved calibration, adaptive quantization, and potentially hybrid precision approaches that combine multiple levels of compression. It is designed to be used in conjunction with modern machine learning workflows, particularly those involving large models that require optimization for deployment. TurboQuant Plus focuses on experimentation and performance tuning, allowing developers to test different configurations and evaluate trade-offs. ...
    Downloads: 47 This Week
    Last Update:
    See Project
  • 3
    GLM-4.6

    GLM-4.6

    Agentic, Reasoning, and Coding (ARC) foundation models

    ...The model achieves superior coding performance, excelling in benchmarks and practical coding assistants such as Claude Code, Cline, Roo Code, and Kilo Code. Its reasoning capabilities have been strengthened, including improved tool usage during inference and more effective integration within agent frameworks. GLM-4.6 also enhances writing quality, producing outputs that better align with human preferences and role-playing scenarios. Benchmark evaluations demonstrate that it not only outperforms GLM-4.5 but also rivals leading global models such as DeepSeek-V3.1-Terminus and Claude Sonnet 4.
    Downloads: 102 This Week
    Last Update:
    See Project
  • 4
    DINOv3

    DINOv3

    Reference PyTorch implementation and models for DINOv3

    ...DINOv3 removes the need for complex augmentations or momentum encoders, streamlining the pipeline while maintaining or improving feature quality. The model supports multiple backbone architectures, including Vision Transformers (ViT), and can handle larger image resolutions with improved stability during training. The learned embeddings generalize robustly across tasks like classification, retrieval, and segmentation without fine-tuning, showing state-of-the-art transfer performance among self-supervised models.
    Downloads: 22 This Week
    Last Update:
    See Project
  • Securely stream and govern industrial data to power intelligent operations with agentic insights. Icon
    Securely stream and govern industrial data to power intelligent operations with agentic insights.

    For IoT Developers, Solution Architects, Technical Architects, CTOs, OT/IT Engineers

    Trusted MQTT Platform — Fully-managed and cloud-native MQTT platform for bi-directional IoT data movement.
    Learn More
  • 5
    GLM-4.7

    GLM-4.7

    Advanced language and coding AI model

    ...The model introduces stronger “thinking before acting” behavior, improving stability and accuracy in complex agent frameworks like Claude Code, Cline, and Roo Code. GLM-4.7 also advances “vibe coding,” producing cleaner, more modern UIs, better-structured webpages, and visually improved slide layouts. Its tool-use capabilities are substantially enhanced, with notable improvements in browsing, search, and tool-integrated reasoning tasks. Overall, GLM-4.7 shows broad performance upgrades across coding, reasoning, chat, creative writing, and role-play scenarios.
    Downloads: 75 This Week
    Last Update:
    See Project
  • 6
    FlashInfer

    FlashInfer

    FlashInfer: Kernel Library for LLM Serving

    FlashInfer is a kernel library designed to enhance the serving of Large Language Models (LLMs) by optimizing inference performance. It provides a high-performance framework that integrates seamlessly with existing systems, aiming to reduce latency and improve efficiency in LLM deployments. FlashInfer supports various hardware architectures and is built to scale with the demands of production environments.
    Downloads: 36 This Week
    Last Update:
    See Project
  • 7
    GPT-SoVITS

    GPT-SoVITS

    1 min voice data can also be used to train a good TTS model

    GPT‑SoVITS is a state-of-the-art voice conversion and TTS system that enables zero‑shot and few‑shot synthesis based on a short vocal sample (e.g., 5 seconds). It supports cross‑lingual speech synthesis across English, Chinese, Japanese, Korean, Cantonese, and more. It's powered by VITS architecture enhanced for few‑sample adaptation and real‑time usability.
    Downloads: 57 This Week
    Last Update:
    See Project
  • 8
    Claw Code

    Claw Code

    AI agent harness for AI coding agents

    Claw Code is an open-source AI agent harness project focused on building better tools for orchestrating and managing autonomous coding agents. It originated as a clean-room reimplementation inspired by the architecture of Claude Code, aiming to replicate core concepts without using proprietary code. The project provides a Python-based foundation for experimenting with agent workflows, tool integration, and task execution pipelines. It emphasizes harness engineering—how agents are structured,...
    Downloads: 118 This Week
    Last Update:
    See Project
  • 9
    Text Generation Inference

    Text Generation Inference

    Large Language Model Text Generation Inference

    Text Generation Inference is a high-performance inference server for text generation models, optimized for Hugging Face's Transformers. It is designed to serve large language models efficiently with optimizations for performance and scalability.
    Downloads: 11 This Week
    Last Update:
    See Project
  • The Secure And Reliable File Transfer Solution That You Control. Icon
    The Secure And Reliable File Transfer Solution That You Control.

    Helping IT professionals responsibly secure the world's data

    Cerberus offers a variety of secure file transfer solutions to fit businesses of any size or business sector, including finance, technology, education, publishing, law offices, local, state, and federal government agencies, hospitals and many more.
    Learn More
  • 10
    xLSTM

    xLSTM

    Neural Network architecture based on ideas of the original LSTM

    ...The project introduces a new recurrent neural network design that incorporates exponential gating mechanisms and enhanced memory structures to overcome limitations of traditional LSTM models. By introducing innovations such as matrix-based memory and improved normalization techniques, xLSTM improves the ability of recurrent networks to capture long-range dependencies in sequential data. The architecture aims to provide competitive performance with transformer-based models while maintaining advantages such as linear computational scaling and efficient memory usage for long sequences. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    DeepSeek-V3

    DeepSeek-V3

    Powerful AI language model (MoE) optimized for efficiency/performance

    DeepSeek-V3 is a robust Mixture-of-Experts (MoE) language model developed by DeepSeek, featuring a total of 671 billion parameters, with 37 billion activated per token. It employs Multi-head Latent Attention (MLA) and the DeepSeekMoE architecture to enhance computational efficiency. The model introduces an auxiliary-loss-free load balancing strategy and a multi-token prediction training objective to boost performance. Trained on 14.8 trillion diverse, high-quality tokens, DeepSeek-V3...
    Downloads: 175 This Week
    Last Update:
    See Project
  • 12
    SetFit

    SetFit

    Efficient few-shot learning with Sentence Transformers

    SetFit is an efficient and prompt-free framework for few-shot fine-tuning of Sentence Transformers. It achieves high accuracy with little labeled data - for instance, with only 8 labeled examples per class on the Customer Reviews sentiment dataset, SetFit is competitive with fine-tuning RoBERTa Large on the full training set of 3k examples.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    HunyuanCustom

    HunyuanCustom

    Multimodal-Driven Architecture for Customized Video Generation

    ...The architecture builds on HunyuanVideo, with added modules for identity reinforcement and modality-specific condition injection. Text-image fusion module based on LLaVA for improved multimodal understanding. Applicable to single- and multi-subject scenarios, video editing/replacement, singing avatars etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    WikiChat

    WikiChat

    WikiChat is an improved RAG

    WikiChat is a chatbot framework designed to interactively retrieve and summarize Wikipedia information, allowing users to ask questions and get context-aware responses?
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    MiniCPM4.1

    MiniCPM4.1

    Achieving 3+ generation speedup on reasoning tasks

    ...One of its key innovations is the hybrid reasoning mode, which allows developers to control whether the model engages in deeper reasoning processes or faster responses depending on the use case. The model also supports both dense and sparse attention mechanisms, enabling more efficient computation depending on the selected inference framework. With improved pretraining on longer sequences and enhanced scaling techniques, MiniCPM4.1 delivers better performance in long-context tasks and complex problem solving.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Gymnasium

    Gymnasium

    An API standard for single-agent reinforcement learning environments

    Gymnasium is a fork of OpenAI Gym, maintained by the Farama Foundation, that provides a standardized API for reinforcement learning environments. It improves upon Gym with better support, maintenance, and additional features while maintaining backward compatibility.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 17
    WhisperX

    WhisperX

    Automatic Speech Recognition with Word-level Timestamps

    WhisperX is an advanced speech recognition system built on top of OpenAI’s Whisper model, designed to improve transcription accuracy and timing precision for long-form audio. It addresses key limitations of standard Whisper implementations by introducing voice activity detection and forced alignment techniques to produce word-level timestamps. The system enables batched inference, significantly increasing transcription speed while maintaining high accuracy. It is particularly effective for...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 18
    DSPy

    DSPy

    DSPy: The framework for programming—not prompting—language models

    Developed by the Stanford NLP Group, DSPy (Declarative Self-improving Python) is a framework that enables developers to program language models through compositional Python code rather than relying solely on prompt engineering. It facilitates the construction of modular AI systems and provides algorithms for optimizing prompts and weights, enhancing the quality and reliability of language model outputs.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 19
    AGiXT

    AGiXT

    AGiXT is a dynamic AI Automation Platform

    AGiXT is a dynamic Artificial Intelligence Automation Platform engineered to orchestrate efficient AI instruction management and task execution across a multitude of providers. Our solution infuses adaptive memory handling with a broad spectrum of commands to enhance AI's understanding and responsiveness, leading to improved task completion. The platform's smart features, like Smart Instruct and Smart Chat, seamlessly integrate web search, planning strategies, and conversation continuity, transforming the interaction between users and AI. By leveraging a powerful plugin system that includes web browsing and command execution, AGiXT stands as a versatile bridge between AI models and users. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    DeepSeek-OCR 2

    DeepSeek-OCR 2

    Visual Causal Flow

    DeepSeek-OCR-2 is the second-generation optical character recognition system developed to improve document understanding by introducing a “visual causal flow” mechanism, enabling the encoder to reorder visual tokens in a way that better reflects semantic structure rather than strict raster scan order. It is designed to handle complex layouts and noisy documents by giving the model causal reasoning capabilities that mimic human visual scanning behavior, enhancing OCR performance on documents...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 21
    CO3D (Common Objects in 3D)

    CO3D (Common Objects in 3D)

    Tooling for the Common Objects In 3D dataset

    CO3Dv2 (Common Objects in 3D, version 2) is a large-scale 3D computer vision dataset and toolkit from Facebook Research designed for training and evaluating category-level 3D reconstruction methods using real-world data. It builds upon the original CO3Dv1 dataset, expanding both scale and quality—featuring 2× more sequences and 4× more frames, with improved image fidelity, more accurate segmentation masks, and enhanced annotations for object-centric 3D reconstruction. CO3Dv2 enables research in multi-view 3D reconstruction, novel view synthesis, and geometry-aware representation learning. Each of the thousands of sequences in CO3Dv2 captures a common object (from categories like cars, chairs, or plants) from multiple real-world viewpoints. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CodiumAI Cover-Agent

    CodiumAI Cover-Agent

    CodiumAI Cover-Agent: An AI-Powered Tool for Automated Test Generation

    CodiumAI Cover Agent aims to help efficiently increasing code coverage, by automatically generating qualified tests to enhance existing test suites.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Audiogen Codec

    Audiogen Codec

    48khz stereo neural audio codec for general audio

    AGC (Audiogen Codec) is a convolutional autoencoder based on the DAC architecture, which holds SOTA. We found that training with EMA and adding a perceptual loss term with CLAP features improved performance. These codecs, being low compression, outperform Meta's EnCodec and DAC on general audio as validated from internal blind ELO games. We trained (relatively) very low compression codecs in the pursuit of solving a core issue regarding general music and audio generation, low acoustic quality, and audible artifacts, which hinder industry use for these models. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 24
    CogVideo

    CogVideo

    Text and image to video generation: CogVideoX and CogVideo

    ...Built on large-scale Transformer and diffusion architectures, it enables multimodal generation across text-to-video, image-to-video, and video continuation tasks. The latest CogVideoX models offer higher resolution outputs, longer video durations, and improved controllability through prompt engineering. The project includes tools for inference, fine-tuning, and optimization, making it suitable for both research and production use. It supports efficient deployment on a range of GPUs, including consumer hardware with quantization techniques. Overall, CogVideo provides a powerful framework for generating high-quality AI videos and experimenting with cutting-edge multimodal AI systems.
    Downloads: 21 This Week
    Last Update:
    See Project
  • 25
    x-transformers

    x-transformers

    A simple but complete full-attention transformer

    ...Proposes adding learned tokens, akin to CLS tokens, named memory tokens, that is passed through the attention layers alongside the input tokens. You can also use the l2 normalized embeddings proposed as part of fixnorm. I have found it leads to improved convergence when paired with small initialization (proposed by BlinkDL). The small initialization will be taken care of as long as l2norm_embed is set to True.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • Next
MongoDB Logo MongoDB