TensorRT LLM provides users with an easy-to-use Python API
Development repository for the Triton language and compiler
AI agents running research on single-GPU nanochat training
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
A sound cloning tool with a web interface, using your voice
A NumPy-compatible array library accelerated by CUDA
Public CI, Docker images for popular JAX libraries
Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model
Run your own AI cluster at home with everyday devices
Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real
Fast and memory-efficient exact attention
Traditional Mandarin LLMs for Taiwan
AI video generator optimized for low VRAM and older GPUs use
Free open source tool for real-time PC hardware sensor monitoring
Performance meets Productivity
A nearly-live implementation of OpenAI's Whisper
The Triton Inference Server provides an optimized cloud
Document content and metadata extraction microservice
A set of Docker images for training and serving models in TensorFlow
Style-Bert-VITS2: Bert-VITS2 with more controllable voice styles
InvokeAI is a leading creative engine for Stable Diffusion models
950 line, minimal, extensible LLM inference engine built from scratch
Simplest working implementation of Stylegan2
Ongoing research training transformer models at scale
Interface for OuteTTS models