Training and deploying machine learning models on Amazon SageMaker
The official Python client for the Huggingface Hub
FlashInfer: Kernel Library for LLM Serving
DoWhy is a Python library for causal inference
A high-throughput and memory-efficient inference and serving engine
Adversarial Robustness Toolbox (ART) - Python Library for ML security
PyTorch library of curated Transformer models and their components
Data manipulation and transformation for audio signal processing
A library for accelerating Transformer models on NVIDIA GPUs
AIMET is a library that provides advanced quantization and compression
Library for serving Transformers models on Amazon SageMaker
A unified framework for scalable computing
Uncover insights, surface problems, monitor, and fine tune your LLM
Official inference library for Mistral models
Pytorch domain library for recommendation systems
PyTorch extensions for fast R&D prototyping and Kaggle farming
MII makes low-latency and high-throughput inference possible
Lightweight Python library for adding real-time multi-object tracking
Probabilistic reasoning and statistical analysis in TensorFlow
A Unified Library for Parameter-Efficient Learning
Deep learning optimization library: makes distributed training easy
Replace OpenAI GPT with another LLM in your app
Fast inference engine for Transformer models
A GPU-accelerated library containing highly optimized building blocks
State-of-the-art diffusion models for image and audio generation