Training and deploying machine learning models on Amazon SageMaker
Powering Amazon custom machine learning chips
Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Single-cell analysis in Python
The official Python client for the Huggingface Hub
FlashInfer: Kernel Library for LLM Serving
Everything you need to build state-of-the-art foundation models
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Uplift modeling and causal inference with machine learning algorithms
Operating LLMs in production
Gaussian processes in TensorFlow
Uncover insights, surface problems, monitor, and fine tune your LLM
A unified framework for scalable computing
AIMET is a library that provides advanced quantization and compression
Easiest and laziest way for building multi-agent LLMs applications
DoWhy is a Python library for causal inference
Optimizing inference proxy for LLMs
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Superduper: Integrate AI models and machine learning workflows
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Integrate, train and manage any AI models and APIs with your database
Trainable models and NN optimization tools
State-of-the-art Parameter-Efficient Fine-Tuning
Multilingual Automatic Speech Recognition with word-level timestamps