Training and deploying machine learning models on Amazon SageMaker
Powering Amazon custom machine learning chips
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
A high-throughput and memory-efficient inference and serving engine
The official Python client for the Huggingface Hub
Everything you need to build state-of-the-art foundation models
Single-cell analysis in Python
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Uplift modeling and causal inference with machine learning algorithms
A Pythonic framework to simplify AI service building
Uncover insights, surface problems, monitor, and fine tune your LLM
Operating LLMs in production
Gaussian processes in TensorFlow
A unified framework for scalable computing
Set of comprehensive computer vision & machine intelligence libraries
AIMET is a library that provides advanced quantization and compression
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
DoWhy is a Python library for causal inference
State-of-the-art Parameter-Efficient Fine-Tuning
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Optimizing inference proxy for LLMs
Integrate, train and manage any AI models and APIs with your database
GPU environment management and cluster orchestration
MII makes low-latency and high-throughput inference possible