Training and deploying machine learning models on Amazon SageMaker
Trainable models and NN optimization tools
Uplift modeling and causal inference with machine learning algorithms
Bring the notion of Model-as-a-Service to life
Lightweight Python library for adding real-time multi-object tracking
GPU environment management and cluster orchestration
Integrate, train and manage any AI models and APIs with your database
A library for accelerating Transformer models on NVIDIA GPUs
A unified framework for scalable computing
Superduper: Integrate AI models and machine learning workflows
A high-performance ML model serving framework, offers dynamic batching
The Triton Inference Server provides an optimized cloud
Powering Amazon custom machine learning chips
Phi-3.5 for Mac: Locally-run Vision and Language Models
Replace OpenAI GPT with another LLM in your app
Uncover insights, surface problems, monitor, and fine tune your LLM
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Probabilistic reasoning and statistical analysis in TensorFlow
Standardized Serverless ML Inference Platform on Kubernetes
State-of-the-art Parameter-Efficient Fine-Tuning
Library for OCR-related tasks powered by Deep Learning
Libraries for applying sparsification recipes to neural networks
Pytorch domain library for recommendation systems
Low-latency REST API for serving text-embeddings
A set of Docker images for training and serving models in TensorFlow