Ongoing research training transformer models at scale
A set of Docker images for training and serving models in TensorFlow
The Triton Inference Server provides an optimized cloud
MII makes low-latency and high-throughput inference possible
Library for OCR-related tasks powered by Deep Learning
Trainable models and NN optimization tools
Probabilistic reasoning and statistical analysis in TensorFlow
A unified framework for scalable computing
Library for serving Transformers models on Amazon SageMaker
Powering Amazon custom machine learning chips
Deep learning optimization library: makes distributed training easy
OpenMMLab Model Deployment Framework
A computer vision framework to create and deploy apps in minutes
Implementation of model parallel autoregressive transformers on GPUs
Toolkit for allowing inference and serving with MXNet in SageMaker