Implementation of TurboQuant (ICLR 2026)
AIMET is a library that provides advanced quantization and compression
Accessible large language models via k-bit quantization for PyTorch
From-scratch PyTorch implementation of Google's TurboQuant
Libraries for applying sparsification recipes to neural networks
Minimal and clean examples of machine learning algorithms
Neural Network Compression Framework for enhanced OpenVINO
An implementation of a deep learning recommendation model (DLRM)
Open-source large language model family from Tencent Hunyuan
Build AI-powered semantic search applications
Pytorch domain library for recommendation systems
MiniSom is a minimalistic implementation of the Self Organizing Maps
Library to facilitate federated learning research
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs
A unified library of SOTA model optimization techniques
Machine learning on FPGAs using HLS
Z80-μLM is a 2-bit quantized language model
A Python package for extending the official PyTorch
Low-code framework for building custom LLMs, neural networks
A Python package for segmenting geospatial data with the SAM
48khz stereo neural audio codec for general audio
The data structure for multimodal data
Pretrained (Language) Models for Probabilistic Time Series Forecasting
Build cross-modal and multimodal applications on the cloud
Topic Modelling for Humans