Fast, flexible and easy to use probabilistic modelling in Python
MoBA: Mixture of Block Attention for Long-Context LLMs
Wan2.2: Open and Advanced Large-Scale Video Generative Model
Open-source, high-performance AI model with advanced reasoning
A Powerful Native Multimodal Model for Image Generation
A Next-Generation Training Engine Built for Ultra-Large MoE Models
Qwen3-Coder is the code version of Qwen3
From nobody to big model (LLM) hero
Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Wan2.1: Open and Advanced Large-Scale Video Generative Model
Powerful AI language model (MoE) optimized for efficiency/performance
Open-weight, large-scale hybrid-attention reasoning model
Decentralized deep learning in PyTorch. Built to train models
Fully automatic censorship removal for language models
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Mixture-of-Experts Vision-Language Models for Advanced Multimodal
Ling is a MoE LLM provided and open-sourced by InclusionAI
kaldi-asr/kaldi is the official location of the Kaldi project
Open-source large language model family from Tencent Hunyuan
Large-language-model & vision-language-model based on Linear Attention
Qwen3-omni is a natively end-to-end, omni-modal LLM
Open-source, high-performance Mixture-of-Experts large language model
Run Mixtral-8x7B models in Colab or consumer desktops
Towards Ultimate Expert Specialization in Mixture-of-Experts Language