Cosmos-RL is a scalable reinforcement learning framework designed specifically for physical AI systems such as robotics, autonomous agents, and multimodal models. It provides a distributed training architecture that separates policy learning and environment rollout processes, enabling efficient and asynchronous reinforcement learning at scale. The framework supports multiple parallelism strategies, including tensor, pipeline, and data parallelism, allowing it to leverage large GPU clusters effectively. It is built with compatibility in mind, supporting popular model families such as LLaMA, Qwen, and diffusion-based world models, as well as integration with Hugging Face ecosystems. cosmos-rl also includes support for advanced RL algorithms, low-precision training, and fault-tolerant execution, making it suitable for large-scale production workloads.

Features

  • Distributed reinforcement learning with asynchronous architecture
  • Support for multiple parallelism strategies including tensor and pipeline
  • Compatibility with LLMs, vision-language models, and diffusion models
  • Low-precision training support such as FP8 and FP4
  • Fault-tolerant and elastic distributed execution
  • Integration with PyTorch and Hugging Face ecosystems

Project Samples

Project Activity

See All Activity >

License

Apache License V2.0

Follow Cosmos-RL

Cosmos-RL Web Site

Other Useful Business Software
Collect! is a highly configurable debt collection software Icon
Collect! is a highly configurable debt collection software

Everything that matters to debt collection, all in one solution.

The flexible & scalable debt collection software built to automate your workflow. From startup to enterprise, we have the solution for you.
Learn More
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Cosmos-RL!

Additional Project Details

Programming Language

Python

Related Categories

Python Reinforcement Learning Frameworks

Registered

2026-03-18