Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audio codec. Mimi processes 24 kHz audio, down to a 12.5 Hz representation with a bandwidth of 1.1 kbps, in a fully streaming manner (latency of 80ms, the frame size), yet performs better than existing, non-streaming, codecs like SpeechTokenizer (50 Hz, 4kbps), or SemantiCodec (50 Hz, 1.3kbps). Moshi models two streams of audio: one corresponds to Moshi, and the other one to the user. At inference, the stream from the user is taken from the audio input, and the one for Moshi is sampled from the model's output. Along these two audio streams, Moshi predicts text tokens corresponding to its own speech, its inner monologue, which greatly improves the quality of its generation. A small Depth Transformer models inter codebook dependencies for a given time step, while a large, 7B parameter Temporal Transformer models the temporal dependencies.

Features

  • Converts JSON to Java and Kotlin objects
  • Annotation-based mapping for flexibility
  • Supports custom adapters for complex models
  • Lightweight with minimal dependencies
  • JSON serialization and deserialization

Project Samples

Project Activity

See All Activity >

Categories

Speech

License

Apache License V2.0

Follow Moshi

Moshi Web Site

Other Useful Business Software
Collect! is a highly configurable debt collection software Icon
Collect! is a highly configurable debt collection software

Everything that matters to debt collection, all in one solution.

The flexible & scalable debt collection software built to automate your workflow. From startup to enterprise, we have the solution for you.
Learn More
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Moshi!

Additional Project Details

Operating Systems

Linux, Mac, Windows

Programming Language

Python

Related Categories

Python Speech Software

Registered

2024-11-04