Local TTS model

Sesame CSM

Conversational Speech Model - generates speech with natural turn-taking, backchannels and interruptions. Built specifically for multi-turn dialogue with real-time response generation.

Quality
9.5/10
Speed
7.5/10
Size
3.5GB
Languages
1+

Quick answer

Sesame CSM is a local speech model from Sesame AI. It is best suited for dialogue, streaming, realtime, emotion workflows. Check the license before commercial use.

Model details

Hardware

gpuapple

Formats

pytorchsafetensors

Voices

Built-in conversational voices

Latency

low

License

Apache 2.0

Release

2025-03

Install command

pip install sesame-csm

Features

dialoguestreamingrealtimeemotion

Languages: en

Context: Turn-taking, backchannels, interruptions

Related TTS models

NVIDIA

Parakeet TDT 0.6B v2

Quality 9.4 · Speed 10 · 1.1GB · CC-BY-4.0

NVIDIA's SOTA lightweight ASR - 0.6B params, #1 on Open ASR Leaderboard for English. TDT (Token-and-Duration Transducer) decoding makes it 50× faster than Whisper Large v3 on GPU. Real-time streaming with word-level timestamps.

streamingrealtimelow-latency
hexgrad

Kokoro TTS

Quality 9.2 · Speed 9.8 · 0.33GB · Apache 2.0

Ultra-lightweight yet stunning quality. 82M params only - runs on CPU in real-time. Best quality-to-size ratio of any TTS model.

realtimestreaminglow-latencymultilingual
Kyutai

Kyutai STT 2.6B

Quality 9.4 · Speed 9.5 · 2.7GB · CC-BY-4.0

Production-grade streaming ASR from Kyutai (makers of Moshi). Delay-streaming transformer with 500ms latency, word-level timestamps, speaker diarization. Top of Open ASR Leaderboard for real-time French + English.

streamingrealtimelow-latencymultilingual
Speech Research (SWivid)

F5-TTS v1.1

Quality 9.5 · Speed 9.2 · 1.6GB · MIT

Iterative upgrade over the original F5-TTS. Faster convergence via improved flow-matching schedule, better Chinese prosody, cross-lingual cloning. Now with streaming inference and improved CFM sampler.

realtimecloningstreamingmultilingual
OpenAI

Whisper v3 Turbo

Quality 9.1 · Speed 9.5 · 1.6GB · MIT

OpenAI's optimized Whisper v3 with 4 decoder layers instead of 32. 8× faster than Whisper Large v3 with only minor accuracy trade-off. 99 languages supported. New gold standard for fast local transcription.

streamingrealtimemultilinguallow-latency
jamiepine / Community

Voicebox

Quality 9 · Speed 9.5 · 0.05GB · MIT

Desktop app & orchestrator for local TTS - not a model. Provides a UI studio, voice profile management, and a local API. Generates audio via swappable backends (Qwen3 TTS, Kokoro, Piper, XTTS…). Think of it as a front-end shell that runs on top of your installed TTS models.

streamingrealtimelow-latency