KaniTTS2 Revolutionizes Real-Time TTS with Frame-Level Encoding and Open-Source Training
KaniTTS2, a groundbreaking text-to-speech model with frame-level position encodings, has been released with open-source training code, enabling real-time conversational AI even on low-memory GPUs. With support for three languages and voice cloning capabilities, it sets a new standard for accessible, high-performance speech synthesis.

KaniTTS2 Revolutionizes Real-Time TTS with Frame-Level Encoding and Open-Source Training
A new frontier in text-to-speech (TTS) technology has emerged with the release of KaniTTS2, a 400-million-parameter model engineered for real-time conversational AI applications. Developed by the team at NineNineSix AI, the model leverages frame-level position encodings — a novel architectural refinement that enhances temporal precision in speech generation — and runs efficiently on consumer-grade hardware. Unlike traditional TTS systems requiring high-end GPUs and extensive computational resources, KaniTTS2 achieves a remarkable 0.2 real-time factor (RTF) on an RTX 5080 with just 3GB of VRAM, making it one of the fastest and most accessible TTS models to date.
According to the release posted on Reddit’s r/LocalLLaMA, KaniTTS2 is built on LiquidAI’s LFM2 backbone and integrates Nvidia’s NanoCodec for optimized audio encoding. The model has been pretrained on approximately 10,000 hours of multilingual speech data, trained across eight NVIDIA H100 GPUs in just six hours — a testament to the efficiency of its training pipeline. The system supports English, Spanish, and Kyrgyz out of the box, with plans to expand language coverage in future updates. An English-specific variant is also available for users seeking higher fidelity in that language.
Perhaps the most transformative aspect of KaniTTS2 is its complete open-source release. The full pretraining codebase, hosted on GitHub, empowers researchers, developers, and linguistic communities to train custom TTS models from scratch — even for low-resource languages or regional accents. The framework includes advanced features such as Fully Sharded Data Parallel (FSDP) for multi-GPU training, Flash Attention 2 for memory-efficient computation, and YAML-configurable training pipelines that simplify hyperparameter tuning. Additionally, built-in attention analysis metrics allow users to validate layer isolation and model convergence, ensuring transparency and reproducibility in model development.
Voice cloning is another standout feature. By incorporating speaker embeddings, KaniTTS2 can replicate the vocal characteristics of a target speaker using as little as a few seconds of reference audio. This capability has profound implications for personalized virtual assistants, audiobook narration, and accessibility tools for individuals with speech impairments. The model’s ability to maintain natural prosody and intonation — even under low-latency constraints — positions it as a leading candidate for deployment in live customer service bots, interactive gaming NPCs, and real-time translation systems.
The release is licensed under Apache 2.0, ensuring commercial and academic freedom. Developers can experiment with live demos via Hugging Face Spaces and engage directly with the development team on Discord. This community-driven approach reflects a broader industry shift toward democratizing AI: rather than locking cutting-edge models behind proprietary APIs, NineNineSix AI is providing the tools for anyone to adapt the technology to their unique linguistic or cultural context.
Industry analysts note that KaniTTS2’s combination of speed, accuracy, and openness could disrupt the TTS market, currently dominated by closed systems from Google, Amazon, and Microsoft. By lowering the barrier to entry — both in hardware requirements and technical expertise — KaniTTS2 may accelerate innovation in underrepresented languages and regional dialects, fostering more inclusive voice technologies globally. As the team prepares to add more languages and optimize for edge devices, the open-source community is already beginning to contribute fine-tuned checkpoints and domain-specific datasets, signaling a new era of collaborative AI development.


