Hume AI releases TADA-1B, a 1 billion parameter text-to-speech model
Hume AI has released TADA-1B, a 1 billion parameter text-to-speech model available on Hugging Face under an MIT license. The model, which combines speech and language capabilities, has already accumulated over 3,100 downloads since its January 12 release.
Hume AI has released TADA-1B, a 1 billion parameter open-source text-to-speech model designed to bridge speech synthesis and language understanding in a single architecture.
Model Specifications
TADA-1B is available on Hugging Face under the permissive MIT license, making it freely usable for both research and commercial applications. The model is built on a Llama-based architecture and includes optimized safetensors formatting for efficient inference.
The model supports English language synthesis and was released on January 12, 2026. According to the Hugging Face model card, the work is associated with arxiv:2602.23068, suggesting a corresponding research paper detailing the architecture and training methodology.
Adoption and Accessibility
Since its release, TADA-1B has generated significant initial interest, accumulating 3,158 downloads and 69 likes on Hugging Face—metrics indicating early adoption within the open-source AI community. The 1 billion parameter size positions it as a lightweight alternative to larger text-to-speech systems, potentially enabling deployment on resource-constrained hardware.
The safetensors format used for model distribution ensures compatibility with modern inference frameworks and reduces security risks associated with pickle-based model loading.
What This Means
TADA-1B represents an incremental advance in open-source speech synthesis, particularly in combining LLM-style architectures with TTS capabilities. The MIT licensing and modest 1B parameter count make it genuinely accessible to researchers and developers seeking to build speech applications without proprietary dependencies. However, early download metrics suggest adoption remains limited compared to established TTS baselines. The associated arxiv paper (2602.23068) will be critical for evaluating claims about audio quality, latency, and comparative performance against existing methods.
For teams needing lightweight, permissively-licensed text-to-speech, TADA-1B offers a viable open alternative—but actual quality benchmarks against Bark, Edge TTS, or commercial APIs remain unstated.
Related Articles
Rakuten releases RakutenAI-3.0, 671B-parameter Japanese-optimized mixture-of-experts model
Rakuten Group has released RakutenAI-3.0, a 671 billion parameter mixture-of-experts (MoE) model designed specifically for Japanese language tasks. The model activates 37 billion parameters per token and supports a 128K context window. It is available under the Apache License 2.0 on Hugging Face.
Nvidia releases Nemotron 3 Super: 120B MoE model with 1M token context
Nvidia has released Nemotron 3 Super, a 120-billion parameter hybrid Mamba-Transformer Mixture-of-Experts model that activates only 12 billion parameters during inference. The open-weight model features a 1-million token context window, multi-token prediction capabilities, and pricing at $0.10 per million input tokens and $0.50 per million output tokens.
NVIDIA releases Nemotron 3 Content Safety 4B for multimodal, multilingual moderation
NVIDIA released Nemotron 3 Content Safety 4B, an open-source multimodal safety model designed to moderate content across text, images, and multiple languages. Built on Gemma-3 4B-IT with a 128K context window, the model achieved 84% average accuracy on multimodal safety benchmarks and supports over 140 languages through culturally-aware training data.
Xiaomi launches MiMo-V2-Pro with 1T parameters, matches Claude Opus on coding at 80% lower cost
Xiaomi shipped three AI models simultaneously designed to form a complete agent platform. MiMo-V2-Pro, a 1-trillion-parameter Mixture-of-Experts model with 42 billion active parameters per request, scores 78% on SWE-bench Verified and 81 points on ClawEval—nearly matching Claude Opus 4.6 while costing $1 per million input tokens versus $5 for Opus.