Google launches Veo 3.1 Lite, cutting video generation costs by half
Google announced Veo 3.1 Lite, a cost-reduced video generation model priced at less than 50% of Veo 3.1 Fast's cost. The model supports text-to-video and image-to-video generation at 720p or 1080p resolution with customizable durations of 4s, 6s, or 8s, rolling out today on the Gemini API and Google AI Studio.
Google launches Veo 3.1 Lite, cutting video generation costs by half
Google announced Veo 3.1 Lite today, positioning itself as committed to video generation as OpenAI exits the space. The new model is priced at less than 50% of Veo 3.1 Fast's cost, making it Google's most cost-effective video generation offering to date.
Model hierarchy and capabilities
Veo 3.1 Lite sits beneath Veo 3.1 Fast in Google's video generation lineup, with Veo 3.1 remaining at the top tier. The model is designed for "high-volume video applications" and supports:
- Text-to-video generation
- Image-to-video generation
- Resolution options: 720p or 1080p
- Aspect ratios: 16:9 (landscape) and 9:16 (portrait)
- Customizable video duration: 4s, 6s, or 8s (with cost adjusting per duration)
- Generation speed matching Veo 3.1 Fast
Pricing and availability
Exact pricing for Veo 3.1 Lite has not been disclosed, but Google confirmed it costs "less than 50%" of Veo 3.1 Fast's current rate. Veo 3.1 Fast itself is receiving a price reduction on April 7, 2026.
Veo 3.1 Lite is available immediately through:
- Gemini API
- Google AI Studio
Integration and commitment
Veo technology is already integrated across multiple Google products including YouTube Shorts, Google Photos, Google Vids, the Gemini app, and the dedicated Flow tool. Google's statement suggests additional announcements are forthcoming: "Our commitment to making video generation more available to developers doesn't stop with the release of Veo 3.1 Lite. Stay tuned for more updates soon!"
The announcement comes days after OpenAI announced it would discontinue Sora video generation, ceding that market to competitors like Google, Runway, and others.
What this means
Google is doubling down on video generation at the moment when OpenAI is retreating. By tiering Veo pricing with a cost-optimized Lite version, Google targets volume use cases where Veo 3.1 and Veo 3.1 Fast pricing may have been prohibitive. The sub-50% pricing positions Veo 3.1 Lite for developers building high-throughput video applications. The cryptic commitment to future announcements suggests Google plans additional pricing tiers or capability expansions soon.
Related Articles
Google releases Gemini 3.1 Flash Lite with 1M context at $0.25 per million input tokens
Google has released Gemini 3.1 Flash Lite, a high-efficiency multimodal model with a 1,048,576 token context window priced at $0.25 per million input tokens and $1.50 per million output tokens. The model supports text, image, video, audio, and PDF inputs with four thinking levels for cost-performance optimization.
Baidu Releases Qianfan-OCR-Fast Model with 66K Context at $0.68 Per 1M Input Tokens
Baidu has released Qianfan-OCR-Fast, a multimodal model specialized for optical character recognition tasks. The model offers a 66,000 token context window and is priced at $0.68 per 1M input tokens and $2.81 per 1M output tokens.
Tencent Releases Hy3 Preview: Mixture-of-Experts Model with 262K Context and Configurable Reasoning
Tencent has released Hy3 preview, a Mixture-of-Experts model with a 262,144 token context window priced at $0.066 per million input tokens and $0.26 per million output tokens. The model features three configurable reasoning modes—disabled, low, and high—designed for agentic workflows and production environments.
Microsoft Releases Fara-7B: 7B Parameter Computer Use Agent Trained in 2.5 Days on 64 H100s
Microsoft Research has released Fara-7B, a 7-billion parameter small language model designed for computer automation tasks. The model, which took 2.5 days to train on 64 H100 GPUs, can navigate websites to complete tasks like booking restaurants and shopping, using screenshots as input with a 128K token context window.
Comments
Loading...