Google's Gemini adds Lyria 3 music generation from text and images
Google has integrated Lyria 3, its music generation model, directly into the Gemini app. Users can now create custom 30-second music tracks from text descriptions and images without additional tools or subscriptions.
Google Integrates Lyria 3 Music Generation Into Gemini App
Google has rolled out Lyria 3 music generation capabilities within the Gemini app, allowing users to create 30-second audio tracks from text prompts and image inputs. The feature is now available to Gemini app users.
What Lyria 3 Does
Lyria 3 generates original music compositions based on user-provided text descriptions and images. Users input creative prompts—specifying genre, mood, instrumentation, or other characteristics—and the model produces corresponding audio files. The system can also generate music inspired by visual inputs, adding another dimension to music creation workflows.
Integration Details
The feature is embedded directly in the Gemini app interface, eliminating the need for users to navigate to separate music generation tools. This positions music creation as a native capability alongside Gemini's text, image, and document analysis features.
Google has not disclosed specific technical specifications for Lyria 3 in this announcement, including model size, training data composition, or detailed performance benchmarks against competing music generation systems.
Broader Context
This release follows Google's earlier introduction of Lyria and Music AI Essentials framework. Google DeepMind has been developing music generation technology as part of its broader multimodal AI strategy. The integration into Gemini represents a shift toward making generative music capabilities as accessible as text and image generation.
The 30-second generation limit suggests design decisions around computational efficiency and user experience, though Google has not explained the rationale for this constraint.
Industry Position
Music generation remains an emerging category in generative AI. Competitors including OpenAI-backed projects, Meta's research divisions, and specialized startups have developed similar tools. Google's approach of integrating music generation into its flagship AI assistant positions music creation as a mainstream feature rather than a specialized tool.
The lack of disclosed pricing or usage limits suggests the feature may be part of Gemini's existing subscription tier structure, though specific details remain unavailable.
What This Means
Google is treating music generation as a standard multimodal AI capability, not a separate product. By embedding Lyria 3 into Gemini, the company signals confidence in both the technology's stability and its appeal to mainstream users. For music creators, marketers, and content producers, this lowers barriers to generative music experimentation. For competitors, it intensifies pressure to include music generation in their own platforms. The feature's limitations—30-second maximum length, unspecified quality standards—remain important constraints relative to specialized music generation tools.