Product Thumbnail

Lyria 3 by Google Deepmind

Turn any photo or thought into a custom song inside Gemini

Funny
Music
Artificial Intelligence

The Gemini app now features the most advanced music generation model Lyria 3, empowering anyone to make 30-second tracks using text or images in beta. Turn photos, vibes, or ideas into 30-second songs with vocals and cover art straight from the Gemini app/website.

Top comment

I’m excited to hunt something that genuinely expands what “creative tools” mean in 2026.

Meet Gemini with Lyria 3: Google DeepMind’s most advanced generative music model, now built directly into the Gemini app.

This isn’t just another “type a prompt, get a beat” tool. It’s a full-stack AI music collaborator.

How Lyria 3 works?

Lyria 3 is a high-fidelity music generation model that lets anyone create 30-second custom tracks using:

  • Text prompts

  • Photos

  • Videos

Describe a vibe like: “A comical R&B slow jam about a sock finding their match” …and Gemini generates a polished, structured, catchy track, complete with lyrics and vocals.

Upload a hiking photo of your dog? It composes a song inspired by it.

This is music as expression, not production complexity.

Core features and capabilities

1️⃣ Text → Full Track

  • Genre, mood, tempo control

  • Lyrics auto-generated (or bring your own)

  • Realistic, nuanced vocals

  • Instrumental-only options

2️⃣ Image/Video → Soundtrack

  • Turn visuals into high-fidelity music

  • Mood-aware lyric writing

  • Perfect for social content

3️⃣ Creative Control

  • Direct vocal style

  • Specify acoustic preferences

  • Adjust complexity and arrangement

4️⃣ Professional-Grade Output

  • Clean, structured tracks

  • Export-ready audio

  • Custom AI-generated cover art

5️⃣ Built-In Safety & Watermarking

  • All tracks embedded with SynthID

  • AI content verification inside Gemini

Why Lyria 3 is different?

There are plenty of AI music tools out there. Here’s where this feels distinct:

  1. Musicality, not just loop generation

  2. No lyrics required

  3. Multi-language support

  4. Visual-to-Music pipeline

  5. Responsible AI Layer

Use cases I am excited about:

  • Personalized birthday songs in seconds

  • Instant background music for Shorts/Reels

  • Community creators generating meme music

  • Emotional gifts (parents, partners, inside jokes)

  • Indie creators prototyping soundtrack ideas

  • Marketers testing vibe-based audio branding

This lowers the barrier between idea → emotional artifact.

My personal take: What’s powerful here isn’t that AI can “make music.” It’s that this removes friction between feeling something and expressing it musically.

Over to you!

Would you use this for content? For fun? Or as a serious creative tool? Let me know in the comments. :)

Comment highlights

This looks great! Would be nice to be able songs across smaller languages as well.

Lyria 3 feels like the first AI music tool that actually removes the “production tax” — I can go from a vibe (or even a photo/video) to a structured, listenable track without juggling DAWs, sample packs, and endless tweaking. Quick question: how much granular control do creators get for iteration (e.g., keep the same melody/voice but change tempo/lyrics/arrangement), and what are the licensing/commercial-use terms for tracks generated in Gemini?

Google should consider creating a sheet music generator with complex harmonies and counterpoint.

music is a very interesting field. i was experimenting with elevenlabs and i am exited to see how googles compares to this.

Already thinking about how we can use this for CoreSight. Amazing achievement, congrats!

Has anyone gotten some good results with this they'd like to share? The generations feels pretty good, but still a bit slop-y.