Product Thumbnail

Lyria 3 by Google Deepmind

Turn any photo or thought into a custom song inside Gemini

Funny
Music
Artificial Intelligence

The Gemini app now features the most advanced music generation model Lyria 3, empowering anyone to make 30-second tracks using text or images in beta. Turn photos, vibes, or ideas into 30-second songs with vocals and cover art straight from the Gemini app/website.

Top comment

I’m excited to hunt something that genuinely expands what “creative tools” mean in 2026.

Meet Gemini with Lyria 3: Google DeepMind’s most advanced generative music model, now built directly into the Gemini app.

This isn’t just another “type a prompt, get a beat” tool. It’s a full-stack AI music collaborator.

How Lyria 3 works?

Lyria 3 is a high-fidelity music generation model that lets anyone create 30-second custom tracks using:

  • Text prompts

  • Photos

  • Videos

Describe a vibe like: “A comical R&B slow jam about a sock finding their match” …and Gemini generates a polished, structured, catchy track, complete with lyrics and vocals.

Upload a hiking photo of your dog? It composes a song inspired by it.

This is music as expression, not production complexity.

Core features and capabilities

1️⃣ Text → Full Track

  • Genre, mood, tempo control

  • Lyrics auto-generated (or bring your own)

  • Realistic, nuanced vocals

  • Instrumental-only options

2️⃣ Image/Video → Soundtrack

  • Turn visuals into high-fidelity music

  • Mood-aware lyric writing

  • Perfect for social content

3️⃣ Creative Control

  • Direct vocal style

  • Specify acoustic preferences

  • Adjust complexity and arrangement

4️⃣ Professional-Grade Output

  • Clean, structured tracks

  • Export-ready audio

  • Custom AI-generated cover art

5️⃣ Built-In Safety & Watermarking

  • All tracks embedded with SynthID

  • AI content verification inside Gemini

Why Lyria 3 is different?

There are plenty of AI music tools out there. Here’s where this feels distinct:

  1. Musicality, not just loop generation

  2. No lyrics required

  3. Multi-language support

  4. Visual-to-Music pipeline

  5. Responsible AI Layer

Use cases I am excited about:

  • Personalized birthday songs in seconds

  • Instant background music for Shorts/Reels

  • Community creators generating meme music

  • Emotional gifts (parents, partners, inside jokes)

  • Indie creators prototyping soundtrack ideas

  • Marketers testing vibe-based audio branding

This lowers the barrier between idea → emotional artifact.

My personal take: What’s powerful here isn’t that AI can “make music.” It’s that this removes friction between feeling something and expressing it musically.

Over to you!

Would you use this for content? For fun? Or as a serious creative tool? Let me know in the comments. :)

Comment highlights

music is a very interesting field. i was experimenting with elevenlabs and i am exited to see how googles compares to this.

Not bad at all. Gave it a spin and I love that it’s missing that subtle 'AI hum' you usually get with competitors. Perfect for creating chill, low-stakes background music like smooth jazz.

Already thinking about how we can use this for CoreSight. Amazing achievement, congrats!

Has anyone gotten some good results with this they'd like to share? The generations feels pretty good, but still a bit slop-y.