Published 18 February 2026

Lyria 3

Model Cards are intended to provide developers with essential, summarized information on models, including overviews of known limitations and mitigation approaches. Model cards may be updated from time to time; for example, to include updated evaluations as the model is improved or revised. The model card below covers Lyria 3 and subsequent versions.

Published: February 2026

Model information

Description

Lyria 3 is a music generation system capable of synthesizing high-quality audio from a text prompt.

Inputs

Text

Outputs

Audio (music), text (lyrics)

Architecture

Lyria 3 utilizes latent diffusion, applied to temporal audio latents.


Model data

Training dataset

Lyria 3 was trained on audio data. Audio datasets were annotated with text captions at different levels of detail.

Training data processing

Data filtering and preprocessing included techniques such as deduplication, safety filtering in-line with Google's commitment to advancing AI safely and responsibly, and quality filtering to mitigate risks and improve training data reliability and compliance. Once data is collected, it is cleaned and preprocessed to make it suitable for training.


Implementation and sustainability

Hardware

Lyria 3 was trained using Google’s Tensor Processing Units (TPUs). TPUs are specifically designed to handle large-scale computation involved in training and serving generative AI models, increasing training throughput and efficiency considerably when compared to CPUs. TPUs come with large amounts of high-bandwidth memory, allowing for the handling of large models and batch sizes during training, which can lead to better model quality. TPU Pods (large clusters of TPUs) also provide a scalable solution for handling the growing complexity of large foundation models. Training can be distributed across multiple TPU devices for faster and more efficient processing.

The efficiencies gained through the use of TPUs are aligned with Google's commitment to operate sustainably.

Software

Training was done using JAX and ML Pathways.


Distribution

Lyria 3 is distributed in the following channels; respective documentation shared in line:

Our models are available to downstream providers via an application program interface (API) and subject to relevant terms of use. There is no required hardware or software to use the model.


Evaluation

Approach

Lyria 3 was evaluated across a range of benchmarks, including human and automated evaluations. In close collaboration with music experts, we have developed an evaluation framework that tests the model's ability to represent concepts (such as genres, moods, instruments) in- and out-of-distribution across a broad range of curated user prompts. We sampled several music generation models for comparison, including Lyria 2.

We focus on a range of quality dimensions including music quality and aesthetics, vocal quality, audio fidelity, and prompt adherence.

Results

Lyria 3 improved significantly compared to Lyria 2 on audio fidelity, highlighting the technical quality and clarity of the generated audio. With lyrics, Lyria 3 demonstrates better prompt adherence, following both simple and more complex instructions more accurately.


Intended usage and limitations

Benefit and Intended Usage

Lyria 3 is Google’s most capable music generation model to date. Lyria can be used to generate high-quality music in a wide range of genres and styles.

Acceptable Usage

Google’s Generative AI Prohibited Use Policy applies to uses of the model in accordance with the applicable terms of service. Additionally, the model should not be integrated into certain systems (also found in Google’s Generative AI Prohibited Use Policy), including those that: (1) engage in dangerous or illicit activities, or otherwise violate applicable laws or regulations, (2) violate the rights of others, including privacy and intellectual property rights, (3) compromise the security of others’ or Google’s services, (4) engage in sexually explicit, violent, hateful, or harmful activities, (5) engage in misinformation, misrepresentation, or misleading activities.


Ethics and content safety

Evaluation Approach

Lyria 3 was developed in partnership with internal safety and responsibility teams. A range of evaluations and red-teaming activities were conducted to help improve the model and inform decision-making. These evaluations and activities align with Google's AI Principles and responsible AI approach. Evaluation types included but were not limited to:

  • Training/Development Evaluations including automated and human evaluations carried out continuously throughout and after the model’s training, to monitor progress and performance
  • Human Red Teaming conducted by specialist teams who sit outside of the model development team deliberately trying to spot weaknesses and ensure the model adheres to safety policies and desired outcomes
  • Ethics & Safety Reviews were conducted ahead of the model’s release

In addition to the evaluations above, system-level safety evaluations and reviews are run within the context of specific applications that models are deployed within.

Risks and Mitigations

Safety and responsibility was built into Lyria 3 throughout the training and deployment lifecycle, including pre-training, post-training, and product-level mitigations.

Mitigations include, but are not limited to:

  • dataset filtering;
  • conditional pre-training;
  • supervised fine-tuning;
  • reinforcement learning from human and critic feedback;
  • safety policies and desiderata;
  • product-level mitigations such as safety filtering and applying SynthID watermarking.