Work done at Zee

3

music composers involved

8

AI guidelines

Designing the Input Experience for AI Music Creation

+ Guidelines for Designing Generative AI Platforms

At Zee Entertainment, background music is a key part of the daily TV show production. It is resource and time-intensive process. In 2024, as generative AI began transforming creative tools, the company identified an opportunity to accelerate this workflow.

I was part of the team from the start, focusing on how creators would interact with such a system. As the product designer, I led foundational research to understand music composer workflows and translated these insights into interface design. The outcome was a detailed concept for AI-assisted music creation, along with a set of guidelines for designing generative AI platforms.

ROLE

Foundational Research, Concept Design, Concept Testing

TIMELINE

3 months

TEAM

3 data scientists • 1 sound designer

Introduction

Zee Entertainment is an Indian media house that produces daily television shows.

  1. Each episode runs for ~20–25 minutes

  2. Episodes air daily

  3. Background music is present across all scenes

  4. Music is created on a daily basis by music composers

We set out to design an intuitive interface that enables professional composers to create music using generative AI.

A woman watching Kundali Bhagya (a popular soap opera)

Understanding Music Workflows & Existing Tools

To understand how AI could assist music creation, I studied existing tools like Suno.com, Beatoven.ai for:

  1. What inputs they accept (prompt vs parameters)

  2. How outputs map to user intent

  3. Whether users feel in control of the outcome

I also visited 3 music composers in Mumbai who create music for Zee’s daily shows.

  1. I observed and spoke to them about their workflows

  2. Gathered feedback on early prototypes(the one from the hackathon) to understand gaps and opportunities.

Insights from music composers' interviews revealed gaps in how musical intent is expressed, interpreted, and controlled in existing AI systems.

Two primary workflows emerged - creating music from scratch and adapting existing tracks - each requiring different levels of control and iteration.

HMW design an input that enables composers to express musical intent and get predictable results?

This concept approaches the input from the perspective of a user who isn’t an expert in music creation, but understands the need for their project and wants to produce music quickly. This concept is assistive in nature and reduces the chances of poor output due to conflicting inputs.

To make the platform relevant for an expert musician, I introduced more parameter-based controls. This assumes that the musician knows what they want and prefers accessing specific options quickly.

With improvements in the dataset and model capability, the system was able to generate higher-quality music from text prompts. This enabled users to freely describe their intent, introducing a more flexible input approach.

To evaluate these approaches, I conducted user testing with composers (with guidance from a senior researcher), focusing on:

  1. How efficiently each concept captures the music brief

  2. Ease of use of the interface

  3. Ability to support experimentation and diverse outputs

These were the takeaways:

  1. Uploading a reference was highly valued, though technical limitations (e.g., vocals) affected output quality

  2. Parameter-only inputs restricted expression, limiting creative flexibility

  3. Text prompts enabled higher experimentation while still capturing intent effectively

  4. Output quality was reliable for shorter durations (~60 seconds), but degraded beyond that

Final Concept

Based on concept testing, I focused on a prompt-based input supported by assistive controls as the final direction. The goal was to balance flexibility of expression with guidance and control, enabling composers to efficiently generate and iterate on music.

Multi-modal input enables users to describe music via text, reference upload, or humming.

Assistive text input uses contextual chips to guide prompt creation without restricting flexibility.

While the broader workspace and editing workflows were explored, this case study focuses on the input experience as the core interaction layer.

Guidelines for Designing Generative AI Platforms

Based on this exploration, I derived a set of guidelines to design effective generative AI experiences. These were shared internally and informed other AI-led initiatives.

Did this in late 2023 - early 2024 :)