Music in Interactive Media

The current explosive growth in interactive media means providers and users can easily create, publish and share emergent content in vlogs, video games, social media, VR and AR. However, the use of music in these media is out of sync with this trend. Music is a difficult thing to do and to come by. There are not enough composers and time to go round for the hundreds of millions of hours of stuff put out there each month. Paying composers or using online music libraries is not an option for many. To get copyright music means using various time-consuming and costly channels. And music libraries don’t provide custom music. Providers and users need an infinite well of music to get over this creation deficit - but where can they get it? Artificial intelligence is the only answer.

Interactive media uses static tracts that are not easily customised by users. Also, musical styles and emotions cannot adapt to the near-infinite possible contexts in their media. The result is that music that does not fit the media content lowers the level of engagement in an experience. A collection of static songs can easily lead to listeners becoming bored and can mean they stream music from external apps, as we’ve discovered in our research. Also from our research, content creators and users want to have music that fits experiences stylistically and emotionally. At present this is not a reality.

Interactive platforms don’t enable creators and users to have convenient, instant music that is customised for their content. Since music is not a resource that is freely available and is not usually produced in realtime, this puts strain on creators and users to find apt music solutions. Content providers and users need instant and convenient music solutions to more easily dress up the content they build or explore.

In contrast to the linear content of songs, like those used in videos and movies, music in interactive media should respond dynamically to internal and external events - such as the changing emotions of the player and the developing storyline. This type of music is adaptive music.

Advancements in AI and music theory have made it possible for Melodrive to build new technology that optimises musical creation for users. Deep adaptive music is Melodrive’s own innovation, and does the following:

  • Generates music content in realtime by cutting-edge AI
  • Adapts emotionally on a granular level
  • Enables infinite possible musical paths, depending on the unique interactions of the user
  • Provides music content in a range of styles
  • Has tailormade musical presets and instrumentation for the user

With deep adaptive music, users with all skill levels can create interactive music on a granular level. This solves the music creation deficit for interactive platforms noted above, allowing millions of users to write and experience immersive soundtracks at the click of a button.

Want to know more about


Want to Know How to
Increase User Engagement With Music?

Check out our white paper "How Music Can Boost User Engagement by 40%". Learn how to leverage music to provide a dramatic increase in user immersion and session time on digital platforms.

Contact us

Are you curious to know about how Melodrive can benefit your business?

Drop us an email at