In addition to providing technological solutions for audio description, VIDEO TO VOICE is actively involved in other fields surrounding accessibility for digital media. For some time, VIDEO TO VOICE has been working in close partnership with the Fraunhofer Institute for Integrated Circuits (IIS) on topics such as automated mixing and MPEG-H Audio. The following article provides detailed analysis of the tools that make up MPEG-H Audio, a technology developed by Fraunhofer IIS to improve accessibility and deliver the best sound experience possible.

This is the second part of our series exploring the object-based approach to audio production and Fraunhofer IIS's involvement in developing MPEG-H Audio system. Part one focused on the fundamentals in traditional channel-based audio, the accessibility problems this approach causes, and the solutions that object-based audio can provide. 

With the basics covered, it's now time to take a closer look at the ground-breaking technology behind MPEG-H Audio and the various tools content creators can take advantage of. Let's start with the MPEG-H Authoring Suite...

What does the MPEG-H Authoring Suite involve?

Authoring in this sense is the step where metadata is created to enable playback devices to deliver on the three key principles of interactivity, immersive sound, and universal delivery.

Metadata is information that describes an audio object's existence, position, and function.

The MPEG-H Authoring Suite (MAS) is the place where content creators can create metadata, such as setting up presets, enabling interactivity options for viewers, and defining positions and properties of audio objects.

On the MAS, content creators can also monitor how their mix will be rendered in different layouts and make any necessary changes.