Sky Sports 4K control room. Higher-quality broadcast channels need to be supported by new immersive audio options.
The world of broadcast audio is about to reach new levels as the industry embraces the future with Next Generation Audio (NGA). While precisely what features will be offered remain unknown, several 3D Immersive formats are already under development and soon will find their way into broadcast production and distribution.
Unlike the world of constrained channel based coding, these new NGA codecs will support more channels and/or object based audio coding. For the consumer, there will be two major benefits - a greater sense of involvement or immersion and a degree of personalisation.
Immersive 3D audio is undoubtedly one aspect of Next Generation Audio. In contrast to that approach, personalized audio can be incorporated into traditional and standard channel based formats, in the same way as stereo or mono. The key to enabling such features is Object-Based Audio (OBA).
Inside Object-based audio
Object Based Audio (OBA) will give users the option of personalising their experience by selecting from a number of audio sources and controlling the level and maybe even the position in the mix. With OBA, an “object” is essentially an audio stream with accompanying descriptive metadata. The metadata carries information about where and how to render the object in a mix that is being reproduced.
That might sound complicated, but in fact it is very straightforward. Two versions of a commentary with one mono FX stream is already an OBA format, and having the ability to choose between one commentary track and the other is already personalised audio. This system works fine as long as both commentary tracks reach the recipients home as separated audio channels and are not mixed into the audio bed.
In common parlance, Auto-Mix means balancing dynamic input levels so they have equal power output at the summation point. This can also be described as conference auto-mixing, where un-used microphone channels receive less gain and therefore noise and crosstalk is automatically reduced.
Figure 1. Jünger’s audio technology, enables separate feeds to be automatically mixed into a program feed as shown by this diagram.
Another Auto-Mix method is A/B crossfade, in which a crossfade from source A to source B is automatically performed in response to a pre-defined trigger. If a sequence of audio elements is being used to create the audio programme, then a typical procedure would involve sequentially switching the sources – for example, presentation, clip, promo, presentation, clip and so on.
A real mix is the result of Auto-Voice-Over mixing, in which one audio element is laid over the audio bed. This kind of Auto-Mix can be triggered by the producer or by an automation system that takes a level controlled ‘voice’ input and lays it over the audio bed in a process known as ‘ducking’.
The question one needs to ask is which of these Auto-Mix methods is most relevant.
Choosing your tools
One of the major challenges for the production industry will be to create OBA production strategies. This means completely rethinking how a final mix is created, because with OBA, it will be performed at home by the viewer rather than by a mixer in a post facility.
Keep in mind that as soon as a post house mixes objects (different language commentaries, for example) into the audio bed, they are gone and are no longer available for personalisation by the viewer. To give viewers the chance for personalised audio, the production workflow must change and deliver separate ‘unmixed’ channels so that the home receiver and decoder can finish the final mix. This is very different to how we currently mix for surround or for standard stereo audio.
Take the next step
A first step is education. Help the audio production staff understand this new way of working whereby they no longer create a final mix. Metadata is key to successful implementation.
Review current workflows. As content is created or added to a mix, be sure the accompanying metadata survives. That metadata is the key to any object based audio tracks surviving postproduction for delivery to the broadcast transmission facility.
As consumers seek out new customisable and immersive audio environments, broadcasters who can supply them will benefit. Content and program production facilities can plan now for the necessary tools to enable these new features. A related benefit of implementing the changes will be a faster and more cost-effective production workflow.
Peter Poers, Managing Director, Jünger Audio.
You might also like...
Part one of this four-part series introduces immersive audio, the terminology used, the standards adopted, and the key principles that make it work.
Every Super Bowl is a showcase of the latest broadcast technology, whether video or audio. For the 53rd Super Bowl broadcast, CBS Sports will use almost exclusively IP and network-based audio.
This year’s Super Bowl LIII telecast on CBS will be produced and broadcast into millions of living rooms by employing the usual plethora of traditional live production equipment, along with a few wiz bang additions like 4K UHD and a…
Networked modular audio stageboxes have been around for a while and were hailed as a convenient alternative to clunky snakes and the huge patch bays that came with them. Unlike analog stage- and wallboxes, which usually only transmit signals to…
Quality Control is one of the many areas where IT and broadcast use similar terms, but the meaning is quite different. Whereas IT focuses on guaranteeing bit rates and packet delivery to improve quality of service and hence quality of…