How Latency Affects Computer Audio Editing Systems

Latency is the millisecond delay between the time an audio signal enters an electronic system and when it emerges. Though a simple concept to understand, latency can play havoc in a range of audio applications, including well-known issues in live broadcasting.

Latency can be caused by many factors, including both analog-to-digital and digital-to-analog conversion, buffering, digital signal processing, transmission time and the audio speed in the transmission medium.

This delay can be a critical performance consideration in several pro audio applications. Included are recording interfaces with personal computers, sound reinforcement, in-ear monitoring systems and live broadcasting. For example, that annoying delay between a questioner in the studio and a reporter in the field on a live TV news remote is one kind of latency.

In this article, however, we will focus on latency issues in basic audio editing systems using a personal computer. Latency often comes between an audio interface and the computer itself. The goal is to lessen the delay to times that are too short for the human ear to detect. Since around 6ms is audible as a delay, latency needs to be below this threshold.

We can never get rid of all latency. Despite very fast computer processing, users can still get a few milliseconds latency as the audio travels from the interface, through the computer and then sent back to the interface to be played out of the computer. The sound must travel through multiple processors and wires to get to the listener's ears.

The good news is with the latest editing systems, most latency is unnoticeable by users. If it remains consistent, it can be fixed in post-production. This has been aided by ever faster computers and speedy connectivity.

Thunderbolt Connection Cable

Thunderbolt Connection Cable

There are a grab bag of tricks to tackle latency. Super-fast Thunderbolt connectivity between devices has helped. By adjusting buffer sizes, latency can also be reduced. Direct monitoring can help solve the problem. This is when the input is sent straight to the headphones, greatly reducing the latency.

Other causes of latency can be found within the processing of the audio — either in the converters or plugins. A step that can be taken to reduce latency is to decrease the number of active processors and plugins being used simultaneously in the system. This can be done either by waiting until mix down to apply them or by rendering the audio tracks with effects turned on so the processing power is reduced.

As to buffer sizes, computers at times process large amounts of simultaneous data. This can cause a problem for audio, which requires a constant stream of samples due to sound being a continuous waveform.

Buffer size is dependent on factors such as how many plugins are loaded on a track, and the computer’s processing power. If the buffer size is too low, users may encounter errors during playback or may hear clicks and pops. If the buffer size is set too high, a lot of latency can be heard, making the process frustrating.

UA Apollo X4 Audio Interface

UA Apollo X4 Audio Interface

When introducing additional audio to an editing session, a larger buffer size might be needed to accurately record the signal. This increase in buffer size will allow more time for the audio to be captured without distortion.

It is important for users to find the appropriate buffer size for the session as this can vary depending on the number of tracks, plugins or audio files. Sample rate can also be used to increase or decrease latency. This is the number of audio samples captured per second. Most audio pros set the sample rates at 48 kHz, while the bit depth is set at 24.

Experimenting with both buffer size and sample rate is a good way to effectively deal with latency. When both recording and editing audio, these settings are also important to the sonic quality of your audio. The balance between optimal performance without latency and audio quality is important.

The good news is audio interfaces have improved dramatically in recent years. Most now come with plugins and drivers that are optimized to lower buffer sizes with higher stability in larger sessions. A good, brand name audio interface can dramatically reduce latency problems. 

You might also like...

Production Control Room Tools At NAB 2024

As we approach the 2024 NAB Show we discuss the increasing demands placed on production control rooms and their crew, and the technologies coming to market in this key area of live broadcast production.

Designing IP Broadcast Systems: Where Broadcast Meets IT

Broadcast and IT engineers have historically approached their professions from two different places, but as technology is more reliable, they are moving closer.

Audio At NAB 2024

The 2024 NAB Show will see the big names in audio production embrace and help to drive forward the next generation of software centric distributed production workflows and join the ‘cloud’ revolution. Exciting times for broadcast audio.

SD/HD/UHD & SDR/HDR Video Workflows At NAB 2024

Here is our run down of some of the technology at the 2024 NAB Show that eases the burden of achieving effective workflows that simultaneously support multiple production and delivery video formats.

Standards: Part 7 - ST 2110 - A Review Of The Current Standard

Of all of the broadcast standards it is perhaps SMPTE ST 2110 which has had the greatest impact on production & distribution infrastructure in recent years, but much has changed since it’s 2017 release.