The Need for a Replacement Timecode Standard

The SMPTE timecode standard has served us well for almost half a century. But the equipment it was designed for is history, and soon it will be too. The industry needs a replacement.

If there is one tool that distinguishes professional video production from amateur it is SMPTE timecode, officially defined in the SMPTE ST 12M suite of specifications as a set of cooperating standards.

The concept of timecode was originally developed in 1967 by EECO (Electronic Engineering Company of Santa Ana, CA) as an "hours:minutes:seconds:frames" numbering system first used on 2” helical scan quad tape. EECO made the intellectual property public, and in the March, 1970 issue of the “Journal of the SMPTE”, it was proposed as an industry standard. Then on April 2nd, 1975, SMPTE timecode was approved by the American National Standards Institute.

But now, half a century later, as Peter Symes of Symes TV Consulting, said in his SMPTE 2018 presentation that the original timecode is in need of revision.

“SMPTE ST 12M is a wonderful 50 year old standard, but it has a number of deficiencies,” Symes began, “and back in 2007 Hans Hoffman of the EBU, and member of SMPTE, and I set up a Task Force to look into it.”

SMPTE timecode has outlived those who first proposed it.<br />(click to expand)

SMPTE timecode has outlived those who first proposed it.
(click to expand)

The Task Force reported that the goals for the existing timecode replacement had to support all known frame rates and still be attractive to other industries.

“At the time there were two competing proposals for a timecode replacement, but they could not be reconciled,” Symes said. “So we set some requirements, including multiple media rates (both low and high), a time span that exceeded 24 hours, and retro compatibility with SMPTE ST 12.”

Above all, they wanted the new version to have a “Digital Birth Certificate” (DBC) which meant including a precise time of acquisition, a persistent identifier that would stipulate whatever camera created the recording or stream, and the media rate.”

A drafting group was formed, chaired by John Willkie of Luxio to define the overall data structure, the content of each data field, and the rules for parsing the data to extract the information.

The group initiated a project, making it as extensible as possible in case new suggestions were added to the proposal.

Hohn Wilkie gets drafted to structure the data<br />(click to expand)

Hohn Wilkie gets drafted to structure the data
(click to expand)

“Eventually, getting tired of calling the goal of this project the ‘thingy’, we settled on the ‘TLX Project’ for ‘Time Label eXtensible’,” Symes said, “and came up with the idea of a tiered structure of profiles so that people could know what items are required in the DBC and what are not.”

They tried to leave the profile as flexible as possible, and Symes was clear that this is still a project in process.

Thanks to KLV's interoperability, it has also been adopted by the Motion Imagery Standards Board.<br />(click to expand)

Thanks to KLV's interoperability, it has also been adopted by the Motion Imagery Standards Board.
(click to expand)

“A profile is not terribly difficult to define,” he said. You just say ‘this is mandatory and ‘this is optional’ and set ‘if this, then that’ rules and set the maximum size. Then let people work with it.”

The TLX structure will be set as KLV (Key-Length-Value), a data encoding standard, often used to embed information in video feeds.

Much work remains for all of us<br />(click to expand)<br />

Much work remains for all of us
(click to expand)

“We are also looking at other profiles, in fact, this TLX project is still very fluid,” Symes said. “So if you want to get involved, you are heartily invited. We are working on the document suite being called ST 2120.”

“And, if you have a better sentence than the second one in this slide, feel free to suggest it.”

His main goal was to invite maximum participation.

“If you have ideas, or even better ideas than we’ve come up with, please join us,” he said.

Let us know what you think…

Log-in or Register for free to post comments…

You might also like...

ZombieLoad And Other Things That Go Bump In The Night

May 14, 2019 may not have seemed a particularly important date for those who edit and color grade on Mac’s and PC’s. But it was. By chance, that day I went looking for the May Windows 10 Feature Update (1903). I was sur…

Building On IP COTS

Transitioning to IP improves flexibility and scalability, both of which are achievable using COTS IT equipment. But can COTS solve every challenge? Or does broadcasting still have some unique and more demanding requirements that need further attention? In this article,…

HDR - Part 2 - Brightness Encoding

Dealing with brightness in camera systems sounds simple. Increase the light going into the lens; increase the signal level coming out of the camera, and in turn increase the amount of light coming out of the display. In reality, it’s…

Color and Colorimetry – Part 3

The human visual system (HVS) sees color using a set of three overlapping filters, which are extremely broad. As a result, the HVS is completely incapable of performing any precise assessment of an observed spectrum.

HDR - Part 1 - The State of HDR

Over the century or so we’ve been making moving images, a lot of improvements have been dreamed up. Some of them, like stereo 3D and high frame rate, have repeatedly suffered a lukewarm reception. Other things, like HD, and e…