Standards: Part 8 - Standards For Designing & Building DAM Workflows

This article is all about content/asset management systems and their workflow. Most broadcasters will invest in a proprietary vendor solution. This article is designed to foster a better understanding of how such systems work, and offers some alternate thinking for those interested in building their own systems.


This article is part of our growing series on Standards. There is an overview of all 26 articles in Part 1 -  An Introduction To Standards.


Buying a vendor asset management solution is a pretty safe and common solution but it doesn’t mean you don’t need to understand or put effort into molding the system to your requirements. Even the most flexible products will require extra work to customize them to your exact needs. Gaining a better fundamental understanding of how the system works can help you achieve more effective systems.

For those seeking to build their own systems, this article advocates challenging convention to see what benefits some more lateral approaches might bring.

There are useful metadata standards for the data model and components used to build workflows but the controlling framework is not easily standardized. Use open-standards if possible and proprietary standards when necessary.

File Based Workflows

Processing media files is quite straightforward. Deploying a Content Management System (CMS) to keep track of the media solves problems but it also introduces some complexity. The CMS has to perform several tasks:

  • Creation and maintenance of metadata.
  • Ingest and tagging of new content.
  • Physical storage management with redundancy in case of loss.
  • Library and archive curation.
  • Search and retrieval.
  • Automated background processes to create multiple alternatively formatted versions for deployment.
  • Workflow orchestration, scheduling and control.

We must be diligent (and possibly somewhat draconian) when it comes to adding metadata at the point of ingest for this to work reliably.

Metadata & Data Models

Each metadata standard serves a different discipline. No single standard covers everything and some schemas are informal. Metadata needs to:

  • Describe content.
  • Identify assets.
  • Locate assets within the storage.
  • Index the assets.
  • Describe technical processes operating on the content.
  • Provide administrative support.
  • Control access permissions.
  • Manage version control.
  • Log access and usage.
  • Track changes.
  • Manage long-term archives.
  • Facilitate preservation and conservation.

These involve the use of many different standards. Here are some that you might find useful. Use their profiles to constrain the range of values:

Standard Description
ISO 15836 Dublin Core.
EBUCore Based on Dublin Core.
ISO 19115 Geographic information.
ISO 14721 Open Archival Information System (OAIS).
PREMIS Preservation Metadata: Implementation Strategies.
NISO MIX Image content description.
PBCore Used by Public Service Broadcasters.
ISO 15938 MPEG-7 - Multimedia descriptions.

 

These are starting points when designing your own data model to steer the workflow. More standards are described at the Digital Curation Centre (DCC) web site.

Spine & Role Pattern

Repurpose the RDBMS spine and role pattern to invoke optional schemas as needed to support your workflow without having unreasonably wide database tables.

The spine record describes where the media is located but it is not swamped by irrelevant data. Additional properties are joined using relational links only when they are needed. The camera and lens details facilitate visual effects work. The shooting location helps when searching for related clips. Collect relevant properties together into a single role only when they are used together.

Research the use of Object-Oriented Patterns to discover some additional robust and well-conceived designs to repurpose for your own needs. These are well documented with much helpful guidance.

Learn From Other Disciplines - Semantic Searching

Scientific Visualisation includes graphical animations of weather patterns and other huge data sources that have too much information for a human to grasp. Animation provides 'time' as an additional axis to display dynamic content. Translating a problem into a different domain reveals an alternative analytical solution. This is an elegant and often simple approach which is fascinating and now informs all of my engineering designs.

The IN-SPIRE system was developed by Pacific Northwest Labs in the mid 1990s. It revolutionized the analysis of textual information and presents the deductions graphically. It was the precursor to what we now think of as Artificial Intelligence.

Moving things to a different domain allows us to deploy radically different algorithms. Arrange database search output onto an X-Y grid using two results criteria for the axes. Then apply spatial-geometric techniques to find related items.

IN-SPIRE uses this technique to take an existing document and parse it to create filters that locate a collection of similar documents. This goes way beyond simple keyword searching. I think we could use AI techniques like this to implement video searching tools that are hugely powerful.

The IN-SPIRE system encourages us to develop more sophisticated processing of our search results to analyze textual information embedded inside our media.

Phonetic Searching

Simple phonetic techniques will improve the reliability and yield of your search queries.

Store a phonetic degeneration of the searchable keywords in an auxiliary column. Apply the same algorithm to the input search key. This locates matching records in low-quality ingested content derived from live sub-titling systems and OCR scans of burned-in text.

The search will correctly match 'PENNSYLVANIA' and 'PENCILVANEYA' as a search key because both will converge on the same phonetic degeneration 'PENSILVANA'.

This is remarkably simple to build and has negligible effects on system performance. Add these enhancements to further improve the yield from your search results:

• Add numeric values as fully spelled out text.
• Add derived values where users may miss-type common characters such as 1 vs small letter L or Capital O vs. zero (0).
• Add Stemming support to match words that derive from a common root (swim, swum, swimming).
• Add Lemmatisation support which matches related words that are semantically similar as you would find in a Thesaurus (swim vs. diving and pool).

Document-centric Vs. Application-centric Design

If you conceive your workflow based on individual apps, you are limited by the file formats they support. Working out a streamlined path through the workflow becomes more difficult.

Turn the problem inside out by making the workflow document-centric rather than application-centric. Then find apps and tools that can operate on those document types. This simplifies the interchange between different parts of your workflow. The concept is illustrated here with a simple text document.

Controlling Tools & Software Apps With Scripts

When the media repository contains many thousands of items, a manual solution is no longer practical. Automation is best implemented as script-driven command-line processes.

Many applications are built on a foundation library that supports a screen-based interaction and a script-driven command-line. In fact, there may be many levels of scripting (like an onion skin). Carefully studying the internals of your applications reveals this functionality:

These apps support command-line-scripted control:

Tool Scripting interface
Blender A free and open-source tool for rendering animated 3D graphics. Includes a basic video editor that supports cutting, splicing masking and colour-grading.
ffmpeg A powerful command-line tool for converting video and audio media.
DaVinci Resolve Call-to-action in headless mode from the command-line. Internal script-driven tools can edit video and manufacture visual effects.
Adobe applications Historically these have supported command-line operating modes. Usually controllable with PowerShell and AppleScript. They also have internal scripting support (called actions).
Media 100 An ancient video editor that only works on older legacy macOS systems but is extremely scriptable.
OpenOffice An open-source tool for manipulating documents. Install the command-line tools in a server and call them from PHP.
GhostScript An open-source command-line tool to convert PostScript content into pixel images.
ImageMagick An open-source command-line tool that works similarly to Photoshop.
Windows apps Use PowerShell programming to control apps.
macOS apps Use AppleScript, Apple Events and OSA to control apps.

 

Bridging Control Between Different Environments

If scripting is not directly available at the command-line, there may be a way to bridge across and call it indirectly. Learning how to do this is another example of moving a problem to a different domain. Remote execution or triggering via service calls facilitates cross-platform integration.


In the 1990's the Showtime network automatically created interstitial stings and trails by steering an AppleScript-driven Media 100 video editor from an Excel Spreadsheet.


In macOS, the GUI level scripting is accessible from the command-line shell through the Open Scripting Architecture (OSA) with the osascript command. On Windows you can install Cygwin and call PowerShell scripts from there. On Linux it is all natively supported.

This opens up your workflow processing choices to utilise apps that could only ever be run manually on a Windows or macOS platform. Now they are accessible from your main workflow orchestration running on Linux.

Workflow Orchestration, Scheduling & Control

Build the workflow from a collection of processes running in a distributed fashion across multiple machines. Those processes implement these activities:

  • Movers.
  • Converters.
  • Renderers.
  • Filters.
  • Front-ends to hardware.

Avoid hard-coding the behaviors and use a data-driven approach and present the control surface in a web-based user interface. A PHP scripted back-end creates data files to initiate tasks and forward the necessary controls. Command-line driven tools wrapped in shell scripts then acquire the parameters to steer those tasks.

Here are some of the high-level constructs we use to build these processes:

Component Description
Watch folder Trigger activity when a file arrives.
Queue Based on watch folders. Task names are sorted by the file-system which implements prioritized execution for free.
Queue manager Load jobs into several queues like a multi-lane highway. Large jobs are processed separately to small quickly completed jobs. This improves latency and throughput.
Scheduler Run CPU intensive work overnight or load the queues with housekeeping tasks.
Distributed execution Delegate work to multiple machines.
Service listener Fire up processes on demand and kill them when completed to save resources.

 

Micro-service Architectures

Construct micro-service architectures from small component modules which are easily virtualized on cloud-based systems. They follow Object Oriented principles and only communicate via a well-defined API with their inner workings concealed. Data is often passed in a JSON structure but this is not mandatory.

The API for a micro-service can be implemented in a variety of ways:

Type Description
REST Each call stands-alone and session-state information is not preserved.
Endpoints Essentially a stateful web page called with a URL that may also pass additional information.
Resource control HTTP methods are used to POST (install/edit) or DELETE resources inside the service.
Gateway Aggregate several services into a single location dispatching requests to child processes.

 

A micro-service responder could be built with inexpensive devices such as a Raspberry Pi single board computer. These can access real-world hardware using Internet of Things (IoT) principles. This makes remotely operated steerable cameras, lights, auto-scripters and other hardware accessible to your workflow.

Conclusion

Think differently by moving problems to a different domain so they can be implemented with less work. Less code equals easier maintenance and therefore fewer things to go wrong. Apply radically different domain-specific algorithmic techniques. Micro-services help to reduce complexity. Bridging techniques allow you to exploit previously unreachable features. Searches can be made much smarter with quite simple extensions. Build your workflow with small re-usable functional blocks.


These techniques and many others are covered fully in my earlier book:
Developing Quality Metadata.
(ISBN 978-0-240-80869-7).

Part of a series supported by

You might also like...

What Are The Long-Term Implications Of AI For Broadcast?

We’ve all witnessed its phenomenal growth recently. The question is: how do we manage the process of adopting and adjusting to AI in the broadcasting industry? This article is more about our approach than specific examples of AI integration;…

The Big Guide To OTT: Part 10 - Monetization & ROI

Part 10 of The Big Guide To OTT features four articles which tackle the key topic of how to monetize OTT content. The articles discuss addressable advertising, (re)bundling, sports fan engagement and content piracy.

Next-Gen 5G Contribution: Part 2 - MEC & The Disruptive Potential Of 5G

The migration of the core network functionality of 5G to virtualized or cloud-native infrastructure opens up new capabilities like MEC which have the potential to disrupt current approaches to remote production contribution networks.

Designing IP Broadcast Systems: Addressing & Packet Delivery

How layer-3 and layer-2 addresses work together to deliver data link layer packets and frames across networks to improve efficiency and reduce congestion.

The Business Cost Of Poor Streaming Quality

Poor quality streaming loses viewers at an alarming rate especially when we consider the unintended consequences of poor error reporting on streaming players.