Crossing the Uncanny Valley with Volumetric Capture
Microsoft has adopted technology from IO Industries to raise Virtual Reality production to new heights of realism and flexibility.
The technology of video production, especially as it relates to VR (Virtual Reality) and AR (Augmented Reality), is about to take a giant leap forward. In fact, it’s about to jump a valley—the famous uncanny valley.
The concept of the “uncanny valley” was identified by the robotics professor Masahiro Mori as “bukimi no tani genshō” (不気味の谷現象) in 1970. The term was first translated as “uncanny valley” in the 1978 book Robots: Fact, Fiction, and Prediction, written by Jasia Reichardt, but it has become a guidon for animators, c.g. artists and computer game creators as a warning that, just as Commander John J. Adams (Leslie Nielsen) said to Altaira (Anne Francis) at the end of 1956’s “Forbidden Planet”, “We are, after all, not God.”
The abyss of the uncanny valley. BTW, “bunraku” refers to large Japanese puppets manipulated by actors (click to expand).
As this graph shows, crafting an imitation of life-like humans can be entertaining, up to a point. But if you come too close, and there is still something unidentifiable missing, it becomes creepy.
Uncanny.
That’s why, up to now, most motion capture approaches, no matter how pseudo god-like, have been better at creating Gollum than Swan Lake.
But the world of VR is changing.
Microsoft has been pursuing the concept of Volumetric Capture for the past seven years. As Andrew Searle, sales manager at IO Industries, explained to me, “The Microsoft Research Division had been working on holographic, free-viewpoint 3D capture video for some time,” he began, “with the idea of using it in their HoloLens head-mounted display, or Windows Mixed Reality platform. But about two years ago they ran into a roadblock with the scalability of their capture platform, so they reached out to us.”
IO Industries’ use of multiple high speed cameras and capture systems had been successful in scientific and research applications such as fluid dynamics, aerospace airfoils, and weapons testing.
But when I asked Searle if what IO Industries and Microsoft were working on could be referenced to the famous “bullet time” or “time slice” shot from 1999’s “The Matrix”, he acknowledged the similarities but used the differences as definitional.
“In ’The Matrix’, a ramp of cameras was set up around computer hacker Neo (Keanu Reeves) inside a circular green screen,” he explained, “and when Neo jumps they all capture one picture simultaneously. When merged together, all those images spin around Neo in one instance of time. With Volumetric Capture, this is done with moving video from multiple cameras in real time, so the action is not frozen.”
This lets the director move the viewpoint of the audience as the action demands, while maintaining the illusion of 3D depth.
Another useful reference is today’s motion capture techniques, where multiple angles from cameras looking at actors often wearing special suits are combined inside a computer to create a skeletal model of the person onto which a c. g. character can be modeled.
As we have seen in many brilliant motion capture films this can look highly realistic for fantasy characters—but never quite real enough to represent actual humans.
Hence the uncanny valley.
“Since Microsoft’s Volumetric Capture is recording the entire actor from all angles, all the muscles, all the facial expressions, all the soft tissue movements, we can cross that realism gap,” Searle told me.
Microsoft came to IO Industries for several reasons. They wanted a system that could be scaled from small studios to large installations, they needed a recording technology that could handle input from up to 100+ cameras, and cameras that could be synchronized to within a microsecond of each other.”
IO Industries' DVR Express raw recorders offer LTC timestamp, frame triggering and configurable SSD storage.
The system employs one additional rather clever trick. Only half of the cameras are outputting RAW, RGB color video. The other half are recording near-infrared images of the tiny dots projected onto the actors’, props’ and objects’ forms by several laser pattern projectors mounted all around the set.
“These overlapping near-IR cameras give us the texture and depth analysis that the Microsoft processing algorithm needs to provide the ultimate in shape rendering,” Searle said. “Motion capture suits can only fit so many reference dots on their surface. But this technique of using pinpoints of laser light gives us a much higher resolution for spatial and depth localization.”
Victorem camera uses advanced global shutter CMOS sensors with high dynamic range.
Searle wanted to be clear during our interview that the Volumetric Capture is the result of almost 7 years research invested by the Microsoft Research Division, but about two years ago they came to IO Industries for some key empowering technologies.
“They are using our Victorem high speed cameras,” he said. “Within a small form factor, they are capable of recording 2K/4K/HD/UHD and non-standard video formats with shutter speeds in a fraction of a second.”
Most importantly, since a common configuration of a Volumetric Capture utilizes up to 100+ cameras for each take, fully shutter-synchronized with microsecond-level accuracy and recording unprocessed RAW images, they generate over 10 Gigabytes of data per second.
Then to wrangle all that data, IO Industries StudioCap management software merges these files so they can be fed into Microsoft’s own processing engine from which it is output as real time, moving 360-degree imagery.
StudioCap management software gives you live video streaming with focus assist and production database integration.
Microsoft is not the only tech firm going after this Volumetric Capture concept for future production, but they are definitely on the forefront.
And making great progress in both VR and AR production using it.
“We’re going to be on the other side of the uncanny valley any time soon,” Searle concluded.
In August, a new studio called Metastage providing Volumetric Capture capabilities based on the Microsoft approach to VR and AR was opened at Culver Studios in Culver City, California. This joins Microsoft’s own Mixed Reality Capture Studios in San Francisco and Redmond, Washington. The system has also been licensed to London’s Dimension Studios in the UK.
You might also like...
NDI For Broadcast: Part 1 – What Is NDI?
This is the first of a series of three articles which examine and discuss NDI and its place in broadcast infrastructure.
Brazil Adopts ATSC 3.0 For NextGen TV Physical Layer
The decision by Brazil’s SBTVD Forum to recommend ATSC 3.0 as the physical layer of its TV 3.0 standard after field testing is a particular blow to Japan’s ISDB-T, because that was the incumbent digital terrestrial platform in the country. C…
Designing IP Broadcast Systems: System Monitoring
Monitoring is at the core of any broadcast facility, but as IP continues to play a more important role, the need to progress beyond video and audio signal monitoring is becoming increasingly important.
Broadcasting Innovations At Paris 2024 Olympic Games
France Télévisions was the standout video service performer at the 2024 Paris Summer Olympics, with a collection of technical deployments that secured the EBU’s Excellence in Media Award for innovations enabled by application of cloud-based IP production.
HDR & WCG For Broadcast - Expanding Acquisition Capabilities With HDR & WCG
HDR & WCG do present new requirements for vision engineers, but the fundamental principles described here remain familiar and easily manageable.