Volumetric Video Takes Gold on the Live Events Stage
Milano Cortina 2026 marked an important step forward for the technology
The professional video industry's #1 source for news, trends and product and tech information. Sign up below.
You are now subscribed
Your newsletter sign-up was successful
A gold medal final lasts seconds. At Milano Cortina 2026, the short track speed skating final—in which Jordan Stolz and Femke Kok seemingly made short work of obliterating Olympic records—lasted just over half a minute. Britain, meanwhile, took a breathtaking gold in the mixed team snowboard cross by a narrow 0.43 seconds.
In this environment, when the starting gun fires, every framing decision must be locked in. As with all live sport broadcasting, the room for error is small, and there can be no second takes or resets to get a better shot. There’s a production risk that can’t be avoided; cameras can miss moments, angles can obstruct. While traditional broadcast innovation, like rail cameras and first-person view drones, has narrowed that risk and brought the viewer closer to the action, it still operates within a flat, fixed perspective.
How Volumetric Video is Changing Sports and Music
Volumetric video changes the capture model itself. By recording subjects from multiple viewpoints simultaneously, it preserves the spatial performance rather than committing to a single camera angle at the point of capture. Producers can then reposition the virtual camera in post-production, even if the final output is rendered in conventional 2D. The end result is a different relationship between time, perspective and editorial control.
This distinction matters in high-profile sporting arenas. Early deployments, most notably at Paris 2024, showed that volumetric replays could be successfully integrated into top level sports coverage. While often constrained to replay segments that required additional processing time, they showed clear editorial value.
Milano Cortina 2026 marked a further step forward for the technology. AI-powered replay systems built on volumetric capture delivered significantly improved visual fidelity, with some sequences approaching almost cinematic standards. The quality leap was immediately visible, signalling that volumetric workflows are evolving from experimental enhancements into credible broadcast tools.
The uneven pace of adoption across sectors, however, reflects differing priorities. In music and selective creative productions, volumetric capture has enabled directors to defer certain camera decisions until post-production. Artists from Radiohead to A$AP Rocky have captured music videos entirely volumetrically, demonstrating how the boundary between capture and creative decision-making can be collapsed, freeing the shot from being permanently defined on set.
Why Aren’t Blockbusters Keeping Up?
Cinema presents a much tougher challenge. Blockbuster filmmaking is deeply director-led. Framing decisions are deliberate and often central to narrative intent. Sets, lighting plans and blocking are constructed to be seen from specific angles. Volumetric capture fundamentally shifts this by decoupling capture from final framing.
The professional video industry's #1 source for news, trends and product and tech information. Sign up below.
The hesitation, therefore, is not primarily about technical feasibility, but about authorship and embedded workflows. Directors must reconsider how their control ebbs and flows in a spatial medium where perspective is programmable after the cameras stop rolling.
As a result, widespread volumetric capture of entire narrative features remains unlikely in the near term. More commonly, multi-view rigs are being integrated selectively into visual effects pipelines, where the flexibility they provide aligns with existing post-production processes. The technology’s strengths are tangible today, but they are being applied pragmatically rather than universally.
What Needs to Happen Next
Creative convention may shape adoption in cinema, but engineering constraints are still limiting deployment in broadcast. Live production environments require reliability under stress. It is one thing for volumetric workflows to work in controlled demonstrations, but another to scale them predictably.
Visual quality is typically the first compromise under real-time constraints. Techniques such as 4D Gaussian splatting can produce high-fidelity representations, but they introduce latency. Generating these models requires iterative learning processes that cannot be accelerated through forcefully adding in more parallel compute resources. Even emerging single-camera approaches struggle to scale effectively in multi-camera environments, which remain essential for full volumetric capture.
Bandwidth heightens the challenge. Uncompressed volumetric datasets are substantial, and even with compression, bitrates can still reach tens or hundreds of megabits per second. Distributing that at broadcast scale, particularly to large audiences or mobile devices, remains a complex challenge.
In short, the industry’s next breakthrough is unlikely to come from capture hardware alone, and will depend on mature, interoperable ecosystems for compression, transport and decoding of volumetric formats. That said, the direction of travel is clear. Just as first-person view drones and cloud-based production have altered expectations of live sports coverage, volumetric workflows point toward a future in which perspective is no longer permanently fixed at capture. In the near term, they will continue to operate within director-led 2D outputs, adding flexibility upstream without disrupting established viewing experiences. Over time, as delivery systems evolve and extended reality devices mature, the possibility of viewer-controlled viewpoints may move from demonstration to mainstream.
For now, volumetric video has flourished where the stakes are highest and the moments are unrepeatable. In the Olympic arena, where fractions of a second define history, preserving every possible perspective is vital.
Lauri Ilola is Principle Engineer for Nokia
