Earlier this week AMD announced the new AMD Milan-X CPU. This new chip harnesses the Ryzen 3D V-cache technology previously demonstrated in August this year. Providing users with similar features as the AMD EPYC 7003 series + a massive boost in L3 cache capacity.
LED panel calibration is a essential element of any in-camera VFX production. The calibration across panels is designed to address three main areas: colour / brightness formatting, uniformity and seamlessness. Correct colours on the LED stage itself is one of the most key, yet difficult processes of calibration. Brompton Technologies is combating this issue with the new Dynamic Calibration engine in the Tessera processers. Title image credit MGX Films.
Nanite is the new and improved virtualized geometry system implemented in Unreal Engine 5. Nanite provides a internal format for imported meshes, creating “Nanite Meshes” instead of the previous “Static Meshes”. Through the use of new rendering technology Nanite is able to render pixel scale and high poly counts effortlessly. Nanite harnesses new Level of Detail (LOD) rendering methods which allow for differing detail and polycounts from every unique perspective.
The first On-Set GODBOX comes of the production line in the UK.
Looking forward to our trip to Vancouver, one of the largest VR/AR production hubs in the world and home of our Canadian Partners. We can’t wait to see what virtual production gems Vancouver has to offer.
In tech terms the word metaverse refers to a pervasive, always on, digital multi-dimension, that you can see and interact with, think of the internet with millions of web pages you access through a browser, now imagine the internet on everything around you, you can see it, touch it. Also, it’s not one virtual world or dimension, its many, no one company will own the metaverse.
You switch on your AR glasses, and you see a virtual, real-time, overlay of the metaverse, graphics that you can touch over your real world. You can jump in and out of this digital twin universe, it’s AR, VR, you flick through services, apps ways to communicate and collaborate, on/off.
That’s the visual internet or Metaverse dream.
Continue reading The Metaverse, what it is and why its happening
The first independent VP test studio of it’s kind in the UK, the new Digital Catapult and Target3D run research and innovation studio will allow independent and early stage production companies in the media and creative industries to experiment with Virtual Production Technologies, and develop new tools and applications.
The race is on, there’s a whole new breed of creative on the way and it’s going to make the shift from shooting on film to digital, look like Pac Man.
In tech circles there is something called coopetition, its where competitive companies work together to serve and innovate to the benefit of the whole (aka NVIDIA, Intel). As they say, “a rising tide lifts all” and that’s the case now with the creative production industry too.
What is Nvidia Broadcast?
Nvidia Broadcast is one of few software’s which provides a AI driven compositing solution to users. Nvidia Broadcast allows users to apply effects over a in-coming camera feed, in this demonstration we used the “replace background” feature allowing us to get a composite without the need of any greenscreen; the way Nvidia archives this ability is by harnessing the power of both machine learning image analysis and depth perception.
Once we have our composited green background in Nvidia Broadcast we move forwards to streaming the feed into Unreal Engine. From the real-time engine we have the ability to place the talent into any 3D environment and now aren’t limited to the previous 2D dimensions.
See below our tutorial and demonstration video:
Relevant links and software’s:
- Nvidia Broadcast
- Unreal Engine 4
- OBS Studio
- Off-World-Live / Spout2 (OBS plugin)
- Off-World-Live (Unreal Engine Plugin)
Moving forward, this proof of concept has proved itself to work and now needs officially testing back in our OSF lab on a GODBOX™ with a 4K camera feed input.
Dark Bay a virtual production company released an eagerly awaited behind the scenes photos from the set of the new Netflix show, Babelsberg 1899 the hotly anticipated new mystery series from Jantje Friese and Baran bo Odar, creators of global hit Dark.
Shooting in Berlin this is the first behind the scenes image to be officially released by the production team. It shows the LED stage and the sett of the ship. One of Europe’s largest, permanently installed LED studios for mixed reality film productions has officially opened: the DARK BAY Virtual Production Stage at Studio Babelsberg near Berlin. With the support of the ARRI Solutions Group, ARRI was responsible for technical coordination of the installation and all components, including more than 70 ARRI SkyPanels. “For productions of the highest quality, state-of-the-art hardware was combined with engineering services to create an innovative overall system in close coordination with the client and business partners,” comments Markus Zeiler, Executive Board member of ARRI. First production to be shot in the DARK BAY is Netflix’s “1899.” The new series by Jantje Friese and Baran bo Odar is captured with ALEXA Mini LF and lenses specially created by ARRI Rental.
Image copyright. Alex Forge / Netflix.