Using Zero Density Reality Engine™ with AW-UE150 and AW-UE100 for Virtual Studio applications
Zero Density Reality Engine™
Enables access to a powerful, high-quality and affordable AR system for your virtual studio needs!
|Position Data notification|
|FreeD protocol via serial/IP|
|Wider angle of view (75.1°)|
|Simultaneous UHD and HD outputs|
The V-Log function was implemented in the Reality Engine starting with version 2.9, and increases the picture quality acquisition from Panasonic cameras including the UE150, UE100, AK-UC4000 and VariCam LT. The above cameras have the ability to output V-Log, meaning that the video signal can be processed more natively within Zero Density Reality Engine, providing a better output in the keyer, and increases the overall quality in the virtual studio production.
The lens of the camera is already calibrated in the Zero Density system and simply needs to be selected in the Lens node. The camera can of course be used both in Full HD or 4K. In short, using the UE150 and UE100 in combination with Zero Density Reality Engine™ allows many more users to get access to the AR and VR world thanks to this powerful, high-quality and affordable solution.
Reality Engine™ from Zero Density is the ultimate real-time node-based compositor enabling real-time visual effects pipelines featuring video I/O, keying, compositing and rendering in one single software in real-time. As the most photo-realistic virtual studio production solution, Reality Engine™ provides its client the tools to create the most immersive content possible and revolutionize story-telling in broadcasting, media or cinema industry
Reality Engine™ uses Unreal Engine by Epic Games, the most photo-realistic real-time game engine, as the 3D renderer. With the advanced real-time visual effects capabilities, Reality ensures the most photo-realistic composite output possible.
Reality Keyer™ provides spectacular results for keying of contact shadows, transparent objects and sub-pixel details like hair.
As a unique approach in the industry, Zero Density’s Reality Engine™ composites images by using 16 bit ﬂoating point arithmetics. Having the control of full compositing pipeline, all gamma and color space conversions are handled correctly. By using this technology, the real world and the virtual elements are blended perfectly.
Having the tracked camera feed as an input image, Reality Engine™ is able to render reﬂections and refractions of the real environment onto the virtual objects. Real-time 3D reﬂections and the shadows of the virtual objects are also composited over the incoming camera feed.
Take Back the Stadiums: Virtual Fans, Striking Advertisement and More Entertainment
Sports returned to the stadiums, but left audiences at home. This results in high ratings, but misses the traditional atmosphere. Why not use the current circumstances to create new opportunities for sport events in general? View an example of how to leverage high viewership with the latest technology here from Zero Density & Dreamwall with the AW-UE150 PTZ camera.
Showing 4 of 9