Preview: ILM Uses ‘Star Wars’ Assets to Show Potential of Google’s ‘Seurat’ VR Rendering Technology
By Ben Lang
Google’s newly announced Seurat rendering tech purportedly makes use of ‘surface light-fields’ to turn high-quality CGI film assets into detailed virtual environments that can run on mobile VR hardware. The company gave Seurat to ILMxLab, the immersive entertainment division of Industrial Light and Magic, to see what they could do with it using assets directly from Star Wars.
Google just announced Seurat this week, a new rendering technology which could be a graphical breakthrough for mobile VR. Here’s what we know about how it works so far:
Google says Seurat makes use of something called surface light-fields, a process which involves taking original ultra-high quality assets, defining a viewing area for the player, then taking a sample of possible perspectives within that area to determine everything that possibly could be viewed from within it. The high-quality assets are then reduced to a significantly smaller number of polygonsâ€”few enough that the scene can run on mobile VR hardwareâ€”while maintaining the look of high quality assets, including perspective-correct specular lightning.
As a proof of concept, Google teamed with ILMxLab to show what Seurat could do. In the video above, xLab says they took their cinema-quality CGI rendersâ€”those which would normally take a long time to render each individual frame of final movie outputâ€”and ran them through Seurat to make them able to playback in real-time on Google’s mobile VR hardware. You can see a teaser video heading this article.
“When xLab was approached by Google, they said that they could take our ILM renders and make them run in real-time on the VR phone… turns out it’s true,” said Lewey Geselowitz, Senior UX Engineer at ILM.
Star Wars Seurat Preview
I got to see the Star Wars Seurat-renderedÂ experience teased in the video above for myself running on a prototype version of Google’s standalone Daydream headset.
When I put on the headset I was dropped into the same hangar scene as shown in the video. And while there’s no replacing the true high quality ray-traced output that comes from the cinematic rendering process (that can take hours for each frame), this was certainly some of the best graphics I’ve ever seen running on mobile VR hardware. In addition to sharp, highly detailed models, the floor had dynamic specular reflections, evoking the same sort of lightning you would expect from some of the best real-time visuals running on high-end PC headsets.
What’s particularly magic about Seurat is thatâ€”unlike a simple 360 video renderâ€”the scene you’re looking at is truly volumetric, and properly stereoscopic no matter where you look. That means that when you move your head back and forth, you’ll get proper positional tracking and see parallax, just like you’d expect from high-end desktop VR content.Â And because Google’s standalone headset has inside-out tracking, I was literally able to walk around the scene in a room-scale sized area with a properly viewable area that extended all the way from the floor to …read more
Read more here:: roadtovr.com