A new animated lighting system lets movie directors improve shots in seconds.
Perfecting a shot for an animated film or a special-effects sequence is a very incremental process. Every time the director tweaks the lighting, an enormous amount of computation is required to render the new image, and that rendering process can take a very long time.
Now a team of computer scientists from MIT, Tippett Studios, and Industrial Light and Magic (ILM) have devised a system that reduces the amount of time required to render a preview image from nearly an hour to seconds, allowing directors to fine-tune the lighting in a shot immediately. ILM tested the system, called Lightspeed, on the movie Transformers, and it plans on deploying it throughout the whole company in the next couple of weeks.
“We are still rolling it out,” says Christophe Hery, the lead engineer of research and development at ILM. “But potentially, what used to take three or four days to produce might be compressed into a single day.”
The team’s solution is based on the fact that lighting designers are working at the end of the production process. Since everything else in the image has largely been set, much of the data involved in the rendering process is redundant. So, to accelerate the process, Lightspeed identifies and then compresses the data that is not changing in the image each time it’s rendered so as to avoid redundancy.
Next, Lightspeed takes advantage of high-performance graphic processors (GPUs). Traditionally, when a lighting designer renders an image, that work is performed entirely on a central processing unit (CPU). The Lightspeed system, in contrast, caches the redundant data on the CPU and performs the remaining computations for re-executing the lighting programs on the GPU. Managing the data in this way makes previewing an image orders of magnitude faster than running it entirely on the CPU.
“The first big step is eliminating work that doesn’t have to be recomputed every frame,” says Jonathan Ragan-Kelley, a computer scientist at MIT and a Lightspeed team member. “The next big acceleration comes from taking that data [that] lighting designers are editing, and then mapping it onto a processor that can execute it much more efficiently.”
The Lpics preview system used by Pixar Animation Studios employs a similar method to render preview images quickly. But Lpics requires a programmer to manually go through and identify what data in an image is going to change and what isn’t when making different preview images. Moreover, this process has to be redone for Lpics anytime the lighting programs change to capture a different lighting effect, which often happens during production.
The other improvement over Lpics, according to Ragan-Kelley, is that the Lightspeed preview system supports additional effects, such as motion blur and transparency, in which more than one point in a scene contributes to the color of an individual pixel.
“They went for a very nice solution that guarantees accuracy, especially in small scenes with lots of details,” says Fabio Pellacini, a computer scientist at Dartmouth College and one of the creators of Pixar’s Lpics system. “We are seeing improvements coming very quickly online these days, but difficulties remain for handling complex images where light reflects across a variety of objects from different angles. Hopefully, these problems will be solved soon.”
Be there when AI pioneers take center stage at EmTech Digital 2019.Register now