Jump to content

Deep image compositing

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Mild Bill Hiccup (talk | contribs) at 10:22, 26 October 2010 (Compositing deep data images: spelling). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Deep image compositing is a recently emerged way of compositing and rendering digital images. In addition to the usual color and opacity channels a notion of depth is created. This allows multiple samples in the depth of the image to make up the final resulting color. This technique is producing high quality results and removes artefacts around edges that could not be dealt with otherwise.

Deep data

Deep data is image that that holds information about the image along the depth of the image. This data can be stored in different formats.

Function based (integrated)

The data is stored as a function of depth. This results in a function curve that can be used to look up the data at any arbitrary depth. Manipulating the data is harder.

Sample based (deintegrated)

Each sample is considered as an independent piece and can so be manipulated easily. To make sure the data is representing the right detail an additional expand value needs to be introduced.

Generating deep data

3D renderers produce the necessary data as a part of the rendering pipeline. Samples are gathered in depth and then combined. The deep data can be written out before this happens and is so nothing new to the process. Generating deep data from camera data needs a proper depth map. This is used in a couple of cases but still not accurate enough for detailed representation. For basic holdout task this can be sufficient though.

Compositing deep data images

Deep images can be composited like regular images. The depth component makes it easier to determine the layering order. This must traditionally be input by a user. Deep images have that information for themselves and need no user input. Edge artifacts are reduced as transparent pixels have more data to work with.

History

Deep Images have been around in 3D rendering packages for quite a while now. The use of them for holdouts was first done at several VFX houses in shaders. Holdout mattes can be generated at render time. Using them in a more interactive manner was started recently by several companies, SideFX integrated it in their Houdini software and facilities like DreamWorks Animation, Weta, AnimalLogic and DRD studios have implemented interactive solutions.

Resources