Skip to yearly menu bar Skip to main content


Poster

OBJECT 3DIT: Language-guided 3D-aware Image Editing

Oscar Michel · Anand Bhattad · Eli VanderBilt · Ranjay Krishna · Aniruddha Kembhavi · Tanmay Gupta

Great Hall & Hall B1+B2 (level 1) #222
[ ]
Thu 14 Dec 3 p.m. PST — 5 p.m. PST

Abstract:

Existing image editing tools, while powerful, typically disregard the underlying 3D geometry from which the image is projected. As a result, edits made using these tools may become detached from the geometry and lighting conditions that are at the foundation of the image formation process; such edits break the portrayal of a coherent 3D world. 3D-aware generative models are a promising solution, but currently only succeed on small datasets or at the level of a single object. In this work, we formulate the new task of language-guided 3D-aware editing, where objects in an image should be edited according to a language instruction while remaining consistent with the underlying 3D scene. To promote progress towards this goal, we release OBJect: a benchmark dataset of 400K editing examples created from procedurally generated 3D scenes. Each example consists of an input image, editing instruction in language, and the edited image. We also introduce 3DIT: single and multi-task models for four editing tasks. Our models show impressive abilities to understand the 3D composition of entire scenes, factoring in surrounding objects, surfaces, lighting conditions, shadows, and physically-plausible object configurations. Surprisingly, training on only synthetic scenes from \dataset, editing capabilities of 3DIT generalize to real-world images.

Chat is not available.