GECCO: Geometrically-Conditioned Point Diffusion Models

Michał J. Tyszkiewicz (EPFL), Pascal Fua (EPFL), Eduard Trulls (Google)

Published in ICCV, 2023

Teaser (interactive)

Image
Ground truth
GECCO
LinkDescription
Read the paper on arXivRead the paper on arXiv
Get source code on GithubSource code release (JAX and PyTorch)
Play with ShapeNet in ColabPlay with GECCO trained on ShapeNet
Play with Taskonomy in ColabPlay with GECCO trained on Taskonomy
View interactive examples on ShapeNetShapeNet sample gallery (not cherrypicked, may fail in Firefox)
View interactive examples on TaskonomyTaskonomy sample gallery (not cherrypicked, may fail in Firefox)

Abstract

Diffusion models generating images conditionally on text, such as Dall-E 2 and Stable Diffusion, have recently made a splash far beyond the computer vision community. Here, we tackle the related problem of generating point clouds, both unconditionally, and conditionally with images. For the latter, we introduce a novel geometrically-motivated conditioning scheme based on projecting sparse image features into the point cloud and attaching them to each individual point, at every step in the denoising process. This approach improves geometric consistency and yields greater fidelity than current methods relying on unstructured, global latent codes. Additionally, we show how to apply recent continuous-time diffusion schemes. Our method performs on par or above the state of art on conditional and unconditional experiments on synthetic data, while being faster, lighter, and delivering tractable likelihoods. We show it can also scale to diverse indoors scenes.