The first workshop on

AI for 3D Generation

June 17, @CVPR2024, Seattle WA, USA

Remote attendees can join via zoom!

Developing algorithms capable of generating realistic, high quality 3D data at scale has been a long standing problem in Computer Vision and Graphics. We anticipate that having generative models that can reliably synthesize meaningful 3D content will completely revolutionize the workflow of artists and content creators, and will also enable new levels of creativity through ``generative art". Although recently there has been considerable success in generating photorealistic images, the quality and generality of 3D generative models has lagged behind their 2D counterparts. Additionally, efficiently controlling what needs to be generated and scaling these approaches to complex scenes with several static and dynamic objects still remains an open challenge.

In this workshop, we seek to bring together researchers working on generative models for 3D shapes, humans, and scenes to discuss the latest advances, existing limitations and next steps towards developing generative pipelines capable of producing fully controllable 3D environments with multiple humans interacting with each other or with objects in the scene. In the last few years, there has been significant progress in generating 3D objects, humans, and scenes independently, but only recently has the research community shifted their attention towards generating meaningful dynamics and interactions between humans or humans and other scene elements. To this end, in our workshop we look forward to cover the following topics:

  • What is the best representation for generating meaningful variations of 3D objects with texture and high quality details?
  • What is the best representation to enable intuitive control over the generated objects?
  • How to synthesize realistic humans performing plausible actions?
  • How to generate fully controllable 3D environments, where it would be possible to manipulate both the appearance of the scene elements as well as their spatial composition?
  • What is the best representation for generating plausible dynamics and interactions between humans or humans and objects?
  • What are the ethical implications that arise from artificially generated 3D content and how we can address them.

  • April 7, 2024: We have extended the paper submission deadline by a couple of days! The new paper and supplemental material deadline is on April 15 (AoE)!!.
  • January 25, 2024: Workshop website launched, with the tentative list of the invited speakers announced.
Call for Papers
We accept two forms of papers:
  • Long paper: Long papers should not exceed 8 pages excluding references and should use the official CVPR template. Long papers are intended for presenting mature works, should describe novel ideas but also include extensive experimental evaluations that support the proposed ideas.
  • Short paper: Short papers should not exceed 4 pages excluding references and should also use the official CVPR template. Short papers are intended for presenting ideas that are still at an early stage. Although comprehensive analyses and experiments are not necessary for short papers, they should have some basic experiments to support their claims. Moreover, in the short paper track, we encourage submissions focusing on creative contributions demonstrating applications of existing technology into 3D content creation pipelines. For example, we look forward for submissions showcasing how ongoing research on 3D generative AI can be used to facilitate the workflow of experienced as well novice users in various fields such as architectural engineering, product designing, education, art, entertainment etc.

All submissions should anonymized. Papers with more than 4 pages (excluding references) will be reviewed as long papers, and papers with more than 8 pages (excluding references) will be rejected without review. Supplementary material is optional with supported formats: pdf, mp4 and zip. All papers that were not previously presented in a major conference, will be peer-reviewed by three experts in the field in a double-blind manner. In case you are submitting a previously accepted conference paper, your submission does not need to be anonymized. For already accepted conference papers, please also attach a copy of the acceptance notification email in the supplementary material documents.

Please not that the accepted papers will NOT be included in the IEEE/CVF proceedings, but will have a poster presentation the day of the workshop.

Submission Website:

All submissions should follow the CVPR paper format:

Paper Review Timeline:

Paper Submission and supplemental material deadline Monday, April 15, 2024 (AoE time)
Notification to authors Friday, May 10, 2024
Camera ready deadline Wednesday, May 15, 2024

Keynote Speakers
Andrea Vedaldi
University of Oxford and Meta AI
Gordon Wetzstein
Stanford University
Jun-Yan Zhu
Carnegie Mellon University
Varun Jampani
Stability AI
Sergey Tulyakov
Snap Research
Sanja Fidler
University of Toronto and NVIDIA Research
Spotlight Speakers
Duygu Ceylan
Adobe Research
Ruoshi Liu
Columbia University
Jun Gao
University of Toronto and NVIDIA Research
Alex Yu
Luma AI
Aleksander Holynski
UC Berkeley and Google Research
Qianqian Wang
UC Berkeley
Dongsu Zhang
Seoul National University
Despoina Paschalidou
Georgios Pavlakos
UT Austin
Davis Rempe
NVIDIA Research
Angel Xuan Chang
Simon Fraser University
Kai Wang
Amlan Kar
University of Toronto and NVIDIA Research
Daniel Ritchie
Brown University
Kaichun Mo
NVIDIA Research
Manolis Savva
Simon Fraser University
Paul Guerrero
Adobe Research
Siyu Tang
ETH Zurich
Leonidas Guibas
Relevant Workshops on the Content Creation Track @ CVPR2024
Relevant Previous Workshops