PhD Defense

Exploring and Rebuilding Reality: Learning 3D Reconstruction from Capturing Images to Building Photorealistic Representations

Antoine Guédon
IMAGINE / LIGM
École des Ponts ParisTech (ENPC)

Profile picture

  Date & Time


Date: September 25th, 2025
Time: 05:00 PM
Duration of the presentation: Approximately 45 minutes
Duration of the questions: Between 1 and 2 hours
Reception: A reception will be held around 9PM after the defense, at La Petite Place, 202 rue du Faubourg Saint-Antoine, 75012 Paris.

  Map


  Location


Venue: Amphithéâtre Caquot, Coriolis building
Address:
École des Ponts ParisTech
6-8, Av Blaise Pascal - Cité Descartes
Champs-sur-Marne
77455 Marne-la-Vallée cedex 2
France

  Thesis Committee


Thesis Advisors:
Vincent Lepetit (ENPC)
Pascal Monasse (ENPC)

Jury Members:
Valérie Gouet (LaSTIG - IGN) - President
Andrea Tagliasacchi (Simon Fraser University) - Reviewer
Maks Ovsjanikov (Ecole polytechnique) - Reviewer
Claire Dune (Université de Toulon) - Examiner

  Attendance


The defense is public and open to all.
Online streaming: Link is coming soon.
Contact: antoine (dot) guedon (at) enpc.fr for any questions

Thesis Abstract


This thesis addresses two fundamental challenges in computer vision: Autonomous scene exploration and photorealistic 3D reconstruction. While recent advances in neural rendering have revolutionized the field of 3D reconstruction, existing methods face significant limitations. They typically require dense sets of carefully captured images, struggle with geometric ambiguities, and often produce representations that are difficult to edit or integrate into standard graphics pipelines. Additionally, the challenge of efficiently acquiring these input images remains largely unexplored, creating a barrier for non-expert users attempting to create high-quality 3D content.

We present five complementary contributions that progressively tackle these challenges: For acquiring optimal input images through autonomous scene exploration, we first introduce SCONE and MACARONS. For photorealistic 3D reconstruction, we then propose SuGaR, Gaussian Frosting and MAtCha. More details are provided in the following sections.

Together, these contributions advance both autonomous exploration and 3D reconstruction by providing more practical, efficient, and accessible solutions for real-world applications. Our work bridges the gap between neural rendering and traditional computer graphics while making high-quality 3D reconstruction more accessible to non-expert users. The developed methods have potential applications across various domains, from virtual reality and digital content creation to robotics and cultural heritage preservation.

Presented Projects


This thesis encompasses five research projects that participated to advance the state-of-the-art in 3D reconstruction and neural rendering.

  Autonomous Scene Exploration

The first part of this thesis focuses on autonomous exploration strategies for efficiently capturing optimal input images for 3D reconstruction.

SCONE: Surface Coverage Optimization in uNknown Environments

NeurIPS 2022 (Spotlight)

Antoine Guédon, Pascal Monasse, Vincent Lepetit

We introduce SCONE, a novel mathematical framework for surface coverage-based exploration using depth sensors. Unlike previous approaches that rely on reinforcement learning, this supervised learning approach enables efficient exploration of arbitrarily large scenes with unrestricted camera motion. The framework's purely geometric considerations allow for better generalization to unseen scenes, making it particularly valuable for real-world applications.

  Paper     Code

MACARONS: Mapping And Coverage Anticipation with RGB ONline Self-supervision

CVPR 2023

Antoine Guédon, Tom Monnier, Pascal Monasse, Vincent Lepetit

We present MACARONS, which eliminates the need for depth sensors and explicit 3D ground truth by leveraging self-supervised learning from RGB images alone. This advancement enables online learning in large unknown environments, making autonomous exploration more practical and cost-effective. MACARONS learns to generate 3D reconstructions directly from RGB images and uses these reconstructions to supervise its coverage predictions, creating a self-improving system suitable for deployment in real-world scenarios.

  Paper     Code     Project Page

  Photorealistic 3D Reconstruction

The second part of this thesis develops novel neural rendering approaches for high-quality 3D reconstruction from captured images.

SuGaR: Surface-Aligned Gaussian Splatting for Efficient 3D Mesh Reconstruction

CVPR 2024

Antoine Guédon, Vincent Lepetit

We propose SuGaR, a method that combines explicit surface representation with Gaussian splatting for fast, high-quality mesh reconstruction. SuGaR first introduces a regularization strategy for extracting explicit surface meshes from Gaussian Splatting representations. Then, SuGaR introduces a hybrid representation that binds Gaussians to the extracted surface meshes and refines the structure through differentiable rendering. This approach achieves photorealistic rendering within remarkably short optimization times (less than an hour) and enables unprecedented fine-grained editing capabilities based on meshes, including surface sculpting, character animation, and scene compositing—features previously unavailable in radiance field models.

  Paper     Code     Project Page

Gaussian Frosting: Editable Complex Radiance Fields with Real-Time Rendering

ECCV 2024 (Oral)

Antoine Guédon, Vincent Lepetit

We extend SuGaR with Gaussian Frosting, which significantly improves material handling and rendering efficiency while maintaining SuGaR's interactive features. The method introduces a novel occlusion culling strategy and achieves rendering quality comparable to vanilla Gaussian Splatting. Notably, Frosting excels at reconstructing challenging materials like hair, fur, and grass, which are traditionally difficult to represent with pure surface-based methods but essential for building photorealistic scenes and virtual avatars. To democratize these capabilities, we developed a dedicated Blender add-on that enables artists and content creators to edit, sculpt, combine, and animate their reconstructions without programming expertise.

  Paper     Code     Project Page

MAtCha Gaussians: Atlas of Charts for High-Quality Geometry from Sparse Views

CVPR 2025 (Highlight)

Antoine Guédon, Tomoki Ichikawa, Kohei Yamashita, Ko Nishino

Finally, we introduce MAtCha, which tackles the challenge of sparse-view reconstruction. By leveraging learned priors, MAtCha enables high-quality mesh reconstruction from just 3-10 input images—a significant advancement over previous methods that required hundreds of images and complex preprocessing steps. The method produces sharp, detailed meshes of both foreground and background elements while maintaining high geometric quality across the entire scene. This capability might make high-quality 3D reconstruction more accessible to everyday users, potentially democratizing the technology for widespread adoption.

  Paper     Code     Project Page

Defense Schedule


05:00 PM

  • Introduction
  • Project 1: SCONE - Autonomous Exploration
  • Project 2: MACARONS - Self-Supervised Exploration from RGB images
  • Project 3: SuGaR - Surface-Aligned Gaussian Splatting
  • Project 4: Frosting - Editable Radiance Fields
  • Project 5: MAtCha - Sparse-View Reconstruction
  • Conclusions and Future Work

05:45 PM

Questions from the Jury (between 1 and 2.5 hours)

08:00 PM

Results Announcement and short toast at the Lab

After Defense

Reception in La Petite Place Paris 12, 202 rue du Faubourg Saint-Antoine, 75012 Paris

Resources


Contact Information

Antoine Guédon
PhD Candidate
IMAGINE Team, LIGM
École des Ponts ParisTech
✉ antoine (dot) guedon (at) enpc.fr

Links

  Email     GitHub     Scholar

© 2025 Antoine Guédon. You are welcome to copy the code, please attribute the source with a link back to this page.