PanoFloor: Reconstruction and Immersive Exploration of Large Multi-Room Scenes from a Minimal Set of Registered Panoramic Images Using Denoised Density Maps

  • Giovanni Pintore*
  • , Sara Jashari
  • , Marco Agus
  • , Enrico Gobbetti
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We introduce a deep learning approach to automatically generate 3D floor plans and immersive multi-room virtual visit experiences from a small set of co-registered 360° panoramas - down to just one per room. We integrate novel neural networks that leverage panoramic image broad context and large annotated room datasets to build a geometric and visual graph. Nodes represent stereo-viewable multiple-center-of-projection (MCOP) 360° images at the capture locations, while arcs connect them with paths through doors, avoiding clutter and minimizing disocclusions to maximize visual quality. The process starts with depth prediction and floor-plan projection to create a comprehensive but noisy global density map, which is refined via a latent diffusion model. A segmentation network then extracts room layouts, openings, and clutter. This structured representation is lifted to a visual one by creating a 360° stereo-explorable MCOP representation at each node, produced using a view-synthesis network from the original image and its predicted depth map. Arc paths are then computed using an optimization process that considers structural constraints, including openings and obstacles, while minimizing visual discontinuities, occlusions, and disocclusions. Finally, 360° video transitions are synthesized using a specialized view-synthesis network to obtain a fully precomputed WebXR-ready explorable representation that can be efficiently experienced on Head-Mounted-Displays with limited graphics capabilities. The extracted floor plan not only aids in documenting the captured building but can also enhance immersive experiences by serving as a live map of the building. Our experiments show that the method achieves state-of-the-art reconstruction from sparse inputs and supports compelling immersive visits.

Original languageEnglish
Title of host publicationProceedings - 2025 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025
EditorsUlrich Eck, Gun Lee, Alexander Plopski, Missie Smith, Qi Sun, Markus Tatzgern
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages414-424
Number of pages11
ISBN (Electronic)9798331587611
DOIs
Publication statusPublished - 2025
Event24th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025 - Daejeon, Korea, Republic of
Duration: 8 Oct 202512 Oct 2025

Publication series

NameProceedings - 2025 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025

Conference

Conference24th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025
Country/TerritoryKorea, Republic of
CityDaejeon
Period8/10/2512/10/25

Keywords

  • 360
  • AR/MR/VR for architecture
  • Computer vision
  • Machine learning
  • Omnidirectional
  • immersive view

Fingerprint

Dive into the research topics of 'PanoFloor: Reconstruction and Immersive Exploration of Large Multi-Room Scenes from a Minimal Set of Registered Panoramic Images Using Denoised Density Maps'. Together they form a unique fingerprint.

Cite this