@inproceedings{fa03e8ef9c6b4da6a1180a659f0e71e5,
title = "PanoFloor: Reconstruction and Immersive Exploration of Large Multi-Room Scenes from a Minimal Set of Registered Panoramic Images Using Denoised Density Maps",
abstract = "We introduce a deep learning approach to automatically generate 3D floor plans and immersive multi-room virtual visit experiences from a small set of co-registered 360° panoramas - down to just one per room. We integrate novel neural networks that leverage panoramic image broad context and large annotated room datasets to build a geometric and visual graph. Nodes represent stereo-viewable multiple-center-of-projection (MCOP) 360° images at the capture locations, while arcs connect them with paths through doors, avoiding clutter and minimizing disocclusions to maximize visual quality. The process starts with depth prediction and floor-plan projection to create a comprehensive but noisy global density map, which is refined via a latent diffusion model. A segmentation network then extracts room layouts, openings, and clutter. This structured representation is lifted to a visual one by creating a 360° stereo-explorable MCOP representation at each node, produced using a view-synthesis network from the original image and its predicted depth map. Arc paths are then computed using an optimization process that considers structural constraints, including openings and obstacles, while minimizing visual discontinuities, occlusions, and disocclusions. Finally, 360° video transitions are synthesized using a specialized view-synthesis network to obtain a fully precomputed WebXR-ready explorable representation that can be efficiently experienced on Head-Mounted-Displays with limited graphics capabilities. The extracted floor plan not only aids in documenting the captured building but can also enhance immersive experiences by serving as a live map of the building. Our experiments show that the method achieves state-of-the-art reconstruction from sparse inputs and supports compelling immersive visits.",
keywords = "360, AR/MR/VR for architecture, Computer vision, Machine learning, Omnidirectional, immersive view",
author = "Giovanni Pintore and Sara Jashari and Marco Agus and Enrico Gobbetti",
note = "Publisher Copyright: {\textcopyright} 2025 IEEE.; 24th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025 ; Conference date: 08-10-2025 Through 12-10-2025",
year = "2025",
doi = "10.1109/ISMAR67309.2025.00052",
language = "English",
series = "Proceedings - 2025 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "414--424",
editor = "Ulrich Eck and Gun Lee and Alexander Plopski and Missie Smith and Qi Sun and Markus Tatzgern",
booktitle = "Proceedings - 2025 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2025",
address = "United States",
}