ImmersiveDepth: A Hybrid Approach for Monocular Depth Estimation from 360 Images Using Tangent Projection and Multi-Model Integration

Sarshar Dorosti, Xiaosong Yang

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

ImmersiveDepth is a hybrid framework designed to tackle challenges in Monocular Depth Estimation (MDE) from 360-degree images, specifically spherical distortions, occlusions, and texture inconsistencies. By integrating tangent image projection, a combination of convolutional neural networks (CNNs) and transformer models, and a novel multi-scale alignment process, ImmersiveDepth achieves seamless and precise depth predictions. Evaluations on diverse datasets show an average 37% reduction in RMSE compared to Depth Anything V2 and a 25% accuracy boost in low-light conditions over MiDaS v3.1. ImmersiveDepth thus establishes a robust solution for immersive technologies, autonomous systems, and 3D reconstruction.
Original languageEnglish
Title of host publication2025 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)
Pages1392-1393
Number of pages2
DOIs
Publication statusPublished (in print/issue) - 8 Mar 2025
Event2025 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW) - The Palais du Grand Large, Saint-Malo, France
Duration: 8 Mar 202512 Mar 2025
https://ieeevr.org/2025/

Conference

Conference2025 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)
Country/TerritoryFrance
CitySaint-Malo
Period8/03/2512/03/25
Internet address

Bibliographical note

Publisher Copyright:
© 2025 IEEE.

Keywords

  • Monocular depth estimation
  • 360-degree images
  • tangent projection
  • VR
  • AR
  • SfM
  • MVS

Fingerprint

Dive into the research topics of 'ImmersiveDepth: A Hybrid Approach for Monocular Depth Estimation from 360 Images Using Tangent Projection and Multi-Model Integration'. Together they form a unique fingerprint.

Cite this