000152428 001__ 152428 000152428 005__ 20250401114421.0 000152428 037__ $$aTAZ-TFM-2024-811 000152428 041__ $$aeng 000152428 1001_ $$aLasheras Hernández, Blanca 000152428 24200 $$aSingle-view depth from focused plenoptic cameras 000152428 24500 $$aProfundidad monocular con cámara plenóptica enfocada 000152428 260__ $$aZaragoza$$bUniversidad de Zaragoza$$c2024 000152428 506__ $$aby-nc-sa$$bCreative Commons$$c3.0$$uhttp://creativecommons.org/licenses/by-nc-sa/3.0/ 000152428 520__ $$aIn recent years, the research progress in computer vision has boosted the capabilities of machines for interpreting visual data, thereby expanding the complexity and range of tasks that robots could perform in fields such as autonomous driving, medicine, and industrial automation. A principal facet of computer vision is depth estimation, crucial for enabling robots to perceive, navigate, and interact with their environment in an effective and safe manner. Traditional setups, like stereo or multi-camera, face challenges such as calibration intricacies and computational and hardware complexity. Further, their accuracy is limited by the baseline between the cameras. Monocular depth estimation, thus using a single camera, offers a more compact alternative but is however limited by the unobservability of the scale. Light field imaging technologies represent a promising solution to the above issues by capturing both the intensity and direction of light rays not only through the main lens, but also through a large number of microlenses placed within the camera. By these means, depth in front of the camera can be measured owing to depth-dependent refraction at the main lens. Despite their potential, there are limited studies exploring their application to single-view dense depth estimation. This scarcity can be attributed to several factors. The technology remains relatively costly and inaccessible for its widespread adoption, leading to a lack of datasets suitable for training deep neural networks. As a consequence, few projects have used light field imaging for depth estimation, and existing efforts often rely on outdated iterations of the technology. Furthermore, the lack of an open-source geometrical model impedes the development of model-based estimation This thesis explores the potential of focused plenoptic cameras for single-view depth estimation using learning-based methods. The proposed approach integrates techniques from image processing, deep learning, and scale alignment achieved through foundational models and robust statistics, to generate dense metric depth maps. To support this approach, a novel real-world dataset of light field images with stereo depth labels was generated, addressing a current gap in existing resources. Experimental results demonstrate that the developed pipeline can reliably produce accurate metric depth predictions, setting a foundation for further research in this domain.<br /> 000152428 521__ $$aMáster Universitario en Robótica, Gráficos y Visión por Computador 000152428 540__ $$aDerechos regulados por licencia Creative Commons 000152428 691__ $$a7 9 000152428 692__ $$a7.1, 7.a, 7.b, 9.1, 9.5 000152428 700__ $$aCivera Sancho, Javier$$edir. 000152428 700__ $$aStrobl, Klaus H.$$edir. 000152428 7102_ $$aUniversidad de Zaragoza$$bInformática e Ingeniería de Sistemas$$cIngeniería de Sistemas y Automática 000152428 8560_ $$f736056@unizar.es 000152428 8564_ $$s14830216$$uhttps://zaguan.unizar.es/record/152428/files/TAZ-TFM-2024-811.pdf$$yMemoria (eng) 000152428 909CO $$ooai:zaguan.unizar.es:152428$$pdriver$$ptrabajos-fin-master 000152428 950__ $$a 000152428 951__ $$adeposita:2025-04-01 000152428 980__ $$aTAZ$$bTFM$$cEINA 000152428 999__ $$a20240626132728.CREATION_DATE