MiDaS v3.1 -- A Model Zoo for Robust Monocular Relative Depth Estimation
Paper
•
2307.14460
•
Published
•
8
DPT 3.1 (MiDaS) models, leveraging state-of-the-art vision backbones such as BEiT and Swinv2
Note We release MiDaS v3.1 for monocular depth estimation, offering a variety of new models based on different encoder backbones. This release is motivated by the success of transformers in computer vision, with a large variety of pretrained vision transformers now available.
Note This model gives the highest quality, but is also the most heavy in terms of computation as mentioned in the paper.
Note This model has moderately less quality, but has a better speed-performance trade-off
Note This model is recommended for deployment on embedded devices