Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
pipeline_tag: depth-estimation
|
6 |
+
tags:
|
7 |
+
- monocular depth estimation
|
8 |
+
- single image depth estimation
|
9 |
+
- depth
|
10 |
+
- in-the-wild
|
11 |
+
- zero-shot
|
12 |
+
- depth
|
13 |
+
---
|
14 |
+
# Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation
|
15 |
+
|
16 |
+
This model represents the official LCM checkpoint of the paper titled "Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation".
|
17 |
+
|
18 |
+
[](https://marigoldmonodepth.github.io)
|
19 |
+
[](https://github.com/prs-eth/Marigold)
|
20 |
+
[](https://arxiv.org/abs/2312.02145)
|
21 |
+
[](https://colab.research.google.com/drive/12G8reD13DdpMie5ZQlaFNo2WCGeNUH-u?usp=sharing)
|
22 |
+
[](https://huggingface.co/spaces/toshas/marigold)
|
23 |
+
[](https://www.apache.org/licenses/LICENSE-2.0)
|
24 |
+
<!-- []() -->
|
25 |
+
<!-- []() -->
|
26 |
+
<!-- []() -->
|
27 |
+
<!-- ### [Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation]() -->
|
28 |
+
|
29 |
+
[Bingxin Ke](http://www.kebingxin.com/),
|
30 |
+
[Anton Obukhov](https://www.obukhov.ai/),
|
31 |
+
[Shengyu Huang](https://shengyuh.github.io/),
|
32 |
+
[Nando Metzger](https://nandometzger.github.io/),
|
33 |
+
[Rodrigo Caye Daudt](https://rcdaudt.github.io/),
|
34 |
+
[Konrad Schindler](https://scholar.google.com/citations?user=FZuNgqIAAAAJ&hl=en )
|
35 |
+
|
36 |
+
We present Marigold, a diffusion model and associated fine-tuning protocol for monocular depth estimation. Its core principle is to leverage the rich visual knowledge stored in modern generative image models. Our model, derived from Stable Diffusion and fine-tuned with synthetic data, can zero-shot transfer to unseen data, offering state-of-the-art monocular depth estimation results.
|
37 |
+
|
38 |
+

|
39 |
+
|
40 |
+
|
41 |
+
## 🎓 Citation
|
42 |
+
|
43 |
+
```bibtex
|
44 |
+
@InProceedings{ke2023repurposing,
|
45 |
+
title={Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation},
|
46 |
+
author={Bingxin Ke and Anton Obukhov and Shengyu Huang and Nando Metzger and Rodrigo Caye Daudt and Konrad Schindler},
|
47 |
+
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
|
48 |
+
year={2024}
|
49 |
+
}
|
50 |
+
```
|
51 |
+
|
52 |
+
## 🎫 License
|
53 |
+
|
54 |
+
This work is licensed under the Apache License, Version 2.0 (as defined in the [LICENSE](LICENSE.txt)).
|
55 |
+
|
56 |
+
By downloading and using the code and model you agree to the terms in the [LICENSE](LICENSE.txt).
|
57 |
+
|
58 |
+
[](https://www.apache.org/licenses/LICENSE-2.0)
|