Norod78 commited on
Commit
ca3232d
·
verified ·
1 Parent(s): 7cf7150

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -3
README.md CHANGED
@@ -1,3 +1,46 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ pipeline_tag: depth-estimation
6
+ library_name: coreml
7
+ tags:
8
+ - depth
9
+ - relative depth
10
+ base_model:
11
+ - depth-anything/Depth-Anything-V2-Small
12
+ ---
13
+
14
+ # Depth Anything V2 Small (mlpackage)
15
+
16
+ In this repo you can find:
17
+ * The notebook which was used to convert [depth-anything/Depth-Anything-V2-Small](https://huggingface.co/depth-anything/Depth-Anything-V2-Small) into a CoreML package.
18
+ * Both mlpackage files which can be opened in Xcode and used for Preview and development of macOS and iOS Apps
19
+ * Performence and compute unit mapping report for these models as meassured on an iPhone 16 Pro Max
20
+ * One model uses internal resolution of 518x518 ("Box") and the other 518x392 ("Landscape").
21
+ * The "Landscape" is much faster than "Box" but will also give more "juggy" edges, due to the patch I applied to avoid bicubing upsampling (.diff file is also present in this repo)
22
+
23
+ As a derivative work of Depth-Anything-V2-Small this port is also under apache-2.0
24
+
25
+ ![Xcode Preview](https://huggingface.co/LloydAI/DepthAnything_v2-Small-CoreML/resolve/main/sample_images/Xcode_Preview_DepthAnything_v2_Small_518x392_Landscape.jpg)
26
+
27
+
28
+ ## Citation of original work
29
+
30
+ If you find this project useful, please consider citing:
31
+
32
+ ```bibtex
33
+ @article{depth_anything_v2,
34
+ title={Depth Anything V2},
35
+ author={Yang, Lihe and Kang, Bingyi and Huang, Zilong and Zhao, Zhen and Xu, Xiaogang and Feng, Jiashi and Zhao, Hengshuang},
36
+ journal={arXiv:2406.09414},
37
+ year={2024}
38
+ }
39
+
40
+ @inproceedings{depth_anything_v1,
41
+ title={Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data},
42
+ author={Yang, Lihe and Kang, Bingyi and Huang, Zilong and Xu, Xiaogang and Feng, Jiashi and Zhao, Hengshuang},
43
+ booktitle={CVPR},
44
+ year={2024}
45
+ }
46
+