Spaces:
Running
on
Zero
Running
on
Zero
Upload pyproject.toml with huggingface_hub
Browse files- pyproject.toml +34 -28
pyproject.toml
CHANGED
@@ -1,27 +1,40 @@
|
|
1 |
[project]
|
2 |
name = "monopriors"
|
3 |
-
version = "0.1.
|
4 |
description = "Monocular Geometric Priors"
|
5 |
requires-python = ">=3.10.0"
|
6 |
-
|
7 |
-
"Development Status :: 3 - Alpha",
|
8 |
-
"Programming Language :: Python",
|
9 |
-
]
|
10 |
|
11 |
dependencies = [
|
12 |
-
"
|
13 |
-
"
|
14 |
-
"
|
15 |
-
"
|
16 |
-
"
|
17 |
-
"
|
18 |
-
"gradio
|
19 |
-
"gradio",
|
20 |
-
"
|
21 |
-
"
|
22 |
-
|
23 |
-
"
|
24 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
|
26 |
]
|
27 |
|
@@ -46,13 +59,7 @@ scripts = [".pixi.sh"]
|
|
46 |
|
47 |
[tool.pixi.tasks]
|
48 |
_install-wheel = "python -m pip install $WHL_PATH --force-reinstall"
|
49 |
-
|
50 |
-
_build-wheel = { cmd = "python -m pip install build && python -m build", depends_on = [
|
51 |
-
"_delete-wheel",
|
52 |
-
] }
|
53 |
-
_upload-wheel = { cmd = "python -m pip install twine && python -m twine upload -r testpypi dist/*", depends_on = [
|
54 |
-
"_build-wheel",
|
55 |
-
] }
|
56 |
upload-hf = { cmd = "python tools/upload_to_hf.py", depends_on = [
|
57 |
"_build-wheel",
|
58 |
] }
|
@@ -94,9 +101,8 @@ depends_on = ["_install-wheel"]
|
|
94 |
python = "3.10.*"
|
95 |
pip = "24.*"
|
96 |
cuda = { version = "*", channel = "nvidia/label/cuda-11.8.0" }
|
97 |
-
pytorch
|
98 |
-
|
99 |
-
torchvision = { version = "0.18.*", channel = "pytorch" }
|
100 |
icecream = ">=2.1.3,<2.2"
|
101 |
jaxtyping = ">=0.2.28,<0.3"
|
102 |
einops = ">=0.8.0,<0.9"
|
|
|
1 |
[project]
|
2 |
name = "monopriors"
|
3 |
+
version = "0.1.0"
|
4 |
description = "Monocular Geometric Priors"
|
5 |
requires-python = ">=3.10.0"
|
6 |
+
readme = "README.md"
|
|
|
|
|
|
|
7 |
|
8 |
dependencies = [
|
9 |
+
"opencv-python>=4.10.0.82",
|
10 |
+
"geffnet>=1.0.2",
|
11 |
+
"calibur==0.0.1",
|
12 |
+
"omnidata-tools>=0.0.23",
|
13 |
+
"open3d>=0.18.0",
|
14 |
+
"gradio-imageslider>=0.0.20",
|
15 |
+
"gradio==4.36.1",
|
16 |
+
"gradio-rerun==0.0.3",
|
17 |
+
"hf-transfer==0.1.6",
|
18 |
+
"mmcv-lite==2.2.0",
|
19 |
+
# pypi only
|
20 |
+
"torch==2.3.0",
|
21 |
+
"torchvision==0.18.0",
|
22 |
+
"icecream==2.1.3",
|
23 |
+
"jaxtyping",
|
24 |
+
"einops",
|
25 |
+
"matplotlib",
|
26 |
+
"wandb",
|
27 |
+
"scipy",
|
28 |
+
"timm",
|
29 |
+
"pandas",
|
30 |
+
"pyarrow",
|
31 |
+
"pytorch-lightning",
|
32 |
+
"jupyterlab",
|
33 |
+
"rerun-sdk==0.16.1",
|
34 |
+
"numpy<2.0.0",
|
35 |
+
"beartype",
|
36 |
+
"transformers",
|
37 |
+
"xformers>=0.0.26.post1,<0.0.27",
|
38 |
|
39 |
]
|
40 |
|
|
|
59 |
|
60 |
[tool.pixi.tasks]
|
61 |
_install-wheel = "python -m pip install $WHL_PATH --force-reinstall"
|
62 |
+
_build-wheel = "python -m pip install build && python -m build"
|
|
|
|
|
|
|
|
|
|
|
|
|
63 |
upload-hf = { cmd = "python tools/upload_to_hf.py", depends_on = [
|
64 |
"_build-wheel",
|
65 |
] }
|
|
|
101 |
python = "3.10.*"
|
102 |
pip = "24.*"
|
103 |
cuda = { version = "*", channel = "nvidia/label/cuda-11.8.0" }
|
104 |
+
pytorch = { version = "2.3.0", channel = "pytorch" }
|
105 |
+
torchvision = { version = "0.18.0", channel = "pytorch" }
|
|
|
106 |
icecream = ">=2.1.3,<2.2"
|
107 |
jaxtyping = ">=0.2.28,<0.3"
|
108 |
einops = ">=0.8.0,<0.9"
|