Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
e497915
1
Parent(s):
d32cb1d
moved both cbow and qwen to cache HF dataset
Browse files- .dockerignore +3 -1
- .gitignore +1 -0
- Dockerfile +11 -3
- cbow_logic.py +1 -1
.dockerignore
CHANGED
@@ -31,4 +31,6 @@ models/
|
|
31 |
|
32 |
# Docker or Space-specific
|
33 |
docker-compose.yaml
|
34 |
-
Caddyfile
|
|
|
|
|
|
31 |
|
32 |
# Docker or Space-specific
|
33 |
docker-compose.yaml
|
34 |
+
Caddyfile
|
35 |
+
qwen_model/
|
36 |
+
models/
|
.gitignore
CHANGED
@@ -2,3 +2,4 @@ __pycache__/
|
|
2 |
*.png
|
3 |
**/artifacts/
|
4 |
models/
|
|
|
|
2 |
*.png
|
3 |
**/artifacts/
|
4 |
models/
|
5 |
+
qwen_model/
|
Dockerfile
CHANGED
@@ -22,13 +22,21 @@ RUN mkdir -p /models/clip && \
|
|
22 |
RUN python3 -c "from transformers import AutoTokenizer; AutoTokenizer.from_pretrained('bert-base-uncased').save_pretrained('/models/bert-tokenizer')"
|
23 |
RUN python3 -c "from transformers import CLIPProcessor; CLIPProcessor.from_pretrained('openai/clip-vit-base-patch32').save_pretrained('/models/clip')"
|
24 |
|
|
|
|
|
25 |
RUN mkdir -p /models/cbow && \
|
26 |
-
|
|
|
27 |
|
|
|
|
|
|
|
28 |
RUN mkdir -p /models/qwen && \
|
29 |
-
python3 -c "from transformers import AutoTokenizer;
|
30 |
-
|
31 |
|
|
|
|
|
32 |
EXPOSE 7860
|
33 |
|
34 |
# Install curl if it's not already installed
|
|
|
22 |
RUN python3 -c "from transformers import AutoTokenizer; AutoTokenizer.from_pretrained('bert-base-uncased').save_pretrained('/models/bert-tokenizer')"
|
23 |
RUN python3 -c "from transformers import CLIPProcessor; CLIPProcessor.from_pretrained('openai/clip-vit-base-patch32').save_pretrained('/models/clip')"
|
24 |
|
25 |
+
# RUN mkdir -p /models/cbow && \
|
26 |
+
# python3 -c "import gensim.downloader as api; model = api.load('glove-twitter-200'); model.save('/models/cbow_model.kv')"
|
27 |
RUN mkdir -p /models/cbow && \
|
28 |
+
wget https://huggingface.co/datasets/ClemSummer/cbow-model-cache/resolve/main/cbow_model.kv -O /models/cbow/cbow_model.kv && \
|
29 |
+
wget https://huggingface.co/datasets/ClemSummer/cbow-model-cache/resolve/main/cbow_model.kv.vectors.npy -O /models/cbow/cbow_model.kv.vectors.npy
|
30 |
|
31 |
+
# RUN mkdir -p /models/qwen && \
|
32 |
+
# python3 -c "from transformers import AutoTokenizer; AutoTokenizer.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
33 |
+
# RUN python3 -c "from transformers import AutoModelForCausalLM; AutoModelForCausalLM.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
34 |
RUN mkdir -p /models/qwen && \
|
35 |
+
python3 -c "from transformers import AutoTokenizer; \
|
36 |
+
AutoTokenizer.from_pretrained('ClemSummer/qwen-model-cache', trust_remote_code=True).save_pretrained('/models/qwen')"
|
37 |
|
38 |
+
RUN python3 -c "from transformers import AutoModelForCausalLM; \
|
39 |
+
AutoModelForCausalLM.from_pretrained('ClemSummer/qwen-model-cache', trust_remote_code=True).save_pretrained('/models/qwen')"
|
40 |
EXPOSE 7860
|
41 |
|
42 |
# Install curl if it's not already installed
|
cbow_logic.py
CHANGED
@@ -7,7 +7,7 @@ import shlex
|
|
7 |
|
8 |
|
9 |
class MeaningCalculator:
|
10 |
-
def __init__(self, model_path: str = "/models/cbow_model.kv"):
|
11 |
if not os.path.exists(model_path):
|
12 |
raise FileNotFoundError(f"Model not found at: {model_path}")
|
13 |
self.model = gensim.models.KeyedVectors.load(model_path, mmap='r')
|
|
|
7 |
|
8 |
|
9 |
class MeaningCalculator:
|
10 |
+
def __init__(self, model_path: str = "/models/cbow/cbow_model.kv"):
|
11 |
if not os.path.exists(model_path):
|
12 |
raise FileNotFoundError(f"Model not found at: {model_path}")
|
13 |
self.model = gensim.models.KeyedVectors.load(model_path, mmap='r')
|