ClemSummer commited on
Commit
a6b901d
·
1 Parent(s): 88b5781

HF needs all model downloads to a special read-write cache dir, also affects encoder.py

Browse files
Files changed (1) hide show
  1. vit_captioning/models/encoder.py +7 -2
vit_captioning/models/encoder.py CHANGED
@@ -12,7 +12,10 @@ class ViTEncoder(nn.Module):
12
  super(ViTEncoder, self).__init__()
13
 
14
  #weights = ViT_B_16_Weights.DEFAULT
15
- self.vit = ViTModel.from_pretrained('google/vit-base-patch16-224-in21k')
 
 
 
16
 
17
  def forward(self, pixel_values):
18
 
@@ -30,7 +33,9 @@ from transformers import CLIPModel
30
  class CLIPEncoder(nn.Module):
31
  def __init__(self):
32
  super(CLIPEncoder, self).__init__()
33
- self.clip = CLIPModel.from_pretrained("openai/clip-vit-base-patch32")
 
 
34
 
35
  def forward(self, pixel_values):
36
  # ✅ Directly get the pooled image features (already the final representation)
 
12
  super(ViTEncoder, self).__init__()
13
 
14
  #weights = ViT_B_16_Weights.DEFAULT
15
+
16
+ #self.vit = ViTModel.from_pretrained('google/vit-base-patch16-224-in21k')
17
+ #HF needs all model downloads to a special read-write cache dir
18
+ self.vit = ViTModel.from_pretrained('google/vit-base-patch16-224-in21k', cache_dir="/tmp")
19
 
20
  def forward(self, pixel_values):
21
 
 
33
  class CLIPEncoder(nn.Module):
34
  def __init__(self):
35
  super(CLIPEncoder, self).__init__()
36
+ #self.clip = CLIPModel.from_pretrained("openai/clip-vit-base-patch32")
37
+ #HF needs all model downloads to a special read-write cache dir
38
+ self.clip = CLIPModel.from_pretrained("openai/clip-vit-base-patch32", cache_dir="/tmp")
39
 
40
  def forward(self, pixel_values):
41
  # ✅ Directly get the pooled image features (already the final representation)