Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -28,6 +28,9 @@ from transformers import CLIPTextModelWithProjection, CLIPTextModel
|
|
28 |
from transformers import Blip2Processor, Blip2ForConditionalGeneration
|
29 |
from transformers import Phi3ForCausalLM
|
30 |
from transformers import pipeline
|
|
|
|
|
|
|
31 |
|
32 |
from transformers import InstructBlipProcessor, InstructBlipForConditionalGeneration
|
33 |
torch.backends.cuda.matmul.allow_tf32 = False
|
@@ -38,8 +41,8 @@ torch.backends.cudnn.deterministic = False
|
|
38 |
torch.backends.cudnn.benchmark = False
|
39 |
#torch.backends.cuda.preferred_blas_library="cublas"
|
40 |
# torch.backends.cuda.preferred_linalg_library="cusolver"
|
41 |
-
|
42 |
torch.set_float32_matmul_precision("highest")
|
|
|
43 |
os.putenv("HF_HUB_ENABLE_HF_TRANSFER","1")
|
44 |
os.environ["SAFETENSORS_FAST_GPU"] = "1"
|
45 |
|
|
|
28 |
from transformers import Blip2Processor, Blip2ForConditionalGeneration
|
29 |
from transformers import Phi3ForCausalLM
|
30 |
from transformers import pipeline
|
31 |
+
import torchvision
|
32 |
+
|
33 |
+
torchvision.set_image_backend('accimage')
|
34 |
|
35 |
from transformers import InstructBlipProcessor, InstructBlipForConditionalGeneration
|
36 |
torch.backends.cuda.matmul.allow_tf32 = False
|
|
|
41 |
torch.backends.cudnn.benchmark = False
|
42 |
#torch.backends.cuda.preferred_blas_library="cublas"
|
43 |
# torch.backends.cuda.preferred_linalg_library="cusolver"
|
|
|
44 |
torch.set_float32_matmul_precision("highest")
|
45 |
+
|
46 |
os.putenv("HF_HUB_ENABLE_HF_TRANSFER","1")
|
47 |
os.environ["SAFETENSORS_FAST_GPU"] = "1"
|
48 |
|