Update joycaption.py
Browse files- joycaption.py +7 -7
joycaption.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
import os
|
2 |
-
if os.environ.get("SPACES_ZERO_GPU") is not None:
|
3 |
-
import spaces
|
4 |
else:
|
5 |
class spaces:
|
6 |
@staticmethod
|
@@ -23,10 +23,10 @@ from typing import Union
|
|
23 |
|
24 |
LOAD_IN_NF4 = True
|
25 |
|
26 |
-
if os.environ.get("SPACES_ZERO_GPU") is not None:
|
27 |
-
import subprocess
|
28 |
-
LOAD_IN_NF4 = False # If true, Custom VLM LoRA doesn't work initially. The rest are fine.
|
29 |
-
subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
30 |
|
31 |
BASE_DIR = Path(__file__).resolve().parent # Define the base directory
|
32 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
@@ -474,7 +474,7 @@ def get_repo_gguf(repo_id: str):
|
|
474 |
else: return gr.update(value=files[0], choices=files)
|
475 |
|
476 |
|
477 |
-
|
478 |
def change_text_model(model_name: str=MODEL_PATH, use_client: bool=False, gguf_file: Union[str, None]=None,
|
479 |
is_nf4: bool=True, is_lora: bool=True, progress=gr.Progress(track_tqdm=True)):
|
480 |
global use_inference_client, llm_models
|
|
|
1 |
import os
|
2 |
+
#if os.environ.get("SPACES_ZERO_GPU") is not None:
|
3 |
+
# import spaces
|
4 |
else:
|
5 |
class spaces:
|
6 |
@staticmethod
|
|
|
23 |
|
24 |
LOAD_IN_NF4 = True
|
25 |
|
26 |
+
#if os.environ.get("SPACES_ZERO_GPU") is not None:
|
27 |
+
# import subprocess
|
28 |
+
# LOAD_IN_NF4 = False # If true, Custom VLM LoRA doesn't work initially. The rest are fine.
|
29 |
+
# subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
30 |
|
31 |
BASE_DIR = Path(__file__).resolve().parent # Define the base directory
|
32 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
|
474 |
else: return gr.update(value=files[0], choices=files)
|
475 |
|
476 |
|
477 |
+
#@spaces.GPU
|
478 |
def change_text_model(model_name: str=MODEL_PATH, use_client: bool=False, gguf_file: Union[str, None]=None,
|
479 |
is_nf4: bool=True, is_lora: bool=True, progress=gr.Progress(track_tqdm=True)):
|
480 |
global use_inference_client, llm_models
|