ehristoforu commited on
Commit
840c22c
·
verified ·
1 Parent(s): 7a02b6e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -5
app.py CHANGED
@@ -22,7 +22,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
22
  HF_TOKEN = os.getenv("HF_TOKEN")
23
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
24
 
25
- model_name = "Qwen/Qwen2.5-3B-Instruct"
26
 
27
  model = AutoModelForCausalLM.from_pretrained(
28
  model_name,
@@ -30,9 +30,10 @@ model = AutoModelForCausalLM.from_pretrained(
30
  )
31
  tokenizer = AutoTokenizer.from_pretrained(model_name)
32
 
33
- peft_model = AutoPeftModelForCausalLM.from_pretrained("ehristoforu/CoolQwen2.5-3b-it")
34
- merged_model = peft_model.merge_and_unload()
35
- merged_model.save_pretrained("./coolqwen")
 
36
  tokenizer.save_pretrained("./coolqwen")
37
 
38
  from huggingface_hub import HfApi
@@ -43,7 +44,7 @@ api = HfApi()
43
 
44
  api.upload_folder(
45
  folder_path="./coolqwen",
46
- repo_id="ehristoforu/CoolQwen2.5-3B-IT-fp16",
47
  repo_type="model",
48
  token=HF_TOKEN,
49
  )
 
22
  HF_TOKEN = os.getenv("HF_TOKEN")
23
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
24
 
25
+ model_name = "ehristoforu/cq3bitfp32"
26
 
27
  model = AutoModelForCausalLM.from_pretrained(
28
  model_name,
 
30
  )
31
  tokenizer = AutoTokenizer.from_pretrained(model_name)
32
 
33
+ #peft_model = AutoPeftModelForCausalLM.from_pretrained("ehristoforu/CoolQwen2.5-3b-it")
34
+ #merged_model = peft_model.merge_and_unload()
35
+ #merged_model.save_pretrained("./coolqwen")
36
+ model.save_pretrained("./coolqwen")
37
  tokenizer.save_pretrained("./coolqwen")
38
 
39
  from huggingface_hub import HfApi
 
44
 
45
  api.upload_folder(
46
  folder_path="./coolqwen",
47
+ repo_id="ehristoforu/coolqwen-3b-it",
48
  repo_type="model",
49
  token=HF_TOKEN,
50
  )