jonACE commited on
Commit
8e63da8
·
verified ·
1 Parent(s): 66ee08b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -59,10 +59,10 @@ from peft import LoraConfig, get_peft_model
59
  # Load LLaMA 2 model in 4-bit mode to save memory
60
  model = AutoModelForCausalLM.from_pretrained(
61
  model_name,
62
- load_in_4bit=True, # Use 4-bit quantization for efficiency
63
- device_map="auto"
64
- #device_map="cpu",
65
- #quantization_config=None
66
  )
67
 
68
  # Apply LoRA (efficient fine-tuning)
 
59
  # Load LLaMA 2 model in 4-bit mode to save memory
60
  model = AutoModelForCausalLM.from_pretrained(
61
  model_name,
62
+ #load_in_4bit=True, # Use 4-bit quantization for efficiency
63
+ #device_map="auto"
64
+ device_map="cpu",
65
+ quantization_config=None
66
  )
67
 
68
  # Apply LoRA (efficient fine-tuning)