jonACE commited on
Commit
fb7c227
·
verified ·
1 Parent(s): 2675666

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -2
app.py CHANGED
@@ -56,8 +56,10 @@ from peft import LoraConfig, get_peft_model
56
  # Load LLaMA 2 model in 4-bit mode to save memory
57
  model = AutoModelForCausalLM.from_pretrained(
58
  model_name,
59
- load_in_4bit=True, # Use 4-bit quantization for efficiency
60
- device_map="auto"
 
 
61
  )
62
 
63
  # Apply LoRA (efficient fine-tuning)
 
56
  # Load LLaMA 2 model in 4-bit mode to save memory
57
  model = AutoModelForCausalLM.from_pretrained(
58
  model_name,
59
+ #load_in_4bit=True, # Use 4-bit quantization for efficiency
60
+ #device_map="auto"
61
+ device_map="cpu",
62
+ quantization_config=None
63
  )
64
 
65
  # Apply LoRA (efficient fine-tuning)