AhmedSSabir commited on
Commit
e8e5e08
·
verified ·
1 Parent(s): 5137275

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -2
app.py CHANGED
@@ -14,6 +14,7 @@ import numpy as np
14
  from transformers import AutoTokenizer, AutoModelForCausalLM
15
  #from torch.nn.functional import softmax
16
 
 
17
  # just for the sake of this demo, we use cloze prob to initialize the hypothesis
18
 
19
  #url = "https://github.com/simonepri/lm-scorer/tree/master/lm_scorer/models"
@@ -53,8 +54,14 @@ def get_sim(x):
53
  #tokenizer = GPT2Tokenizer.from_pretrained('distilgpt2')
54
 
55
 
56
- tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-1B")
57
- model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B")
 
 
 
 
 
 
58
 
59
 
60
  #tokenizer = GPT2Tokenizer.from_pretrained('gpt2')
 
14
  from transformers import AutoTokenizer, AutoModelForCausalLM
15
  #from torch.nn.functional import softmax
16
 
17
+
18
  # just for the sake of this demo, we use cloze prob to initialize the hypothesis
19
 
20
  #url = "https://github.com/simonepri/lm-scorer/tree/master/lm_scorer/models"
 
54
  #tokenizer = GPT2Tokenizer.from_pretrained('distilgpt2')
55
 
56
 
57
+
58
+
59
+
60
+ tokenizer = AutoTokenizer.from_pretrained("prithivMLmods/Llama-3.2-1B-GGUF")
61
+ model = AutoModelForCausalLM.from_pretrained("prithivMLmods/Llama-3.2-1B-GGUF")
62
+
63
+ # tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-1B")
64
+ # model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-1B")
65
 
66
 
67
  #tokenizer = GPT2Tokenizer.from_pretrained('gpt2')