Colby commited on
Commit
e9d0c91
·
verified ·
1 Parent(s): 8c53b39

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +10 -8
  2. requirements.txt +2 -0
app.py CHANGED
@@ -17,17 +17,17 @@ import wikipediaapi
17
  wiki_wiki = wikipediaapi.Wikipedia('Organika ([email protected])', 'en')
18
 
19
  ## ctransformers disabled for now
20
- # from ctransformers import AutoModelForCausalLM
21
- # model = AutoModelForCausalLM.from_pretrained(
22
- # "Colby/StarCoder-3B-WoW-JSON",
23
- # model_file="StarCoder-3B-WoW-JSON-ggml.bin",
24
- # model_type="gpt_bigcode"
25
- # )
26
 
27
  # Use a pipeline as a high-level helper
28
  from transformers import pipeline
29
  topic_model = pipeline("zero-shot-classification", model="valhalla/distilbart-mnli-12-9")
30
- model = pipeline("text-generation", model="Colby/StarCoder-3B-WoW-JSON", device=0)
31
 
32
  def merlin_chat(message, history):
33
  chat_text = ""
@@ -78,11 +78,13 @@ def merlin_chat(message, history):
78
  user_msg = {'role': 'user', 'content': message}
79
  prompt = "[" + json.dumps(system_msg) + chat_json + json.dumps(user_msg) + "{'role': 'assistant, 'content': '*recalls \""
80
  for attempt in range(3):
81
- result = model(prompt, max_new_tokens=250, return_full_text=False, handle_long_generation="hole")
 
82
  response = result[0]['generated_text']
83
  start = 0
84
  end = 0
85
  cleanStr = response.lstrip()
 
86
  start = cleanStr.find('{') # this should skip over whatever it recalls to what it says next
87
  if start<=0:
88
  continue
 
17
  wiki_wiki = wikipediaapi.Wikipedia('Organika ([email protected])', 'en')
18
 
19
  ## ctransformers disabled for now
20
+ from ctransformers import AutoModelForCausalLM
21
+ model = AutoModelForCausalLM.from_pretrained(
22
+ "Colby/StarCoder-3B-WoW-JSON",
23
+ model_file="StarCoder-3B-WoW-JSON-ggml.bin",
24
+ model_type="gpt_bigcode"
25
+ )
26
 
27
  # Use a pipeline as a high-level helper
28
  from transformers import pipeline
29
  topic_model = pipeline("zero-shot-classification", model="valhalla/distilbart-mnli-12-9")
30
+ #model = pipeline("text-generation", model="Colby/StarCoder-3B-WoW-JSON", device=0)
31
 
32
  def merlin_chat(message, history):
33
  chat_text = ""
 
78
  user_msg = {'role': 'user', 'content': message}
79
  prompt = "[" + json.dumps(system_msg) + chat_json + json.dumps(user_msg) + "{'role': 'assistant, 'content': '*recalls \""
80
  for attempt in range(3):
81
+ # result = model(prompt, max_new_tokens=250, return_full_text=False, handle_long_generation="hole")
82
+ result = model(prompt, stop=["]"])
83
  response = result[0]['generated_text']
84
  start = 0
85
  end = 0
86
  cleanStr = response.lstrip()
87
+ cleanStr = cleanStr.replace(prompt,"")
88
  start = cleanStr.find('{') # this should skip over whatever it recalls to what it says next
89
  if start<=0:
90
  continue
requirements.txt CHANGED
@@ -1,5 +1,7 @@
 
1
  transformers
2
  torch
3
  spacy
 
4
  rake_nltk
5
  wikipedia-api
 
1
+ ctransformers
2
  transformers
3
  torch
4
  spacy
5
+ nltk
6
  rake_nltk
7
  wikipedia-api