ColeGuion commited on
Commit
5e37f93
·
verified ·
1 Parent(s): aee8870

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -1
app.py CHANGED
@@ -52,7 +52,12 @@ def generate(prompt, history, system_prompt, temperature=0.9, max_new_tokens=256
52
  formatted_prompt = format_prompt(f"{system_prompt} {prompt}", history)
53
  #formatted_prompt = format_prompt_grammar(f"Corrected Sentence: {prompt}", history)
54
  print("\nPROMPT: \n\t" + formatted_prompt)
55
-
 
 
 
 
 
56
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
57
  output = ""
58
 
 
52
  formatted_prompt = format_prompt(f"{system_prompt} {prompt}", history)
53
  #formatted_prompt = format_prompt_grammar(f"Corrected Sentence: {prompt}", history)
54
  print("\nPROMPT: \n\t" + formatted_prompt)
55
+
56
+ stream1 = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
57
+ stream2 = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=False, return_full_text=False)
58
+ print(f">> STREAM1 - '{stream1}'")
59
+ print(f">> STREAM2 - '{stream2}'")
60
+ # Generate text from the HF inference
61
  stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
62
  output = ""
63