Jofthomas HF staff commited on
Commit
7380361
·
verified ·
1 Parent(s): 6b2710b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -8
app.py CHANGED
@@ -8,6 +8,10 @@ tokenizer = AutoTokenizer.from_pretrained("gpt2")
8
  model = AutoModelForCausalLM.from_pretrained("gpt2")
9
 
10
  print("Loading finished.")
 
 
 
 
11
 
12
  print(f"Is CUDA available: {torch.cuda.is_available()}")
13
  # True
@@ -436,9 +440,6 @@ def get_beam_search_html(
436
  do_sample=False,
437
  )
438
  markdown = "The conclusive sequences are the ones that end in an `<|endoftext|>` token or at the end of generation."
439
- markdown += "\n\nThey are ranked by their scores, as given by the formula `score = cumulative_score / (output_length ** length_penalty)`.\n\n"
440
- markdown += "Only the top `num_beams` scoring sequences are returned: in the tree they are highlighted in **<span style='color:var(--secondary-500)!important'>blue</span>**."
441
- markdown += " The non-selected sequences are also shown in the tree, highlighted in **<span style='color:var(--primary-500)!important'>yellow</span>**."
442
  markdown += "\n#### <span style='color:var(--secondary-500)!important'>Output sequences:</span>"
443
  # Sequences are padded anyway so you can batch decode them
444
  decoded_sequences = tokenizer.batch_decode(outputs.sequences)
@@ -484,16 +485,13 @@ with gr.Blocks(
484
  value="Hugging Face is",
485
  )
486
 
487
- n_steps=12
488
- n_beams=1
489
- length_penalty=1
490
- num_return_sequences=3
491
  button = gr.Button()
492
  out_html = gr.Markdown()
493
  out_markdown = gr.Markdown()
494
  button.click(
495
  get_beam_search_html,
496
- inputs=[text, n_steps, n_beams, length_penalty, num_return_sequences],
497
  outputs=[out_html, out_markdown],
498
  )
499
 
 
8
  model = AutoModelForCausalLM.from_pretrained("gpt2")
9
 
10
  print("Loading finished.")
11
+ n_steps=12
12
+ n_beams=1
13
+ length_penalty=1
14
+ num_return_sequences=3
15
 
16
  print(f"Is CUDA available: {torch.cuda.is_available()}")
17
  # True
 
440
  do_sample=False,
441
  )
442
  markdown = "The conclusive sequences are the ones that end in an `<|endoftext|>` token or at the end of generation."
 
 
 
443
  markdown += "\n#### <span style='color:var(--secondary-500)!important'>Output sequences:</span>"
444
  # Sequences are padded anyway so you can batch decode them
445
  decoded_sequences = tokenizer.batch_decode(outputs.sequences)
 
485
  value="Hugging Face is",
486
  )
487
 
488
+
 
 
 
489
  button = gr.Button()
490
  out_html = gr.Markdown()
491
  out_markdown = gr.Markdown()
492
  button.click(
493
  get_beam_search_html,
494
+ inputs=[text],
495
  outputs=[out_html, out_markdown],
496
  )
497