Spaces:
Runtime error
Runtime error
Commit
·
60d5b9f
1
Parent(s):
0a88363
added bloom
Browse files- demo_watermark.py +2 -1
demo_watermark.py
CHANGED
@@ -34,6 +34,7 @@ from watermark_processor import WatermarkLogitsProcessor, WatermarkDetector
|
|
34 |
|
35 |
# FIXME correct lengths for all models
|
36 |
API_MODEL_MAP = {
|
|
|
37 |
"bigscience/bloomz" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
38 |
"google/flan-ul2" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
39 |
"google/flan-t5-xxl" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
@@ -461,7 +462,7 @@ def run_gradio(args, model=None, device=None, tokenizer=None):
|
|
461 |
Longer prompts that end mid-sentence will result in more fluent generations.
|
462 |
|
463 |
The larger models available in this demo are fine-tuned to follow instructions but have different strengths and will showcase different
|
464 |
-
types of watermark behavior. [BLOOMZ
|
465 |
and can generate long and coherent paragraphs and stories given the right prompt.
|
466 |
The FLAN models [FLAN-t5-xxl (11B)](https://huggingface.co/google/flan-t5-xxl) and [FLAN-UL2 (20B)](https://huggingface.co/google/flan-ul2) are fine-tuned on a variety of in-context few-shot learning NLP tasks,
|
467 |
such as reasoning, and question answering.
|
|
|
34 |
|
35 |
# FIXME correct lengths for all models
|
36 |
API_MODEL_MAP = {
|
37 |
+
"bigscience/bloom" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
38 |
"bigscience/bloomz" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
39 |
"google/flan-ul2" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
40 |
"google/flan-t5-xxl" : {"max_length": 2048, "gamma": 0.5, "delta": 2.0},
|
|
|
462 |
Longer prompts that end mid-sentence will result in more fluent generations.
|
463 |
|
464 |
The larger models available in this demo are fine-tuned to follow instructions but have different strengths and will showcase different
|
465 |
+
types of watermark behavior. [BLOOMZ](https://huggingface.co/bigscience/bloomz) is an instruction tuned variant of [BLOOM (175B)](https://huggingface.co/bigscience/bloom) capable of following instructions in dozens of languages zero-shot
|
466 |
and can generate long and coherent paragraphs and stories given the right prompt.
|
467 |
The FLAN models [FLAN-t5-xxl (11B)](https://huggingface.co/google/flan-t5-xxl) and [FLAN-UL2 (20B)](https://huggingface.co/google/flan-ul2) are fine-tuned on a variety of in-context few-shot learning NLP tasks,
|
468 |
such as reasoning, and question answering.
|