Spaces:
Runtime error
Runtime error
Commit
·
ef05671
1
Parent(s):
c3d8220
Delete tokenizer
Browse files- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +0 -1
- tokenizer/tokenizer_config.json +0 -1
- tokenizer/vocab.json +0 -0
tokenizer/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer/special_tokens_map.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<|endoftext|>"}
|
|
|
|
tokenizer/tokenizer_config.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"errors": "replace", "max_len": 77, "bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": "<|endoftext|>", "add_prefix_space": false, "do_lower_case": true, "model_max_length": 77, "name_or_path": "/root/.cache/huggingface/diffusers/models--Linaqruf--anything-v3.0/snapshots/07d0cba1b02a5a34255ab1e26bca50392e39940c/tokenizer", "special_tokens_map_file": "/home/aistudio/.cache/paddlenlp/ppdiffusers/Linaqruf/anything-v3.0/tokenizer/special_tokens_map.json", "tokenizer_class": "CLIPTokenizer"}
|
|
|
|
tokenizer/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|