Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
@@ -22,53 +22,6 @@ HUGGINGFACE_TOKEN = os.getenv("HUGGINGFACE_TOKEN")
|
|
22 |
|
23 |
global_data = {
|
24 |
'models': {},
|
25 |
-
'tokens': {
|
26 |
-
'eos': 'eos_token',
|
27 |
-
'pad': 'pad_token',
|
28 |
-
'padding': 'padding_token',
|
29 |
-
'unk': 'unk_token',
|
30 |
-
'bos': 'bos_token',
|
31 |
-
'sep': 'sep_token',
|
32 |
-
'cls': 'cls_token',
|
33 |
-
'mask': 'mask_token',
|
34 |
-
'n_ctx': 'n_ctx_token',
|
35 |
-
'vocab_size': 'vocab_size_token',
|
36 |
-
'n_embd': 'n_embd_token',
|
37 |
-
'n_head': 'n_head_token',
|
38 |
-
'n_layer': 'n_layer_token',
|
39 |
-
'n_vocab': 'n_vocab_token',
|
40 |
-
'block_size': 'block_size_token',
|
41 |
-
'n_rot': 'n_rot_token',
|
42 |
-
'rope_dim': 'rope_dim_token',
|
43 |
-
'rope_scaling': 'rope_scaling_token',
|
44 |
-
'n_positions': 'n_positions_token',
|
45 |
-
'use_cache': 'use_cache_token',
|
46 |
-
'use_parallel_inference': 'use_parallel_inference_token',
|
47 |
-
'parallel_inference_count': 'parallel_inference_count_token',
|
48 |
-
'use_mlock': 'use_mlock_token',
|
49 |
-
'use_mmap': 'use_mmap_token',
|
50 |
-
'use_cpu': 'use_cpu_token',
|
51 |
-
'f16_kv': 'f16_kv_token',
|
52 |
-
'f16_quant': 'f16_quant_token',
|
53 |
-
'f16_output': 'f16_output_token',
|
54 |
-
'use_flash_attn': 'use_flash_attn_token',
|
55 |
-
'max_seq_len': 'max_seq_len_token',
|
56 |
-
'do_sample': 'do_sample_token',
|
57 |
-
'top_k': 'top_k_token',
|
58 |
-
'top_p': 'top_p_token',
|
59 |
-
'temperature': 'temperature_token',
|
60 |
-
'num_return_sequences': 'num_return_sequences_token',
|
61 |
-
'use_repetition_penalty': 'use_repetition_penalty_token',
|
62 |
-
'repetition_penalty': 'repetition_penalty_token',
|
63 |
-
'no_repeat_ngram_size': 'no_repeat_ngram_size_token',
|
64 |
-
'bad_words_ids': 'bad_words_ids_token',
|
65 |
-
'use_token_logging': 'use_token_logging_token',
|
66 |
-
'use_tensor_parallel': 'use_tensor_parallel_token',
|
67 |
-
'tensor_parallel_size': 'tensor_parallel_size_token',
|
68 |
-
'use_gpu_memory_growth': 'use_gpu_memory_growth_token',
|
69 |
-
'use_multi_gpu_inference': 'use_multi_gpu_inference_token',
|
70 |
-
'multi_gpu_inference_count': 'multi_gpu_inference_count_token'
|
71 |
-
}
|
72 |
}
|
73 |
|
74 |
model_configs = [
|
@@ -381,4 +334,4 @@ def anonymize_ip():
|
|
381 |
Thread(target=anonymize_ip).start()
|
382 |
|
383 |
if __name__ == "__main__":
|
384 |
-
iface.launch(share=True)
|
|
|
22 |
|
23 |
global_data = {
|
24 |
'models': {},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
}
|
26 |
|
27 |
model_configs = [
|
|
|
334 |
Thread(target=anonymize_ip).start()
|
335 |
|
336 |
if __name__ == "__main__":
|
337 |
+
iface.launch(share=True)
|