Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -8,11 +8,14 @@ import torch
|
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
9 |
|
10 |
DESCRIPTION = """\
|
11 |
-
#
|
12 |
|
13 |
-
|
14 |
-
This
|
15 |
-
For
|
|
|
|
|
|
|
16 |
"""
|
17 |
|
18 |
MAX_MAX_NEW_TOKENS = 2048
|
|
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
9 |
|
10 |
DESCRIPTION = """\
|
11 |
+
# EvaByte-SFT
|
12 |
|
13 |
+
EvaByte is a byte-level language model that combines multibyte prediction with the efficient EVA attention mechanism.
|
14 |
+
This page hosts [`EvaByte/EvaByte-SFT`](https://huggingface.co/EvaByte/EvaByte-SFT), fine-tuned via supervised instruction data to enable chat and general instruction-following capabilities.
|
15 |
+
For full details on architecture, training recipe, and benchmarks, see their blog post and the project repository:
|
16 |
+
|
17 |
+
- Blog: <https://hkunlp.github.io/blog/2025/evabyte>
|
18 |
+
- GitHub: <https://github.com/OpenEvaByte/evabyte>
|
19 |
"""
|
20 |
|
21 |
MAX_MAX_NEW_TOKENS = 2048
|