Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: apache-2.0
|
|
5 |
|
6 |
Reka Flash 3 is our latest general-purpose reasoning model that excels in general chat, coding, instruction following, and function calling. At a size of 21B parameters, it performs competitively with proprietary models such as OpenAI o1-mini, making it a good foundation to build many applications that require low latency or on-device deployments. It is currently the best model in its size category.
|
7 |
|
8 |
-
.
|
11 |
|
@@ -60,15 +60,7 @@ If you are using HF or vLLM, the built-in chat_template shall handle prompt form
|
|
60 |
|
61 |
Reka Flash thinks before it produces an output. We use <reasoning> </reasoning> tags to indicate the beginning and the end of its thinking process. For some problems, the model might think for a long time. You can make the model to stop its thinking process by forcing it to output </reasoning> after a certain number of steps. We observe such a budget forcing mechanism will still produce a reasonable output. We show performance on AIME-2024 (cons@16) for various budgets below.
|
62 |
|
63 |
-
|
64 |
-
|--------|-------|
|
65 |
-
| 4k | 40 |
|
66 |
-
| 8k | 46 |
|
67 |
-
| 12k | 50 |
|
68 |
-
| 16k | 56 |
|
69 |
-
| 24k | 60 |
|
70 |
-
| 32k | 60 |
|
71 |
-
| 48k | 63.3 |
|
72 |
|
73 |
|
74 |
### Language Support
|
|
|
5 |
|
6 |
Reka Flash 3 is our latest general-purpose reasoning model that excels in general chat, coding, instruction following, and function calling. At a size of 21B parameters, it performs competitively with proprietary models such as OpenAI o1-mini, making it a good foundation to build many applications that require low latency or on-device deployments. It is currently the best model in its size category.
|
7 |
|
8 |
+

|
9 |
|
10 |
Try it out at [Reka Space](https://space.reka.ai).
|
11 |
|
|
|
60 |
|
61 |
Reka Flash thinks before it produces an output. We use <reasoning> </reasoning> tags to indicate the beginning and the end of its thinking process. For some problems, the model might think for a long time. You can make the model to stop its thinking process by forcing it to output </reasoning> after a certain number of steps. We observe such a budget forcing mechanism will still produce a reasonable output. We show performance on AIME-2024 (cons@16) for various budgets below.
|
62 |
|
63 |
+

|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
|
65 |
|
66 |
### Language Support
|