Update README.md
Browse files
README.md
CHANGED
|
@@ -5,7 +5,7 @@ license: apache-2.0
|
|
| 5 |
|
| 6 |
Reka Flash 3 is our latest general-purpose reasoning model that excels in general chat, coding, instruction following, and function calling. At a size of 21B parameters, it performs competitively with proprietary models such as OpenAI o1-mini, making it a good foundation to build many applications that require low latency or on-device deployments. It is currently the best model in its size category.
|
| 7 |
|
| 8 |
-
.
|
| 11 |
|
|
@@ -60,15 +60,7 @@ If you are using HF or vLLM, the built-in chat_template shall handle prompt form
|
|
| 60 |
|
| 61 |
Reka Flash thinks before it produces an output. We use <reasoning> </reasoning> tags to indicate the beginning and the end of its thinking process. For some problems, the model might think for a long time. You can make the model to stop its thinking process by forcing it to output </reasoning> after a certain number of steps. We observe such a budget forcing mechanism will still produce a reasonable output. We show performance on AIME-2024 (cons@16) for various budgets below.
|
| 62 |
|
| 63 |
-
|
| 64 |
-
|--------|-------|
|
| 65 |
-
| 4k | 40 |
|
| 66 |
-
| 8k | 46 |
|
| 67 |
-
| 12k | 50 |
|
| 68 |
-
| 16k | 56 |
|
| 69 |
-
| 24k | 60 |
|
| 70 |
-
| 32k | 60 |
|
| 71 |
-
| 48k | 63.3 |
|
| 72 |
|
| 73 |
|
| 74 |
### Language Support
|
|
|
|
| 5 |
|
| 6 |
Reka Flash 3 is our latest general-purpose reasoning model that excels in general chat, coding, instruction following, and function calling. At a size of 21B parameters, it performs competitively with proprietary models such as OpenAI o1-mini, making it a good foundation to build many applications that require low latency or on-device deployments. It is currently the best model in its size category.
|
| 7 |
|
| 8 |
+

|
| 9 |
|
| 10 |
Try it out at [Reka Space](https://space.reka.ai).
|
| 11 |
|
|
|
|
| 60 |
|
| 61 |
Reka Flash thinks before it produces an output. We use <reasoning> </reasoning> tags to indicate the beginning and the end of its thinking process. For some problems, the model might think for a long time. You can make the model to stop its thinking process by forcing it to output </reasoning> after a certain number of steps. We observe such a budget forcing mechanism will still produce a reasonable output. We show performance on AIME-2024 (cons@16) for various budgets below.
|
| 62 |
|
| 63 |
+

|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
|
| 65 |
|
| 66 |
### Language Support
|