Update README.md
Browse files
README.md
CHANGED
@@ -48,8 +48,6 @@ NexaQuant on Reasoning Benchmarks Compared to BF16 and LMStudio's Q4_K_M
|
|
48 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/66abfd6f65beb23afa427d8a/Cyh1zVvDHNBT598IkLHkd.png" width="80%" alt="Example" />
|
49 |
</div>
|
50 |
|
51 |
-
The general capacity has also greatly improved:
|
52 |
-
|
53 |
**General Capacity:**
|
54 |
|
55 |
| Benchmark | Full 16-bit | llama.cpp (4-bit) | NexaQuant (4-bit)|
|
@@ -114,9 +112,9 @@ Get the latest version from the [official website](https://lmstudio.ai/).
|
|
114 |
|
115 |
## What's next
|
116 |
|
117 |
-
1.
|
118 |
|
119 |
-
2.
|
120 |
|
121 |
### Follow us
|
122 |
|
|
|
48 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/66abfd6f65beb23afa427d8a/Cyh1zVvDHNBT598IkLHkd.png" width="80%" alt="Example" />
|
49 |
</div>
|
50 |
|
|
|
|
|
51 |
**General Capacity:**
|
52 |
|
53 |
| Benchmark | Full 16-bit | llama.cpp (4-bit) | NexaQuant (4-bit)|
|
|
|
112 |
|
113 |
## What's next
|
114 |
|
115 |
+
1. This model is built for complex problem-solving, which is why it sometimes takes a long thinking process even for simple questions. We recognize this and are working on improving it in the next update.
|
116 |
|
117 |
+
2. Inference Nexa Quantized Deepseek-R1 distilled model on NPU.
|
118 |
|
119 |
### Follow us
|
120 |
|