Update README.md
Browse files
README.md
CHANGED
@@ -76,4 +76,13 @@ model-index:
|
|
76 |
- type: accuracy
|
77 |
name: QA Average
|
78 |
value: 0.3780991480835666
|
79 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
76 |
- type: accuracy
|
77 |
name: QA Average
|
78 |
value: 0.3780991480835666
|
79 |
+
---
|
80 |
+
|
81 |
+
# TinyLlama_v1.1_1bit_BitDistiller
|
82 |
+
This is a 1-bit quantized version of TinyLlama v1.1, trained using BitDistiller with asymmetric quantization and self-distillation (CAKLD) to optimize accuracy retention under extreme compression. The model is fine-tuned on WikiText-2 and Alpaca-cleaned datasets and evaluated on multiple-choice QA benchmarks.
|
83 |
+
|
84 |
+
Key Features:
|
85 |
+
- 1-bit quantization for ultra-efficient inference.
|
86 |
+
- Asymmetric weight clipping to reduce precision loss.
|
87 |
+
- CAKLD knowledge distillation to preserve performance.
|
88 |
+
- Tested on ARC-Challenge, HellaSwag, PIQA, and Winogrande.
|