Update README.md
Browse files
README.md
CHANGED
|
@@ -270,8 +270,11 @@ Fig 5. Performance comparision of Llama-3-8B, Llama-2-13B, Yi-1.5-6B and Llama-3
|
|
| 270 |
|
| 271 |
Llama-3-6B is competive with model within it's category and upto 2x larger than it self across 6 diverse benchmarks.
|
| 272 |
|
| 273 |
-
#### Summary
|
| 274 |
|
|
|
|
|
|
|
|
|
|
| 275 |
|
| 276 |
|
| 277 |
## Citation
|
|
@@ -288,6 +291,17 @@ Llama-3-6B is competive with model within it's category and upto 2x larger than
|
|
| 288 |
}
|
| 289 |
```
|
| 290 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 291 |
|
| 292 |
## References:
|
| 293 |
|
|
|
|
| 270 |
|
| 271 |
Llama-3-6B is competive with model within it's category and upto 2x larger than it self across 6 diverse benchmarks.
|
| 272 |
|
| 273 |
+
#### Summary and future directions:
|
| 274 |
|
| 275 |
+
This experiment was a success! Using this technique, I'll be able to build many variants. This is the first of many new base models I intend to create.
|
| 276 |
+
|
| 277 |
+
Next, I plan to explore different data mixtures and perform full fine-tuning, all of which will contribute to developing other small model as well as larger and more robust models.
|
| 278 |
|
| 279 |
|
| 280 |
## Citation
|
|
|
|
| 291 |
}
|
| 292 |
```
|
| 293 |
|
| 294 |
+
# **Thank You!**
|
| 295 |
+
|
| 296 |
+
I want to extend my heartfelt thanks to the community for their invaluable expertise and unwavering support.
|
| 297 |
+
|
| 298 |
+
This is my most ambitious project yet, and it wouldn't have been possible without the incredible open-source ML community!
|
| 299 |
+
|
| 300 |
+
Developers, I am eager to see and hear about the innovative fine-tunes and applications you create.
|
| 301 |
+
|
| 302 |
+
Users, I am excited to learn about your experiences and use cases.
|
| 303 |
+
|
| 304 |
+
Thank you for your interest and support!
|
| 305 |
|
| 306 |
## References:
|
| 307 |
|