Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,31 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model:
|
3 |
+
- ShubhamSinghCodes/PyNanoLM-big
|
4 |
+
tags:
|
5 |
+
- text-generation-inference
|
6 |
+
- transformers
|
7 |
+
- unsloth
|
8 |
+
- llama
|
9 |
+
- trl
|
10 |
+
- sft
|
11 |
+
- python
|
12 |
+
license: apache-2.0
|
13 |
+
language:
|
14 |
+
- en
|
15 |
+
datasets:
|
16 |
+
- AI-MO/NuminaMath-CoT
|
17 |
+
- TIGER-Lab/MathInstruct
|
18 |
+
- Vezora/Tested-143k-Python-Alpaca
|
19 |
+
- glaiveai/glaive-code-assistant-v2
|
20 |
+
pipeline_tag: text-generation
|
21 |
+
---
|
22 |
+
# Uploaded model
|
23 |
+
|
24 |
+
- **Developed by:** ShubhamSinghCodes
|
25 |
+
- **License:** apache-2.0
|
26 |
+
- **Distilled from model :** ShubhamSinghCodes/PyNanoLM-big
|
27 |
+
|
28 |
+
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
29 |
+
A fast, lite, not entirely stupid model that assists in Python programming. It is a very small, 3.5M parameter model finetuned and distilled from SmolLM. (WIP)
|
30 |
+
|
31 |
+
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|