Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,40 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- Crystalcareai/openhermes_200k_unfiltered
|
5 |
+
- mlabonne/orpo-dpo-mix-40k
|
6 |
+
- jondurbin/airoboros-3.2
|
7 |
+
- abacusai/SystemChat-1.1
|
8 |
+
- trollek/SimpleInstructionJudge-v01
|
9 |
+
- cgato/SlimOrcaDedupCleaned
|
10 |
+
language:
|
11 |
+
- en
|
12 |
+
library_name: transformers
|
13 |
+
base_model: h2oai/h2o-danube3-4b-base
|
14 |
+
tags:
|
15 |
+
- mergekit
|
16 |
+
- magpie
|
17 |
---
|
18 |
+
# LittlePromptMaker-4B-v0.1
|
19 |
+
|
20 |
+
A small model to create prompts the [Magpie](https://arxiv.org/abs/2406.08464) way.
|
21 |
+
|
22 |
+
The secret sauce turned out to be also training on the prompts. I did that last with SystemChat-1.1 in order to be able to steer the prompt generation. It does not work without a system message.
|
23 |
+
|
24 |
+
Now imagine, if you will, having this bad boy generating a bunch of different prompts right, and having another model like, I mean.. [LittleInstructionJudge](https://huggingface.co/trollek/LittleInstructionJudge-4B-v0.1) right, judge all of the instructions right, and then slam a serverfarm with the cream of the crop right.
|
25 |
+
|
26 |
+
In other words, giving it a system prompt like "You are a creative writing partner", "You are an advanced coding assistant", "You are a damn good psychologist", etc, you can can quickly generate prompts for a niche dataset that can then be answered by large model.
|
27 |
+
|
28 |
+
In a different language: Ved hjælp at Husskades indsigt, hvor man udnytter sprogmodellers natur til at skabe tilpasningsdata, kan man med fordel bruge denne sprogmodel til at skrive instruktioner, og endda styre indholdet ved hjælp at system beskeden.
|
29 |
+
|
30 |
+
## Training
|
31 |
+
|
32 |
+
All the datasets were used seperately and merged together using [Model Stock](https://arxiv.org/abs/2403.19522), except for SystemChat-1.1 where I fine-tuned it using [LoRA+](https://arxiv.org/abs/2402.12354) with `train_on_prompt` set to True.
|
33 |
+
|
34 |
+
## Using
|
35 |
+
|
36 |
+
```jinja2
|
37 |
+
<|im_start|>system
|
38 |
+
{{system_message}}<|im_end|>
|
39 |
+
<|im_start|>user
|
40 |
+
```
|