shleeeee commited on
Commit
5e22bcb
·
1 Parent(s): 8f9af74

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -0
README.md ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ko
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - finetune
7
+ ---
8
+ # Model Card for mistral-ko-7b-wiki-neft
9
+ It is a fine-tuned model using Korean and NEFT in the mistral-7b model.
10
+
11
+ ## Model Details
12
+
13
+ * **Model Developers** : shleeeee(Seunghyeon Lee)
14
+ * **Repository** : To be added
15
+ * **Model Architecture** : The mistral-ko-7b-wiki-neft is is a fine-tuned version of the Mistral-7B-v0.1.
16
+ * **Lora target modules** : q_proj, k_proj, v_proj, o_proj,gate_proj
17
+ * **train_batch** : 4
18
+ * **neftune_noise_alpha** : 5
19
+ * **Max_step** : 1000
20
+
21
+ ## Dataset
22
+ Korean Custom Dataset
23
+
24
+ ## Prompt template: Mistral
25
+ ```
26
+ <s>[INST]{['instruction']}[/INST]{['output']}</s>
27
+ ```
28
+
29
+ ## Usage
30
+ ```
31
+ # Load model directly
32
+ from transformers import AutoTokenizer, AutoModelForCausalLM
33
+
34
+ tokenizer = AutoTokenizer.from_pretrained("shleeeee/mistral-7b-wiki")
35
+ model = AutoModelForCausalLM.from_pretrained("shleeeee/mistral-7b-wiki")
36
+
37
+ # Use a pipeline as a high-level helper
38
+ from transformers import pipeline
39
+
40
+ pipe = pipeline("text-generation", model="shleeeee/mistral-7b-wiki")
41
+ ```
42
+
43
+ ## Evaluation
44
+ - To be added