utkmst commited on
Commit
5ba6652
·
verified ·
1 Parent(s): 3218d26

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +101 -4
README.md CHANGED
@@ -11,20 +11,117 @@ datasets:
11
  - OpenAssistant/oasst1
12
  - databricks/databricks-dolly-15k
13
  - Open-Orca/OpenOrca
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
 
16
-
17
  # utkmst/chimera-beta-test2-lora-merged
18
 
19
  ## Model Description
20
  This model is a fine-tuned version of Meta's Llama-3.1-8B-Instruct model, created through LoRA fine-tuning on multiple instruction datasets, followed by merging the adapter weights with the base model.
21
 
22
- ## Training Details
23
  - **Base Model**: meta-llama/Llama-3.1-8B-Instruct
 
 
 
 
 
24
  - **Training Method**: LoRA fine-tuning followed by adapter merging
25
- - **Datasets Used**: OpenAssistant/oasst1, databricks/databricks-dolly-15k, Open-Orca/OpenOrca, and others
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
- (You can add more details from your original card here if desired)
 
 
 
 
 
28
 
29
  ## Usage with Transformers
30
 
 
11
  - OpenAssistant/oasst1
12
  - databricks/databricks-dolly-15k
13
  - Open-Orca/OpenOrca
14
+ - mlabonne/open-perfectblend
15
+ - tatsu-lab/alpaca
16
+ model-index:
17
+ - name: utkmst/chimera-beta-test2-lora-merged
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ dataset:
22
+ type: leaderboard
23
+ name: Overall Leaderboard
24
+ metrics:
25
+ - name: acc_norm
26
+ type: acc_norm
27
+ value: 0.4440
28
+ verified: true
29
+ - name: acc
30
+ type: acc
31
+ value: 0.2992
32
+ verified: true
33
+ - name: exact_match
34
+ type: exact_match
35
+ value: 0.0951
36
+ verified: true
37
+ - task:
38
+ type: text-generation
39
+ dataset:
40
+ type: bbh
41
+ name: BBH (Big Bench Hard)
42
+ metrics:
43
+ - name: acc_norm
44
+ type: acc_norm
45
+ value: 0.4773
46
+ verified: true
47
+ - task:
48
+ type: text-generation
49
+ dataset:
50
+ type: gpqa
51
+ name: GPQA (Google-Patched Question Answering)
52
+ metrics:
53
+ - name: acc_norm
54
+ type: acc_norm
55
+ value: 0.3036
56
+ verified: true
57
+ - task:
58
+ type: text-generation
59
+ dataset:
60
+ type: math
61
+ name: Math
62
+ metrics:
63
+ - name: exact_match
64
+ type: exact_match
65
+ value: 0.0951
66
+ verified: true
67
+ - task:
68
+ type: text-generation
69
+ dataset:
70
+ type: mmlu_pro
71
+ name: MMLU-Pro
72
+ metrics:
73
+ - name: acc
74
+ type: acc
75
+ value: 0.2992
76
+ verified: true
77
+ - task:
78
+ type: text-generation
79
+ dataset:
80
+ type: musr
81
+ name: MUSR (Multi-Step Reasoning)
82
+ metrics:
83
+ - name: acc_norm
84
+ type: acc_norm
85
+ value: 0.4113
86
+ verified: true
87
  ---
88
 
 
89
  # utkmst/chimera-beta-test2-lora-merged
90
 
91
  ## Model Description
92
  This model is a fine-tuned version of Meta's Llama-3.1-8B-Instruct model, created through LoRA fine-tuning on multiple instruction datasets, followed by merging the adapter weights with the base model.
93
 
94
+ ## Architecture
95
  - **Base Model**: meta-llama/Llama-3.1-8B-Instruct
96
+ - **Size**: 8.03B parameters
97
+ - **Type**: Decoder-only transformer
98
+ - **Format**: SafeTensors (full precision)
99
+
100
+ ## Training Details
101
  - **Training Method**: LoRA fine-tuning followed by adapter merging
102
+ - **LoRA Configuration**:
103
+ - Rank: 8
104
+ - Alpha: 16
105
+ - Trainable modules: Attention layers and feed-forward networks
106
+ - **Training Hyperparameters**:
107
+ - Learning rate: 2e-4
108
+ - Batch size: 2
109
+ - Training epochs: 1
110
+ - Optimizer: AdamW with constant scheduler
111
+
112
+ ## Intended Use
113
+ This model is designed for:
114
+ - General purpose assistant capabilities
115
+ - Question answering and knowledge retrieval
116
+ - Creative content generation
117
+ - Instructional guidance
118
 
119
+ ## Limitations
120
+ - Base model limitations including potential hallucinations and factual inaccuracies
121
+ - Limited context window compared to larger models
122
+ - Knowledge cutoff from the base Llama-3.1 model
123
+ - May exhibit biases present in training data
124
+ - Performance on specialized tasks may vary
125
 
126
  ## Usage with Transformers
127