Update README.md
Browse files
README.md
CHANGED
@@ -9,16 +9,16 @@ pipeline_tag: text-generation
|
|
9 |
---
|
10 |
# Model Card: Neural-phi2
|
11 |
|
12 |
-
![Poster Image](
|
13 |
|
14 |
## Model Details
|
15 |
|
16 |
- **Model Name**: Neural-phi2
|
17 |
- **Model Type**: Large Language Model (LLM)
|
18 |
- **Model Architecture**: A finetuned version of the Phi2 model from Microsoft, utilizing Direct Preference Optimization (DPO) on the `distilabel-intel-orca-dpo-pairs` dataset.
|
19 |
-
- **Model Size**: Approximately
|
20 |
- **Training Data**: The model was finetuned on the `distilabel-intel-orca-dpo-pairs` dataset, which consists of chat-like prompts and responses.
|
21 |
-
- **Training Procedure**: The Phi2 model was finetuned using the DPO technique
|
22 |
- Loading and formatting the `distilabel-intel-orca-dpo-pairs` dataset
|
23 |
- Defining the training configuration, including batch size, learning rate, and number of epochs
|
24 |
- Initializing the DPO Trainer and training the model
|
|
|
9 |
---
|
10 |
# Model Card: Neural-phi2
|
11 |
|
12 |
+
![Poster Image](poster.png)
|
13 |
|
14 |
## Model Details
|
15 |
|
16 |
- **Model Name**: Neural-phi2
|
17 |
- **Model Type**: Large Language Model (LLM)
|
18 |
- **Model Architecture**: A finetuned version of the Phi2 model from Microsoft, utilizing Direct Preference Optimization (DPO) on the `distilabel-intel-orca-dpo-pairs` dataset.
|
19 |
+
- **Model Size**: Approximately 2B parameters
|
20 |
- **Training Data**: The model was finetuned on the `distilabel-intel-orca-dpo-pairs` dataset, which consists of chat-like prompts and responses.
|
21 |
+
- **Training Procedure**: The Phi2 model was finetuned using the DPO technique. The training process involved:
|
22 |
- Loading and formatting the `distilabel-intel-orca-dpo-pairs` dataset
|
23 |
- Defining the training configuration, including batch size, learning rate, and number of epochs
|
24 |
- Initializing the DPO Trainer and training the model
|