Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
# Fine-tuned Language Model for Preference Optimization (DPO)
|
2 |
|
3 |
## Model Overview
|
@@ -85,6 +95,4 @@ This model was trained using the Unsloth framework with contributions from Intel
|
|
85 |
|
86 |
## Notebook
|
87 |
|
88 |
-
Access the implementation notebook for this model [here](https://github.com/SURESHBEEKHANI/Advanced-LLM-Fine-Tuning/blob/main/fine_tuning_llama_3_2_3b_dpo_peft.ipynb). This notebook provides detailed steps for fine-tuning and deploying the model.
|
89 |
-
|
90 |
-
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- Intel/orca_dpo_pairs
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
base_model:
|
8 |
+
- unsloth/Llama-3.2-3B-Instruct
|
9 |
+
pipeline_tag: question-answering
|
10 |
+
---
|
11 |
# Fine-tuned Language Model for Preference Optimization (DPO)
|
12 |
|
13 |
## Model Overview
|
|
|
95 |
|
96 |
## Notebook
|
97 |
|
98 |
+
Access the implementation notebook for this model [here](https://github.com/SURESHBEEKHANI/Advanced-LLM-Fine-Tuning/blob/main/fine_tuning_llama_3_2_3b_dpo_peft.ipynb). This notebook provides detailed steps for fine-tuning and deploying the model.
|
|
|
|