SURESHBEEKHANI commited on
Commit
5e90b32
verified
1 Parent(s): 6b1c651

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -1,3 +1,13 @@
 
 
 
 
 
 
 
 
 
 
1
  # Fine-tuned Language Model for Preference Optimization (DPO)
2
 
3
  ## Model Overview
@@ -85,6 +95,4 @@ This model was trained using the Unsloth framework with contributions from Intel
85
 
86
  ## Notebook
87
 
88
- Access the implementation notebook for this model [here](https://github.com/SURESHBEEKHANI/Advanced-LLM-Fine-Tuning/blob/main/fine_tuning_llama_3_2_3b_dpo_peft.ipynb). This notebook provides detailed steps for fine-tuning and deploying the model.
89
-
90
-
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - Intel/orca_dpo_pairs
5
+ language:
6
+ - en
7
+ base_model:
8
+ - unsloth/Llama-3.2-3B-Instruct
9
+ pipeline_tag: question-answering
10
+ ---
11
  # Fine-tuned Language Model for Preference Optimization (DPO)
12
 
13
  ## Model Overview
 
95
 
96
  ## Notebook
97
 
98
+ Access the implementation notebook for this model [here](https://github.com/SURESHBEEKHANI/Advanced-LLM-Fine-Tuning/blob/main/fine_tuning_llama_3_2_3b_dpo_peft.ipynb). This notebook provides detailed steps for fine-tuning and deploying the model.