lperng commited on
Commit
6b42307
·
verified ·
1 Parent(s): 0704da7

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - unsloth
5
+ - llama
6
+ - medical
7
+ - chinese
8
+ ---
9
+
10
+ # medical-model
11
+
12
+ This is a fine-tuned version of the [DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B) model, specifically trained on a medical dataset for enhanced performance in medical question answering and reasoning.
13
+
14
+ ## Model Description
15
+
16
+ - **Base Model:** DeepSeek-R1-Distill-Llama-8B
17
+ - **Fine-tuning Dataset:** [FreedomIntelligence/medical-o1-reasoning-SFT](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-reasoning-SFT)
18
+ - **Fine-tuning Method:** Unsloth's SFTTrainer with LoRA
19
+ - **Language:** Chinese
20
+
21
+ ## Intended Uses & Limitations
22
+
23
+ This model is intended for research and educational purposes related to medical question answering. It can provide insightful responses to medical queries, but it is crucial to understand its limitations:
24
+
25
+ - **Not a Medical Professional:** The model is not a substitute for professional medical advice. Consult with a healthcare provider for any health concerns or before making any decisions based on the model's output.
26
+ - **Potential Biases:** The training data may contain biases that could be reflected in the model's responses. Always critically evaluate the information provided.
27
+ - **Limited Scope:** The model's knowledge is limited to the information present in its training dataset. It may not be able to answer questions outside its domain of expertise.
28
+
29
+ ## How to Use
30
+
31
+ You can use this model with the following code:
32
+ ```
33
+ python from unsloth import FastLanguageModel import torch
34
+
35
+ model, tokenizer = FastLanguageModel.from_pretrained( model_name="lperng/medical-model", load_in_4bit=True, # Optional: Use 4-bit quantization )
36
+
37
+ prompt = "你的医学问题?" # Replace with your medical question
38
+
39
+ inputs = tokenizer([prompt], return_tensors="pt").to("cuda") outputs = model.generate(input_ids=inputs.input_ids, attention_mask=inputs.attention_mask) response = tokenizer.batch_decode(outputs)[0]
40
+
41
+ print(response)
42
+ ```
43
+ ## Training and Evaluation Data
44
+
45
+ This model was trained on the `FreedomIntelligence/medical-o1-reasoning-SFT` dataset. The dataset contains a collection of medical questions and corresponding answers. Evaluation metrics and results will be added in future updates.
46
+
47
+ ## Ethical Considerations
48
+
49
+ - **Potential for Misinformation:** The model's output should be carefully reviewed to ensure accuracy and prevent the spread of misinformation.
50
+ - **Privacy:** When using the model with personal health information, ensure compliance with relevant privacy regulations.
51
+ - **Responsible Use:** Use the model responsibly and ethically, considering its potential impact on individuals and society.
52
+
53
+ ## Citation
54
+
55
+ If you use this model in your work, please cite it as follows:
56
+ ```
57
+ @misc{medical-model, author = {lperng}, title = {medical-model}, year = {2023}, publisher = {Hugging Face}, howpublished = {\url{[redacted link], }
58
+ ```
59
+ ## Disclaimer
60
+
61
+ This model is provided "as is" without warranty of any kind, express or implied. The authors and contributors are not responsible for any damages or liabilities arising from the use of this model.
62
+