alphaaico commited on
Commit
f427b66
·
verified ·
1 Parent(s): e8e7693

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -0
README.md CHANGED
@@ -89,3 +89,24 @@ This model is released under a permissible license.
89
 
90
  Special thanks to the Unsloth team for providing an optimized training pipeline for LLaMA models.
91
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
 
90
  Special thanks to the Unsloth team for providing an optimized training pipeline for LLaMA models.
91
 
92
+ **Disclaimer**
93
+ This model has been saved in the .bin format because it was trained using Unsloth. The .bin format is the default PyTorch serialization method and functions as expected. However, .bin files use Python's pickle module, which can execute arbitrary code during loading.
94
+
95
+ If security is a concern, we strongly recommend loading the model in a sandboxed environment such as staging servers, Kaggle, or Google Colab before deploying in production. You can also convert the model to .safetensors, a more secure and optimized format, using the following approach:
96
+
97
+ ```python
98
+ from transformers import AutoModel
99
+ from safetensors.torch import save_file
100
+
101
+ # Load model
102
+ model = AutoModel.from_pretrained("path/to/model")
103
+ state_dict = model.state_dict()
104
+
105
+ # Convert to safetensors
106
+ save_file(state_dict, "model.safetensors")
107
+
108
+ print("Model converted to safetensors successfully.")
109
+ ```
110
+ Alternatively, you can use our GGUF models, which are optimized for inference with llama.cpp, exllama, and other efficient runtimes. GGUF provides better performance on CPU/GPU and is a more portable option for deployment.
111
+
112
+ Choose the format that best suits your security, performance, and deployment needs.