Update README.md
Browse files
README.md
CHANGED
@@ -89,3 +89,24 @@ This model is released under a permissible license.
|
|
89 |
|
90 |
Special thanks to the Unsloth team for providing an optimized training pipeline for LLaMA models.
|
91 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
89 |
|
90 |
Special thanks to the Unsloth team for providing an optimized training pipeline for LLaMA models.
|
91 |
|
92 |
+
**Disclaimer**
|
93 |
+
This model has been saved in the .bin format because it was trained using Unsloth. The .bin format is the default PyTorch serialization method and functions as expected. However, .bin files use Python's pickle module, which can execute arbitrary code during loading.
|
94 |
+
|
95 |
+
If security is a concern, we strongly recommend loading the model in a sandboxed environment such as staging servers, Kaggle, or Google Colab before deploying in production. You can also convert the model to .safetensors, a more secure and optimized format, using the following approach:
|
96 |
+
|
97 |
+
```python
|
98 |
+
from transformers import AutoModel
|
99 |
+
from safetensors.torch import save_file
|
100 |
+
|
101 |
+
# Load model
|
102 |
+
model = AutoModel.from_pretrained("path/to/model")
|
103 |
+
state_dict = model.state_dict()
|
104 |
+
|
105 |
+
# Convert to safetensors
|
106 |
+
save_file(state_dict, "model.safetensors")
|
107 |
+
|
108 |
+
print("Model converted to safetensors successfully.")
|
109 |
+
```
|
110 |
+
Alternatively, you can use our GGUF models, which are optimized for inference with llama.cpp, exllama, and other efficient runtimes. GGUF provides better performance on CPU/GPU and is a more portable option for deployment.
|
111 |
+
|
112 |
+
Choose the format that best suits your security, performance, and deployment needs.
|