Uploaded model

  • Developed by: msthil2
  • License: apache-2.0
  • Finetuned from model : unsloth/Llama-3.2-3B-Instruct-bnb-4bit

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

PLEASE NOTE: You may need to do model.config.torch_dtype = torch.float16 when loading from the hub. Related to this github which wasn't fixed when this was trained: https://github.com/unslothai/unsloth/issues/404

Downloads last month
24
Safetensors
Model size
1.87B params
Tensor type
F32
FP16
U8
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for msthil2/dad_gpt

Space using msthil2/dad_gpt 1