To run the model run the following code :
!pip install ctransformers[cuda]
from ctransformers import AutoModelForCausalLM
llm = AutoModelForCausalLM.from_pretrained("epsil/Tinyllama-1b-v1.0-gguf", model_file="Tinyllama-1b-v1.0.gguf")
print(llm("AI is going to"))
- Downloads last month
- 6
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.