ZhafranR's picture
Add model card
7705459 verified
---
language:
- en
library_name: llama.cpp
pipeline_tag: text-generation
license: apache-2.0
quantization: q4_k_m
---
# LLaMa_3.2_3B_Instruct_Text2SQL-Q4_K_M-GGUF.gguf
This is a GGUF quantized version of the LLaMa 3.2 3B Text2SQL model.
## Model Details
- **Architecture:** LLaMa 3.2 3B
- **Task:** Text to SQL Generation
- **Quantization:** Q4_K_M
- **Context Length:** 65536 tokens (2^16)
- **Format:** GGUF (Compatible with llama.cpp)
## Usage
```python
from llama_cpp import Llama
# Initialize model
llm = Llama(
model_path="downloaded_model.gguf",
n_ctx=65536, # 64K context
n_threads=8 # Adjust based on your CPU
)
# Generate SQL
response = llm(
"Convert this to SQL: Find all users who signed up in January 2024",
max_tokens=1024,
temperature=0.7
)
print(response['choices'][0]['text'])
```
## Model Source
This is a quantized version of [XeAI/LLaMa_3.2_3B_Instruct_Text2SQL](https://huggingface.co/XeAI/LLaMa_3.2_3B_Instruct_Text2SQL)