Create README.md
Browse filesinitial commit for model card
README.md
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- text-generation-inference
|
5 |
+
- GGUF
|
6 |
+
---
|
7 |
+
|
8 |
+
# DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant
|
9 |
+
|
10 |
+
## Introduction
|
11 |
+
**DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant** is a ... (TODO)
|
12 |
+
|
13 |
+
---
|
14 |
+
|
15 |
+
## How to Use on Your Device
|
16 |
+
Below, we outline multiple ways to run the model locally.
|
17 |
+
|
18 |
+
#### Option 1: Using Nexa SDK
|
19 |
+
|
20 |
+
**Step 1: Install Nexa SDK**
|
21 |
+
|
22 |
+
Follow the installation instructions in Nexa SDK's [GitHub repository](https://github.com/NexaAI/nexa-sdk).
|
23 |
+
|
24 |
+
**Step 2: Run the model with Nexa**
|
25 |
+
|
26 |
+
Execute the following command in your terminal:
|
27 |
+
```bash
|
28 |
+
nexa run <model-path>
|
29 |
+
```
|
30 |
+
|
31 |
+
#### Option 2: Using llama.cpp
|
32 |
+
|
33 |
+
**Step 1: Build llama.cpp on Your Device**
|
34 |
+
|
35 |
+
Follow the "Building the project" instructions in the llama.cpp [repository](https://github.com/ggerganov/llama.cpp) to build the project.
|
36 |
+
|
37 |
+
**Step 2: Run the Model with llama.cpp**
|
38 |
+
|
39 |
+
Once built, run `llama-cli` under `<build_dir>/bin/`:
|
40 |
+
```bash
|
41 |
+
./llama-cli \
|
42 |
+
--model your/local/path/to/DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant \
|
43 |
+
--prompt 'Provide step-by-step reasoning enclosed in <think> </think> tags, followed by the final answer enclosed in \boxed{} tags.' \
|
44 |
+
```
|
45 |
+
|
46 |
+
#### Option 3: Using LM Studio
|
47 |
+
|
48 |
+
**Step 1: Download and Install LM Studio**
|
49 |
+
|
50 |
+
Get the latest version from the [official website](https://lmstudio.ai/).
|
51 |
+
|
52 |
+
**Step 2: Load and Run the Model**
|
53 |
+
|
54 |
+
2. In LM Studio's top panel, search for and select `NexaAIDev/DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant`.
|
55 |
+
3. Click `Download` (if not already downloaded) and wait for the model to load.
|
56 |
+
4. Once loaded, go to the chat window and start a conversation.
|
57 |
+
---
|