xsxszab commited on
Commit
8c3cda2
·
verified ·
1 Parent(s): cae61de

Create README.md

Browse files

initial commit for model card

Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - text-generation-inference
5
+ - GGUF
6
+ ---
7
+
8
+ # DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant
9
+
10
+ ## Introduction
11
+ **DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant** is a ... (TODO)
12
+
13
+ ---
14
+
15
+ ## How to Use on Your Device
16
+ Below, we outline multiple ways to run the model locally.
17
+
18
+ #### Option 1: Using Nexa SDK
19
+
20
+ **Step 1: Install Nexa SDK**
21
+
22
+ Follow the installation instructions in Nexa SDK's [GitHub repository](https://github.com/NexaAI/nexa-sdk).
23
+
24
+ **Step 2: Run the model with Nexa**
25
+
26
+ Execute the following command in your terminal:
27
+ ```bash
28
+ nexa run <model-path>
29
+ ```
30
+
31
+ #### Option 2: Using llama.cpp
32
+
33
+ **Step 1: Build llama.cpp on Your Device**
34
+
35
+ Follow the "Building the project" instructions in the llama.cpp [repository](https://github.com/ggerganov/llama.cpp) to build the project.
36
+
37
+ **Step 2: Run the Model with llama.cpp**
38
+
39
+ Once built, run `llama-cli` under `<build_dir>/bin/`:
40
+ ```bash
41
+ ./llama-cli \
42
+ --model your/local/path/to/DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant \
43
+ --prompt 'Provide step-by-step reasoning enclosed in <think> </think> tags, followed by the final answer enclosed in \boxed{} tags.' \
44
+ ```
45
+
46
+ #### Option 3: Using LM Studio
47
+
48
+ **Step 1: Download and Install LM Studio**
49
+
50
+ Get the latest version from the [official website](https://lmstudio.ai/).
51
+
52
+ **Step 2: Load and Run the Model**
53
+
54
+ 2. In LM Studio's top panel, search for and select `NexaAIDev/DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant`.
55
+ 3. Click `Download` (if not already downloaded) and wait for the model to load.
56
+ 4. Once loaded, go to the chat window and start a conversation.
57
+ ---