helenai commited on
Commit
fc0e6f9
·
verified ·
1 Parent(s): e95b074

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -0
README.md ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ This is the [Deepseek-R1-Distill-Qwen-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) model, convert to OpenVINO with symmetric channel-wise INT4 weight compression.
2
+
3
+ To run inference on this model, install openvino-genai (`pip install openvino-genai`) and run [llm_chat_deepseek.py(https://gist.github.com/helena-intel/554fba91f380df590ecc9245abdad33f)
4
+
5
+ Step-by-step instructions for best results:
6
+
7
+ ```
8
+ pip install --pre --upgrade openvino openvino-genai openvino-tokenizers --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
9
+ pip install huggingface-hub
10
+ huggingface-cli download helenai/DeepSeek-R1-Distill-Qwen-7B-ov-int4 --local-dir DeepSeek-R1-Distill-Qwen-7B-ov-int4
11
+ curl -O https://gist.githubusercontent.com/helena-intel/554fba91f380df590ecc9245abdad33f/raw/04f495164482823aa7e6ba1119a5c43e275d08f5/llm_chat_deepseek.py
12
+ python llm_chat_deepseek.py DeepSeek-R1-Distill-Qwen-7B-ov-int4 GPU
13
+ ```
14
+
15
+ > [!NOTE]
16
+ > The last line specifies the device to run inference. GPU is recommended for recent Intel laptops with integrated graphics, or for Intel discrete graphics. Change to CPU if you do not have an Intel GPU, or to NPU if you have a system with an Intel NPU