Update README.md
Browse files
README.md
CHANGED
@@ -53,11 +53,12 @@ The training process consisted of three distinct stages:
|
|
53 |
- PC Requirements 16GB+ VRAM NVIDIA GPU(rocm also can use. but only fp16.)
|
54 |
- OS Windows WSL2 with CUDA or Linux
|
55 |
- install RWKV-Infer(see how to install) https://github.com/OpenMOSE/RWKV-Infer
|
|
|
56 |
- loadmodel(choose fp16 or fp6 or fp5 (dont choose FP8))
|
57 |
- need 34GB VRAM in FP16, 14GB VRAM in FP5
|
58 |
- Enjoy Text chats via Open-webui or Silly-Tavern :)
|
59 |
```
|
60 |
-
curl http://127.0.0.1:9000/loadmodel -X POST -H "Content-Type: application/json" -d '{"model_filename":"models/
|
61 |
|
62 |
```
|
63 |
3. you can use this model via OpenAI CompatibleAPI http://127.0.0.1:9000/v1 and set modelname "PRWKV7-Phi-4 Preview 0.1"
|
|
|
53 |
- PC Requirements 16GB+ VRAM NVIDIA GPU(rocm also can use. but only fp16.)
|
54 |
- OS Windows WSL2 with CUDA or Linux
|
55 |
- install RWKV-Infer(see how to install) https://github.com/OpenMOSE/RWKV-Infer
|
56 |
+
- make folder "models" and put PRWKV-7-Phi-4-Instruct-Preview-v0.1.pth
|
57 |
- loadmodel(choose fp16 or fp6 or fp5 (dont choose FP8))
|
58 |
- need 34GB VRAM in FP16, 14GB VRAM in FP5
|
59 |
- Enjoy Text chats via Open-webui or Silly-Tavern :)
|
60 |
```
|
61 |
+
curl http://127.0.0.1:9000/loadmodel -X POST -H "Content-Type: application/json" -d '{"model_filename":"models/PRWKV-7-Phi-4-Instruct-Preview-v0.1.pth","model_viewname":"PRWKV7-Phi-4 Preview 0.1","model_strategy":"fp5","template":"phi4"}'
|
62 |
|
63 |
```
|
64 |
3. you can use this model via OpenAI CompatibleAPI http://127.0.0.1:9000/v1 and set modelname "PRWKV7-Phi-4 Preview 0.1"
|