OpenMOSE commited on
Commit
185d1b8
·
1 Parent(s): 4596968

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -53,11 +53,12 @@ The training process consisted of three distinct stages:
53
  - PC Requirements 16GB+ VRAM NVIDIA GPU(rocm also can use. but only fp16.)
54
  - OS Windows WSL2 with CUDA or Linux
55
  - install RWKV-Infer(see how to install) https://github.com/OpenMOSE/RWKV-Infer
 
56
  - loadmodel(choose fp16 or fp6 or fp5 (dont choose FP8))
57
  - need 34GB VRAM in FP16, 14GB VRAM in FP5
58
  - Enjoy Text chats via Open-webui or Silly-Tavern :)
59
  ```
60
- curl http://127.0.0.1:9000/loadmodel -X POST -H "Content-Type: application/json" -d '{"model_filename":"models/rwkv-phi4-stage2.pth","model_viewname":"PRWKV7-Phi-4 Preview 0.1","model_strategy":"fp5","template":"phi4"}'
61
 
62
  ```
63
  3. you can use this model via OpenAI CompatibleAPI http://127.0.0.1:9000/v1 and set modelname "PRWKV7-Phi-4 Preview 0.1"
 
53
  - PC Requirements 16GB+ VRAM NVIDIA GPU(rocm also can use. but only fp16.)
54
  - OS Windows WSL2 with CUDA or Linux
55
  - install RWKV-Infer(see how to install) https://github.com/OpenMOSE/RWKV-Infer
56
+ - make folder "models" and put PRWKV-7-Phi-4-Instruct-Preview-v0.1.pth
57
  - loadmodel(choose fp16 or fp6 or fp5 (dont choose FP8))
58
  - need 34GB VRAM in FP16, 14GB VRAM in FP5
59
  - Enjoy Text chats via Open-webui or Silly-Tavern :)
60
  ```
61
+ curl http://127.0.0.1:9000/loadmodel -X POST -H "Content-Type: application/json" -d '{"model_filename":"models/PRWKV-7-Phi-4-Instruct-Preview-v0.1.pth","model_viewname":"PRWKV7-Phi-4 Preview 0.1","model_strategy":"fp5","template":"phi4"}'
62
 
63
  ```
64
  3. you can use this model via OpenAI CompatibleAPI http://127.0.0.1:9000/v1 and set modelname "PRWKV7-Phi-4 Preview 0.1"