OpenMOSE commited on
Commit
f9a57f6
·
1 Parent(s): ffddee1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -45,6 +45,16 @@ The training process consisted of three distinct stages:
45
  - Utilized a distillation dataset of 900K samples
46
  - Implemented VariableRankPEFT to enhance training efficiency
47
 
 
 
 
 
 
 
 
 
 
 
48
  ## Training Infrastructure
49
  - Hardware: Single AMD MI300X GPU
50
  - Training Duration: 3 days(Stage1,2)
@@ -59,6 +69,10 @@ This work was made possible through the contributions of:
59
  - RecursalAI
60
  - RWKV-Red-Team
61
 
 
 
 
 
62
  ## Limitations
63
 
64
  This is trained Stage2. Stage3 is on going :)
 
45
  - Utilized a distillation dataset of 900K samples
46
  - Implemented VariableRankPEFT to enhance training efficiency
47
 
48
+ ## How to Use
49
+ - install RWKV-Infer(see how to install) https://github.com/OpenMOSE/RWKV-Infer
50
+ - loadmodel(choose fp16 or fp6 or fp5 (dont choose FP8))
51
+ - need 19GB VRAM in FP16, 12GB VRAM in FP6
52
+ ```
53
+ curl http://127.0.0.1:9000/loadmodel -X POST -H "Content-Type: application/json" -d '{"model_filename":"models/rwkv-phi4-stage2.pth","model_viewname":"PRWKV7-Phi-4 Preview 0.1","model_strategy":"fp5","template":"phi4"}'
54
+
55
+ ```
56
+ 3. you can use this model via OpenAI CompatibleAPI http://127.0.0.1:9000/v1 and set modelname "PRWKV7-Phi-4 Preview 0.1"
57
+
58
  ## Training Infrastructure
59
  - Hardware: Single AMD MI300X GPU
60
  - Training Duration: 3 days(Stage1,2)
 
69
  - RecursalAI
70
  - RWKV-Red-Team
71
 
72
+ - https://github.com/OpenMOSE/RWKVInside
73
+ - https://github.com/OpenMOSE/RWKV-LM-RLHF
74
+ - https://github.com/OpenMOSE/RWKV-Infer
75
+
76
  ## Limitations
77
 
78
  This is trained Stage2. Stage3 is on going :)