SmerkyG commited on
Commit
121df86
·
verified ·
1 Parent(s): 17c2f35

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -21
README.md CHANGED
@@ -1,16 +1,16 @@
1
  ---
2
  license: apache-2.0
3
  ---
4
- ### Huggingface RWKV Finch 7B Model
5
 
6
- > HF compatible model for Finch-7B.
7
 
8
  ![Finch Bird](./imgs/finch.jpg)
9
 
10
 
11
  > **! Important Note !**
12
  >
13
- > The following is the HF transformers implementation of the Finch 7B model. This is meant to be used with the huggingface transformers
14
  >
15
  >
16
 
@@ -18,21 +18,21 @@ license: apache-2.0
18
  ## Quickstart with the hugging face transformer library
19
 
20
  ```
21
- model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True).to(torch.float32)
22
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True)
23
  ```
24
 
25
  ## Evaluation
26
 
27
- The following demonstrates the improvements from Eagle 7B to Finch 14B
28
 
29
- | | [Eagle 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 14B](https://huggingface.co/RWKV/v6-Finch-14B-HF) |
30
- | --- | --- | --- | --- |
31
- | [ARC](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/arc) | 39.59% | 41.47% | 46.33% |
32
- | [HellaSwag](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/hellaswag) | 53.09% | 55.96% | 57.69% |
33
- | [MMLU](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/mmlu) | 30.86% | 41.70% | 56.05% |
34
- | [Truthful QA](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/truthfulqa) | 33.03% | 34.82% | 39.27% |
35
- | [Winogrande](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/winogrande) | 67.56% | 71.19% | 74.43% |
36
 
37
  #### Running on CPU via HF transformers
38
 
@@ -59,8 +59,8 @@ User: {instruction}
59
  Assistant:"""
60
 
61
 
62
- model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True).to(torch.float32)
63
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True)
64
 
65
  text = "请介绍北京的旅游景点"
66
  prompt = generate_prompt(text)
@@ -115,8 +115,8 @@ User: {instruction}
115
  Assistant:"""
116
 
117
 
118
- model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True, torch_dtype=torch.float16).to(0)
119
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True)
120
 
121
  text = "介绍一下大熊猫"
122
  prompt = generate_prompt(text)
@@ -162,8 +162,8 @@ User: {instruction}
162
 
163
  Assistant:"""
164
 
165
- model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True).to(torch.float32)
166
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-7B-HF", trust_remote_code=True)
167
 
168
  texts = ["请介绍北京的旅游景点", "介绍一下大熊猫", "乌兰察布"]
169
  prompts = [generate_prompt(text) for text in texts]
@@ -208,8 +208,7 @@ Assistant: 乌兰察布是中国新疆维吾尔自治区的一个县级市,位
208
  ## Links
209
  - [Our wiki](https://wiki.rwkv.com)
210
  - [Recursal.AI Cloud Platform](https://recursal.ai)
211
- - [Featherless Inference](https://featherless.ai/models/RWKV/Finch-14B)
212
- - [Blog article, detailing our model launch](https://blog.rwkv.com/p/rwkv-v6-finch-14b-is-here)
213
 
214
  ## Acknowledgement
215
  We are grateful for the help and support from the following key groups:
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ ### Huggingface RWKV Flock of Finches 36B-A11B Mixture of Experts Model
5
 
6
+ > HF compatible model for Finch-MoE-36B-A11B.
7
 
8
  ![Finch Bird](./imgs/finch.jpg)
9
 
10
 
11
  > **! Important Note !**
12
  >
13
+ > The following is the HF transformers implementation of the Flock of Finches Mixture of Experts 36B-A11B model. This is meant to be used with the huggingface transformers
14
  >
15
  >
16
 
 
18
  ## Quickstart with the hugging face transformer library
19
 
20
  ```
21
+ model = AutoModelForCausalLM.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True).to(torch.float32)
22
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True)
23
  ```
24
 
25
  ## Evaluation
26
 
27
+ The following demonstrates the improvements from Eagle 7B to Flock of Finches 36B-A11B
28
 
29
+ | | [Eagle 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 14B](https://huggingface.co/RWKV/v6-Finch-14B-HF) | [Flock of Finches 36B-A11B]
30
+ | --- | --- | --- | --- | --- |
31
+ | [ARC](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/arc) | 39.59% | 41.47% | 46.33% | 48.21%
32
+ | [HellaSwag](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/hellaswag) | 53.09% | 55.96% | 57.69% | 57.69%
33
+ | [MMLU](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/mmlu) | 30.86% | 41.70% | 56.05% | 55.0%
34
+ | [Truthful QA](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/truthfulqa) | 33.03% | 34.82% | 39.27% |
35
+ | [Winogrande](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/winogrande) | 67.56% | 71.19% | 74.43% | 75.77%
36
 
37
  #### Running on CPU via HF transformers
38
 
 
59
  Assistant:"""
60
 
61
 
62
+ model = AutoModelForCausalLM.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True).to(torch.float32)
63
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True)
64
 
65
  text = "请介绍北京的旅游景点"
66
  prompt = generate_prompt(text)
 
115
  Assistant:"""
116
 
117
 
118
+ model = AutoModelForCausalLM.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True, torch_dtype=torch.float16).to(0)
119
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True)
120
 
121
  text = "介绍一下大熊猫"
122
  prompt = generate_prompt(text)
 
162
 
163
  Assistant:"""
164
 
165
+ model = AutoModelForCausalLM.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True).to(torch.float32)
166
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/Finch-MoE-36B-A11B-HF", trust_remote_code=True)
167
 
168
  texts = ["请介绍北京的旅游景点", "介绍一下大熊猫", "乌兰察布"]
169
  prompts = [generate_prompt(text) for text in texts]
 
208
  ## Links
209
  - [Our wiki](https://wiki.rwkv.com)
210
  - [Recursal.AI Cloud Platform](https://recursal.ai)
211
+ - [Featherless Inference](https://featherless.ai/models/RWKV/)
 
212
 
213
  ## Acknowledgement
214
  We are grateful for the help and support from the following key groups: