pharaouk commited on
Commit
011c685
·
verified ·
1 Parent(s): 9695eaf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -1
README.md CHANGED
@@ -11,6 +11,11 @@ tags:
11
  ---
12
  Thanks to V2RAY for uploading, i shamelessly forked it here [here](https://huggingface.co/v2ray/Mixtral-8x22B-v0.1)
13
 
 
 
 
 
 
14
  # Model Card for Mixtral-8x22B
15
  Converted to HuggingFace Transformers format using the script [here](https://huggingface.co/v2ray/Mixtral-8x22B-v0.1/blob/main/convert.py).
16
 
@@ -19,7 +24,7 @@ The Mixtral-8x22B Large Language Model (LLM) is a pretrained generative Sparse M
19
  ```python
20
  from transformers import AutoModelForCausalLM, AutoTokenizer
21
 
22
- model_id = "v2ray/Mixtral-8x22B-v0.1"
23
  tokenizer = AutoTokenizer.from_pretrained(model_id)
24
 
25
  model = AutoModelForCausalLM.from_pretrained(model_id)
 
11
  ---
12
  Thanks to V2RAY for uploading, i shamelessly forked it here [here](https://huggingface.co/v2ray/Mixtral-8x22B-v0.1)
13
 
14
+
15
+
16
+
17
+
18
+
19
  # Model Card for Mixtral-8x22B
20
  Converted to HuggingFace Transformers format using the script [here](https://huggingface.co/v2ray/Mixtral-8x22B-v0.1/blob/main/convert.py).
21
 
 
24
  ```python
25
  from transformers import AutoModelForCausalLM, AutoTokenizer
26
 
27
+ model_id = "pharaouk/Mixtral-8x22B-v0.1"
28
  tokenizer = AutoTokenizer.from_pretrained(model_id)
29
 
30
  model = AutoModelForCausalLM.from_pretrained(model_id)