tsqn commited on
Commit
416906a
·
verified ·
1 Parent(s): abab495

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -1,37 +1,37 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- wizardcoder-33b-v1.1-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
- *.gguf filter=lfs diff=lfs merge=lfs -text
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wizardcoder-33b-v1.1-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ *.gguf filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,67 +1,67 @@
1
- ---
2
- metrics:
3
- - code_eval
4
- library_name: transformers
5
- tags:
6
- - code
7
- - llama-cpp
8
- - gguf-my-repo
9
- base_model: WizardLMTeam/WizardCoder-33B-V1.1
10
- model-index:
11
- - name: WizardCoder
12
- results:
13
- - task:
14
- type: text-generation
15
- dataset:
16
- name: HumanEval
17
- type: openai_humaneval
18
- metrics:
19
- - type: pass@1
20
- value: 0.799
21
- name: pass@1
22
- verified: false
23
- ---
24
-
25
- # tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF
26
- This model was converted to GGUF format from [`WizardLMTeam/WizardCoder-33B-V1.1`](https://huggingface.co/WizardLMTeam/WizardCoder-33B-V1.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
- Refer to the [original model card](https://huggingface.co/WizardLMTeam/WizardCoder-33B-V1.1) for more details on the model.
28
-
29
- ## Use with llama.cpp
30
- Install llama.cpp through brew (works on Mac and Linux)
31
-
32
- ```bash
33
- brew install llama.cpp
34
-
35
- ```
36
- Invoke the llama.cpp server or the CLI.
37
-
38
- ### CLI:
39
- ```bash
40
- llama-cli --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -p "The meaning to life and the universe is"
41
- ```
42
-
43
- ### Server:
44
- ```bash
45
- llama-server --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -c 2048
46
- ```
47
-
48
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
49
-
50
- Step 1: Clone llama.cpp from GitHub.
51
- ```
52
- git clone https://github.com/ggerganov/llama.cpp
53
- ```
54
-
55
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
56
- ```
57
- cd llama.cpp && LLAMA_CURL=1 make
58
- ```
59
-
60
- Step 3: Run inference through the main binary.
61
- ```
62
- ./llama-cli --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -p "The meaning to life and the universe is"
63
- ```
64
- or
65
- ```
66
- ./llama-server --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -c 2048
67
- ```
 
1
+ ---
2
+ metrics:
3
+ - code_eval
4
+ library_name: transformers
5
+ tags:
6
+ - code
7
+ - llama-cpp
8
+ - gguf-my-repo
9
+ base_model: WizardLMTeam/WizardCoder-33B-V1.1
10
+ model-index:
11
+ - name: WizardCoder
12
+ results:
13
+ - task:
14
+ type: text-generation
15
+ dataset:
16
+ name: HumanEval
17
+ type: openai_humaneval
18
+ metrics:
19
+ - type: pass@1
20
+ value: 0.799
21
+ name: pass@1
22
+ verified: false
23
+ ---
24
+
25
+ # tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF
26
+ This model was converted to GGUF format from [`WizardLMTeam/WizardCoder-33B-V1.1`](https://huggingface.co/WizardLMTeam/WizardCoder-33B-V1.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
+ Refer to the [original model card](https://huggingface.co/WizardLMTeam/WizardCoder-33B-V1.1) for more details on the model.
28
+
29
+ ## Use with llama.cpp
30
+ Install llama.cpp through brew (works on Mac and Linux)
31
+
32
+ ```bash
33
+ brew install llama.cpp
34
+
35
+ ```
36
+ Invoke the llama.cpp server or the CLI.
37
+
38
+ ### CLI:
39
+ ```bash
40
+ llama-cli --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -p "The meaning to life and the universe is"
41
+ ```
42
+
43
+ ### Server:
44
+ ```bash
45
+ llama-server --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -c 2048
46
+ ```
47
+
48
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
49
+
50
+ Step 1: Clone llama.cpp from GitHub.
51
+ ```
52
+ git clone https://github.com/ggerganov/llama.cpp
53
+ ```
54
+
55
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
56
+ ```
57
+ cd llama.cpp && LLAMA_CURL=1 make
58
+ ```
59
+
60
+ Step 3: Run inference through the main binary.
61
+ ```
62
+ ./llama-cli --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -p "The meaning to life and the universe is"
63
+ ```
64
+ or
65
+ ```
66
+ ./llama-server --hf-repo tsqn/WizardCoder-33B-V1.1-Q8_0-GGUF --hf-file wizardcoder-33b-v1.1-q8_0.gguf -c 2048
67
+ ```
wizardcoder-33b-v1.1-q6_k-00001-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42d0436430fdb0651109e7b622a7b8afad83fc3995123f6c10b34b076d66280f
3
+ size 4940616736
wizardcoder-33b-v1.1-q6_k-00002-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:244bb71784231abee40ba969f277a47bf6e26c91c0cd208a5d53f94b29f9d294
3
+ size 4898846720
wizardcoder-33b-v1.1-q6_k-00003-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb841e6defcd59f49ac1936c28e012530dc4056d151e0e0f3a9453141f7e47a
3
+ size 4995184896
wizardcoder-33b-v1.1-q6_k-00004-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eeae3e6ad9f01b82f727468a4243cc0adc0aacaf418b248367936f7d0a57852
3
+ size 4898818016
wizardcoder-33b-v1.1-q6_k-00005-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b02ee360df49ad2d3e5477625e0cd6d8955b15051ddef79ee9bdac439702fd77
3
+ size 4898846720
wizardcoder-33b-v1.1-q6_k-00006-of-00006.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef4a1adf3885f9bcb5c286c0b70fd671f024c80130fa1d28ea15e437b1ea3812
3
+ size 2723399008