brittlewis12 commited on
Commit
7aa20a2
·
verified ·
1 Parent(s): 93f8fc9

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +141 -0
README.md ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-3-27b-it
3
+ pipeline_tag: text-generation
4
+ inference: true
5
+ language:
6
+ - en
7
+ license: gemma
8
+ model_creator: google
9
+ model_name: gemma-3-27b-it
10
+ model_type: gemma3
11
+ quantized_by: brittlewis12
12
+ tags:
13
+ - gemma
14
+ ---
15
+
16
+ # Gemma 3 27B IT GGUF
17
+
18
+ **Original model**: [Gemma 3 27B IT](https://huggingface.co/google/gemma-3-27b-it)
19
+
20
+ **Model creator**: [Google DeepMind](https://huggingface.co/google)
21
+
22
+ > Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
23
+
24
+ > Gemma 3 has a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions. Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.
25
+
26
+ This repo contains GGUF format model files for Google DeepMind’s Gemma 3 27B IT (instruction-tuned).
27
+
28
+ ### What is GGUF?
29
+
30
+ GGUF is a file format for representing AI models. It is the third version of the format,
31
+ introduced by the llama.cpp team on August 21st 2023.
32
+
33
+ Converted with llama.cpp build b4875 (revision [7841fc7](https://github.com/ggml-org/llama.cpp/commits/7841fc723e059d1fd9640e5c0ef19050fcc7c698)),
34
+ using [autogguf-rs](https://github.com/brittlewis12/autogguf-rs).
35
+
36
+ ### Prompt template: [Gemma Instruct](https://huggingface.co/google/gemma-3-27b-it/raw/main/tokenizer_config.json)
37
+
38
+ ```
39
+ {{system_prompt}}
40
+ <start_of_turn>user
41
+ {{prompt}}<end_of_turn>
42
+ <start_of_turn>model
43
+
44
+
45
+ ```
46
+
47
+ ---
48
+
49
+ ## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!
50
+
51
+ ![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)
52
+
53
+ [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
54
+ - create & save **Characters** with custom system prompts & temperature settings
55
+ - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
56
+ * or, use an API key with the chat completions-compatible model provider of your choice -- ChatGPT, Claude, Gemini, DeepSeek, & more!
57
+ - make it your own with custom **Theme colors**
58
+ - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggml-org/llama.cpp), with **haptics** during response streaming!
59
+ - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
60
+ * if you **already have the app**, download Gemma 3 27B IT now!
61
+ * <cnvrsai:///models/search/hf?id=brittlewis12/gemma-3-27b-it-GGUF>
62
+ - follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date
63
+
64
+ ### Gemma 3 27B IT in cnvrs on macOS
65
+
66
+ ![gemma-3 in cnvrs](https://cdn-uploads.huggingface.co/production/uploads/63b64d7a889aa6707f155cdb/hf_Z7YjT26hK8fXo2EPah.png)
67
+
68
+ ---
69
+
70
+ ## Original Model Evaluation
71
+
72
+ > These models were evaluated against a large collection of different datasets and
73
+ metrics to cover different aspects of text generation:
74
+
75
+ #### Reasoning and factuality
76
+
77
+ | Benchmark | Metric | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
78
+ | ------------------------------ |----------------|:--------------:|:-------------:|:--------------:|:--------------:|
79
+ | [HellaSwag][hellaswag] | 10-shot | 62.3 | 77.2 | 84.2 | 85.6 |
80
+ | [BoolQ][boolq] | 0-shot | 63.2 | 72.3 | 78.8 | 82.4 |
81
+ | [PIQA][piqa] | 0-shot | 73.8 | 79.6 | 81.8 | 83.3 |
82
+ | [SocialIQA][socialiqa] | 0-shot | 48.9 | 51.9 | 53.4 | 54.9 |
83
+ | [TriviaQA][triviaqa] | 5-shot | 39.8 | 65.8 | 78.2 | 85.5 |
84
+ | [Natural Questions][naturalq] | 5-shot | 9.48 | 20.0 | 31.4 | 36.1 |
85
+ | [ARC-c][arc] | 25-shot | 38.4 | 56.2 | 68.9 | 70.6 |
86
+ | [ARC-e][arc] | 0-shot | 73.0 | 82.4 | 88.3 | 89.0 |
87
+ | [WinoGrande][winogrande] | 5-shot | 58.2 | 64.7 | 74.3 | 78.8 |
88
+ | [BIG-Bench Hard][bbh] | few-shot | 28.4 | 50.9 | 72.6 | 77.7 |
89
+ | [DROP][drop] | 1-shot | 42.4 | 60.1 | 72.2 | 77.2 |
90
+
91
+ [hellaswag]: https://arxiv.org/abs/1905.07830
92
+ [boolq]: https://arxiv.org/abs/1905.10044
93
+ [piqa]: https://arxiv.org/abs/1911.11641
94
+ [socialiqa]: https://arxiv.org/abs/1904.09728
95
+ [triviaqa]: https://arxiv.org/abs/1705.03551
96
+ [naturalq]: https://github.com/google-research-datasets/natural-questions
97
+ [arc]: https://arxiv.org/abs/1911.01547
98
+ [winogrande]: https://arxiv.org/abs/1907.10641
99
+ [bbh]: https://paperswithcode.com/dataset/bbh
100
+ [drop]: https://arxiv.org/abs/1903.00161
101
+
102
+ #### STEM and code
103
+
104
+ | Benchmark | Metric | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
105
+ | ------------------------------ |----------------|:-------------:|:--------------:|:--------------:|
106
+ | [MMLU][mmlu] | 5-shot | 59.6 | 74.5 | 78.6 |
107
+ | [MMLU][mmlu] (Pro COT) | 5-shot | 29.2 | 45.3 | 52.2 |
108
+ | [AGIEval][agieval] | 3-5-shot | 42.1 | 57.4 | 66.2 |
109
+ | [MATH][math] | 4-shot | 24.2 | 43.3 | 50.0 |
110
+ | [GSM8K][gsm8k] | 8-shot | 38.4 | 71.0 | 82.6 |
111
+ | [GPQA][gpqa] | 5-shot | 15.0 | 25.4 | 24.3 |
112
+ | [MBPP][mbpp] | 3-shot | 46.0 | 60.4 | 65.6 |
113
+ | [HumanEval][humaneval] | 0-shot | 36.0 | 45.7 | 48.8 |
114
+
115
+ [mmlu]: https://arxiv.org/abs/2009.03300
116
+ [agieval]: https://arxiv.org/abs/2304.06364
117
+ [math]: https://arxiv.org/abs/2103.03874
118
+ [gsm8k]: https://arxiv.org/abs/2110.14168
119
+ [gpqa]: https://arxiv.org/abs/2311.12022
120
+ [mbpp]: https://arxiv.org/abs/2108.07732
121
+ [humaneval]: https://arxiv.org/abs/2107.03374
122
+
123
+ #### Multilingual
124
+
125
+ | Benchmark | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
126
+ | ------------------------------------ |:-------------:|:-------------:|:--------------:|:--------------:|
127
+ | [MGSM][mgsm] | 2.04 | 34.7 | 64.3 | 74.3 |
128
+ | [Global-MMLU-Lite][global-mmlu-lite] | 24.9 | 57.0 | 69.4 | 75.7 |
129
+ | [WMT24++][wmt24pp] (ChrF) | 36.7 | 48.4 | 53.9 | 55.7 |
130
+ | [FloRes][flores] | 29.5 | 39.2 | 46.0 | 48.8 |
131
+ | [XQuAD][xquad] (all) | 43.9 | 68.0 | 74.5 | 76.8 |
132
+ | [ECLeKTic][eclektic] | 4.69 | 11.0 | 17.2 | 24.4 |
133
+ | [IndicGenBench][indicgenbench] | 41.4 | 57.2 | 61.7 | 63.4 |
134
+
135
+ [mgsm]: https://arxiv.org/abs/2210.03057
136
+ [flores]: https://arxiv.org/abs/2106.03193
137
+ [xquad]: https://arxiv.org/abs/1910.11856v3
138
+ [global-mmlu-lite]: https://huggingface.co/datasets/CohereForAI/Global-MMLU-Lite
139
+ [wmt24pp]: https://arxiv.org/abs/2502.12404v1
140
+ [eclektic]: https://arxiv.org/abs/2502.21228
141
+ [indicgenbench]: https://arxiv.org/abs/2404.16816