File size: 2,888 Bytes
a708e66 ac99ef3 a708e66 ac99ef3 5ba1fb5 700a117 b64d4f3 5d27613 7d5e5d2 5d27613 36bd97e 5d27613 36bd97e 5d27613 36bd97e 5d27613 700a117 8701283 5ba1fb5 f71a5a2 ac99ef3 bc2e5de ac99ef3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 |
---
license: apache-2.0
language:
- en
- fr
- de
- es
- it
tags:
- moe
---
Converted version of [Mixtral Instruct](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) to 4-bit using bitsandbytes. For more information about the model, refer to the model's page.
## Impact on performance
Impact of quantization on a set of models.
Evaluation of the model was conducted using the PoLL (Pool of LLM) technique, assessing performance on **100 French questions** with scores aggregated from six evaluations
(two per evaluator). The evaluators included GPT-4o, Gemini-1.5-pro, and Claude3.5-sonnet.
Performance Scores (on a scale of 5):
| Model | Score | # params (Billion) | size (GB) |
|---------------------------------------------:|:--------:|:------------------:|:---------:|
| gpt-4o | 4.13 | N/A | N/A |
| gpt-4o-mini | 4.02 | N/A | N/A |
| Qwen/Qwen2.5-32B-Instruct | 3.99 | 32.8 | 65.6 |
| cmarkea/Qwen2.5-32B-Instruct-4bit | 3.98 | 32.8 | 16.4 |
| mistralai/Mixtral-8x7B-Instruct-v0.1 | 3.71 | 46.7 | 93.4 |
| **cmarkea/Mixtral-8x7B-Instruct-v0.1-4bit** | **3.68** | **46.7** | **23.35** |
| meta-llama/Meta-Llama-3.1-70B-Instruct | 3.68 | 70.06 | 140.12 |
| gpt-3.5-turbo | 3.66 | 175 | 350 |
| cmarkea/Meta-Llama-3.1-70B-Instruct-4bit | 3.64 | 70.06 | 35.3 |
| TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ | 3.56 | 46.7 | 46.7 |
| meta-llama/Meta-Llama-3.1-8B-Instruct | 3.25 | 8.03 | 16.06 |
| mistralai/Mistral-7B-Instruct-v0.2 | 1.98 | 7.25 | 14.5 |
| cmarkea/bloomz-7b1-mt-sft-chat | 1.69 | 7.07 | 14.14 |
| cmarkea/bloomz-3b-dpo-chat | 1.68 | 3 | 6 |
| cmarkea/bloomz-3b-sft-chat | 1.51 | 3 | 6 |
| croissantllm/CroissantLLMChat-v0.1 | 1.19 | 1.3 | 2.7 |
| cmarkea/bloomz-560m-sft-chat | 1.04 | 0.56 | 1.12 |
| OpenLLM-France/Claire-Mistral-7B-0.1 | 0.38 | 7.25 | 14.5 |
The impact of quantization is negligible.
## Prompt Pattern
Here is a reminder of the command pattern to interact with the model if the add_special_tokens option is disabled (otherwise, do not include the BOS symbol and space at the beginning of the sequence):
```verbatim
<s> [INST] {user_prompt_1} [/INST] model_answer_1</s> [INST] {user_prompt_2} [/INST] model_answer_2</s>
``` |