bpawar commited on
Commit
f8345df
·
verified ·
1 Parent(s): 659ae5a

Initial commit

Browse files
Files changed (1) hide show
  1. README.md +107 -6
README.md CHANGED
@@ -1,6 +1,107 @@
1
- ---
2
- license: other
3
- license_name: nvidia-community-model-license
4
- license_link: >-
5
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-community-models-license/
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: nvidia-community-model-license
4
+ license_link: >-
5
+ https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-community-models-license/
6
+ language:
7
+ - en
8
+ base_model:
9
+ - mistralai/Mistral-Nemo-Instruct-2407
10
+ ---
11
+
12
+
13
+
14
+ # Mistral-Nemo-12B-Instruct-ONNX-INT4
15
+
16
+
17
+ ### Model Description
18
+
19
+ Mistral-NeMo is a Large Language Model (LLM) composed of 12B parameters. This model leads accuracy on popular benchmarks across common sense reasoning, coding, math, multilingual and multi-turn chat tasks; it significantly outperforms existing models smaller or similar in size.
20
+
21
+ We downloaded Mistral Nemo 12B instruct model in Pytorch bfloat16 format from HuggingFace. We used Onnxruntime-GenAI to convert the model from Pytorch FP16 format to ONNX FP16 format. We used TensorRT Model Optimizer - Windows tool to convert the model from ONNX FP16 format to ONNX INT4 fomat.  We have posted the Mistral Nemo 12B ONNX INT4 model files here. 
22
+
23
+ This model is ready for commercial/non-commercial use. 
24
+
25
+ ### Third-Party Community Consideration
26
+
27
+ This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; See link to Non-NVIDIA [Mistral-Nemo-12B-Instruct Model Card](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)
28
+
29
+ ### License/Terms of Use:
30
+
31
+ GOVERNING TERMS: Use of this model is governed by the NVIDIA Open Model License Agreement (found at https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf ). ADDITIONAL INFORMATION: Apache License, Version 2.0 (found at https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md ).
32
+
33
+ ### Reference(s):
34
+
35
+ Mistral NeMo 12B [Blogpost](https://mistral.ai/news/mistral-nemo/)
36
+
37
+ ### Model Architecture:
38
+
39
+ Mistral NeMo, a 12B model built in collaboration with NVIDIA. Mistral NeMo offers a large context window of up to 128k tokens. Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. As it relies on standard architecture, Mistral NeMo is easy to use as a drop-in replacement in any system using Mistral 7B. 
40
+
41
+ **Architecture Type:** Transformer <br>
42
+
43
+ **Network Architecture:** Mistral <br>
44
+
45
+ **Input**
46
+
47
+ * Input Type: Text
48
+
49
+ * Input Format: String
50
+
51
+ * Input Parameters: 1D
52
+
53
+ * Other Properties Related to Input: max_tokens, temperature, top_p, stop, frequency_penalty, presence_penalty, seed
54
+
55
+ **Output**
56
+
57
+ * Output Type: Text
58
+
59
+ * Output Format: String
60
+
61
+ * Output Parameters: 1D
62
+
63
+ ## Software Integration:
64
+
65
+ * **Supported Hardware Platform(s):**  Nvidia Ampere and newer GPUs. 6GB or higher VRAM GPUs are recommended. Higher VRAM may be required for larger context length use cases. 
66
+
67
+ * **Supported Operating System(s):**  Windows 
68
+
69
+ ## Model Version(s):  1.0 
70
+
71
+ ## Training, Testing, and Evaluation Datasets:  
72
+
73
+ Refer to [Mistral-Nemo-12B-Instruct Model Card](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) for the details. 
74
+
75
+ ### Evaluation Dataset:
76
+ Link: https://people.eecs.berkeley.edu/~hendrycks/data.tar
77
+
78
+ * Data Collection Method by dataset  - Unknown
79
+
80
+ * Labeling Method by dataset  - Not Applicable
81
+
82
+ ## Evaluation Results:
83
+
84
+ **MMLU (5# shots):**
85
+
86
+ With GenAI ORT->DML backend, we got below accuracy numbers on a desktop RTX 4090 GPU system. 
87
+
88
+ "overall_accuracy": 66.74
89
+
90
+ **Test configuration:**
91
+
92
+ * **GPU:** RTX 4090.  
93
+
94
+ * **Windows 11:** 23H2
95
+
96
+ * **NVIDIA Graphics driver:** R565 or higher
97
+
98
+ ## Inference:
99
+
100
+ We used GenAI ORT->DML backend for inference. The instructions to use this backend are given in readme.txt file available under Files section. 
101
+
102
+
103
+ ## Ethical Considerations:
104
+
105
+ NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications.  When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. 
106
+
107
+ Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).