ONNX
bpawar commited on
Commit
d6708bb
·
verified ·
1 Parent(s): 4fe167c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +120 -6
README.md CHANGED
@@ -1,6 +1,120 @@
1
- ---
2
- license: other
3
- license_name: nvidia-open-model-license
4
- license_link: >-
5
- https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: nvidia-open-model-license
4
+ license_link: >-
5
+ https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf
6
+ base_model:
7
+ - google/gemma-2b-it
8
+ ---
9
+
10
+
11
+ # Gemma-2b-it ONNX INT4
12
+
13
+ ## Model Developer: Google
14
+
15
+ ## Model Description
16
+
17
+ The NVIDIA Gemma-2b-it ONNX INT4 model is the quantized version of the Google Gemma-2b-it model which is a text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. It is well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure. For more information, please check [here](https://huggingface.co/google/gemma-2b-it). The NVIDIA Gemma-2b-it ONNX INT4 model is quantized with [TensorRT Model Optimizer](https://github.com/NVIDIA/TensorRT-Model-Optimizer).
18
+
19
+ This model is ready for commercial and research use case.
20
+
21
+
22
+ Steps followed to generate this quantized model:
23
+
24
+ * 1. Download Google Gemma-2b-it model in Pytorch bfloat16 format from HuggingFace.
25
+
26
+ * 2. Convert PyTorch model to ONNX FP16 using onnxruntime-genai model builder.
27
+
28
+ * 3. Quantize Gemma-2b-it ONNX FP16 model to Gemma-2b-it ONNX INT4 AWQ model using TensorRT Model Optimizer – Windows.
29
+
30
+ ## Third-Party Community Consideration
31
+ This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see link to the [Gemma-2b-it Model Card](https://huggingface.co/google/gemma-2b-it).
32
+
33
+
34
+ ## License/Terms of Use:
35
+ GOVERNING TERMS: Use of this model is governed by the NVIDIA Open Model License Agreement (found at https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf). ADDITIONAL INFORMATION: Gemma Terms of Use (found at https://ai.google.dev/gemma/terms).
36
+
37
+ ## Reference:
38
+
39
+ Gemma-2b-it [Model Card](https://huggingface.co/google/gemma-2b-it)
40
+
41
+ ## Model Architecture:
42
+
43
+ **Architecture Type:** Transformer <br>
44
+
45
+ **Network Architecture:** Gemma <br>
46
+
47
+ **Input**
48
+
49
+ * Input Type: Text
50
+
51
+ * Input Format: String
52
+
53
+ * Input Parameters: Sequence (1D)
54
+
55
+ * Other Properties Related to Input: Text strings can include a question, prompt, or a document to be summarized. Primarily for English language
56
+
57
+ **Output**
58
+
59
+ * Output Type: Text
60
+
61
+ * Output Format: String
62
+
63
+ * Output Parameters: Sequence (1D)
64
+
65
+ * Other properties: Generated English-language text in response to the input, such as an answer to a question, or a summary of a document.
66
+
67
+ ## Software Integration:
68
+
69
+ * **Supported Hardware Microarchitecture Compatibility :** Nvidia Ampere and newer GPUs. 6GB or higher VRAM GPUs are recommended. Higher VRAM may be required for larger context length use cases. 
70
+
71
+ * **Supported Operating System(s):**  Windows 
72
+
73
+ ## Model Version(s):  1.0 
74
+
75
+ ## Training, Testing, and Evaluation Datasets:  
76
+ Refer to [Gemma-2b-it Model Card](https://huggingface.co/google/gemma-2b-it ) for the details.
77
+
78
+ ### Calibration Dataset: cnn_daily mail used for calibration.
79
+
80
+ Link: https://huggingface.co/datasets/abisee/cnn_dailymail
81
+
82
+ * Data Collection Method by dataset: Automated
83
+
84
+ * Labeling Method by dataset: [Unknown]
85
+
86
+ ### Evaluation Dataset:
87
+
88
+ Link: https://people.eecs.berkeley.edu/~hendrycks/data.tar
89
+
90
+ * Data Collection Method by dataset  - Unknown
91
+
92
+ * Labeling Method by dataset  - Not Applicable
93
+
94
+ ## Evaluation Results:
95
+
96
+ **MMLU (5# shots):**
97
+
98
+ With GenAI ORT->DML backend, we got below mentioned accuracy numbers on a desktop RTX 4090 GPU system. 
99
+
100
+ "overall_accuracy": 37.26
101
+
102
+ **Test configuration:**
103
+
104
+ * **GPU:** RTX 4090, RTX 3090.  
105
+
106
+ * **Windows 11:** 23H2
107
+
108
+ * **NVIDIA Graphics driver:** R565 or higher
109
+
110
+ ## Inference:
111
+ Inference Backend: [Onnxruntime-GenAI-DirectML](https://onnxruntime.ai/docs/genai/howto/install.html#directml)
112
+
113
+ We used GenAI ORT->DML backend for inference. The instructions to use this backend are given in readme.txt file available under Files section. 
114
+
115
+
116
+ ## Ethical Considerations:
117
+
118
+ NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications.  When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. 
119
+
120
+ Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).