Update my_app.aiconfig.json
Browse files- my_app.aiconfig.json +2 -2
my_app.aiconfig.json
CHANGED
@@ -14,13 +14,13 @@
|
|
14 |
"description": "In this notebook, we compare the individual performance of HF hosted LLMs () on a few example questions from the GLUE benchmarks (https://gluebenchmark.com/tasks).\n\nExample questions taken from \"What is the GLUE Benchmark\" medium post - https://angelina-yang.medium.com/what-is-the-glue-benchmark-for-nlu-systems-61127b3cab3f",
|
15 |
"prompts": [
|
16 |
{
|
17 |
-
"name": "
|
18 |
"input": "{{SST_2_ex_prompt}}",
|
19 |
"metadata": {
|
20 |
"model": {
|
21 |
"name": "Text Generation",
|
22 |
"settings": {
|
23 |
-
"model": "mistralai/Mistral-7B-v0.1"
|
24 |
}
|
25 |
},
|
26 |
"tags": null,
|
|
|
14 |
"description": "In this notebook, we compare the individual performance of HF hosted LLMs () on a few example questions from the GLUE benchmarks (https://gluebenchmark.com/tasks).\n\nExample questions taken from \"What is the GLUE Benchmark\" medium post - https://angelina-yang.medium.com/what-is-the-glue-benchmark-for-nlu-systems-61127b3cab3f",
|
15 |
"prompts": [
|
16 |
{
|
17 |
+
"name": "mistral_7b_instruct_v0.1",
|
18 |
"input": "{{SST_2_ex_prompt}}",
|
19 |
"metadata": {
|
20 |
"model": {
|
21 |
"name": "Text Generation",
|
22 |
"settings": {
|
23 |
+
"model": "mistralai/Mistral-7B-Instruct-v0.1"
|
24 |
}
|
25 |
},
|
26 |
"tags": null,
|