Spaces:
Paused
Paused
Remove Gemma 2 from HuggingChat (#1319)
Browse files- chart/env/prod.yaml +0 -28
chart/env/prod.yaml
CHANGED
|
@@ -92,34 +92,6 @@ envVars:
|
|
| 92 |
"top_p" : 0.9
|
| 93 |
}
|
| 94 |
},
|
| 95 |
-
{
|
| 96 |
-
"name" : "google/gemma-2-27b-it",
|
| 97 |
-
"tokenizer": "google/gemma-2-27b-it",
|
| 98 |
-
"description": "Gemma 2 27B is the latest release in the Gemma family of models built by Google, with performance on par with Llama 3 70B. ",
|
| 99 |
-
"modelUrl": "https://huggingface.co/google/gemma-2-27b-it",
|
| 100 |
-
"websiteUrl": "https://huggingface.co/google/gemma-2-27b-it",
|
| 101 |
-
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/google-logo.png",
|
| 102 |
-
"parameters": {
|
| 103 |
-
"truncate" : 6146,
|
| 104 |
-
"max_new_tokens" : 2044,
|
| 105 |
-
"temperature": 0.95,
|
| 106 |
-
"top_k": 100,
|
| 107 |
-
"stop" : ["<end_of_turn>", "<eos>", "<end_of_turn><eos>"]
|
| 108 |
-
},
|
| 109 |
-
"preprompt" : "",
|
| 110 |
-
"promptExamples" : [
|
| 111 |
-
{
|
| 112 |
-
"title": "Write a poem",
|
| 113 |
-
"prompt": "Write a poem to help me remember the first 10 elements on the periodic table, giving each element its own line."
|
| 114 |
-
}, {
|
| 115 |
-
"title": "Code a snake game",
|
| 116 |
-
"prompt": "Code a basic snake game in python, give explanations for each step."
|
| 117 |
-
}, {
|
| 118 |
-
"title": "Assist in a task",
|
| 119 |
-
"prompt": "How do I make a delicious lemon cheesecake?"
|
| 120 |
-
}
|
| 121 |
-
]
|
| 122 |
-
},
|
| 123 |
{
|
| 124 |
"name" : "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 125 |
"description" : "The latest MoE model from Mistral AI! 8x7B and outperforms Llama 2 70B in most benchmarks.",
|
|
|
|
| 92 |
"top_p" : 0.9
|
| 93 |
}
|
| 94 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 95 |
{
|
| 96 |
"name" : "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 97 |
"description" : "The latest MoE model from Mistral AI! 8x7B and outperforms Llama 2 70B in most benchmarks.",
|