|
@@ -83,6 +83,37 @@ model_cards = {
|
|
|
"dummy": { "layers": 8, "repo": { "DummyInferenceEngine": "dummy", }, },
|
|
|
}
|
|
|
|
|
|
+pretty_name = {
|
|
|
+ "llama-3.2-1b": "Llama 3.2 1B",
|
|
|
+ "llama-3.2-3b": "Llama 3.2 3B",
|
|
|
+ "llama-3.1-8b": "Llama 3.1 8B",
|
|
|
+ "llama-3.1-70b": "Llama 3.1 70B",
|
|
|
+ "llama-3.1-70b-bf16": "Llama 3.1 70B (BF16)",
|
|
|
+ "llama-3.1-405b": "Llama 3.1 405B",
|
|
|
+ "llama-3.1-405b-8bit": "Llama 3.1 405B (8-bit)",
|
|
|
+ "gemma2-9b": "Gemma2 9B",
|
|
|
+ "gemma2-27b": "Gemma2 27B",
|
|
|
+ "nemotron-70b": "Nemotron 70B",
|
|
|
+ "nemotron-70b-bf16": "Nemotron 70B (BF16)",
|
|
|
+ "mistral-nemo": "Mistral Nemo",
|
|
|
+ "mistral-large": "Mistral Large",
|
|
|
+ "deepseek-coder-v2-lite": "Deepseek Coder V2 Lite",
|
|
|
+ "deepseek-coder-v2.5": "Deepseek Coder V2.5",
|
|
|
+ "llava-1.5-7b-hf": "LLaVa 1.5 7B (Vision Model)",
|
|
|
+ "qwen-2.5-coder-1.5b": "Qwen 2.5 Coder 1.5B",
|
|
|
+ "qwen-2.5-coder-3b": "Qwen 2.5 Coder 3B",
|
|
|
+ "qwen-2.5-coder-7b": "Qwen 2.5 Coder 7B",
|
|
|
+ "qwen-2.5-coder-14b": "Qwen 2.5 Coder 14B",
|
|
|
+ "qwen-2.5-coder-32b": "Qwen 2.5 Coder 32B",
|
|
|
+ "qwen-2.5-7b": "Qwen 2.5 7B",
|
|
|
+ "qwen-2.5-math-7b": "Qwen 2.5 7B (Math)",
|
|
|
+ "qwen-2.5-14b": "Qwen 2.5 14B",
|
|
|
+ "qwen-2.5-72b": "Qwen 2.5 72B",
|
|
|
+ "qwen-2.5-math-72b": "Qwen 2.5 72B (Math)",
|
|
|
+ "llama-3-8b": "Llama 3 8B",
|
|
|
+ "llama-3-70b": "Llama 3 70B",
|
|
|
+}
|
|
|
+
|
|
|
def get_repo(model_id: str, inference_engine_classname: str) -> Optional[str]:
|
|
|
return model_cards.get(model_id, {}).get("repo", {}).get(inference_engine_classname, None)
|
|
|
|