From d9f2364d0921b27615cc3104ab4c95d3c1860be6 Mon Sep 17 00:00:00 2001 From: Pierce Kelaita Date: Mon, 30 Sep 2024 10:57:46 -0700 Subject: [PATCH] [models] update groq models --- l2m2/model_info.py | 80 ++++++++++++++++++++++++++++++++++++++++++++-- 1 file changed, 78 insertions(+), 2 deletions(-) diff --git a/l2m2/model_info.py b/l2m2/model_info.py index 54b1fe0..65a50be 100644 --- a/l2m2/model_info.py +++ b/l2m2/model_info.py @@ -455,9 +455,25 @@ class ModelEntry(TypedDict): "extras": {}, }, }, + "gemma-2-9b": { + "groq": { + "model_id": "gemma2-9b-it", + "params": { + "temperature": { + "default": PROVIDER_DEFAULT, + "max": 2.0, + }, + "max_tokens": { + "default": PROVIDER_DEFAULT, + "max": 2**16 - 1, + }, + }, + "extras": {}, + }, + }, "llama-3-8b": { "groq": { - "model_id": "llama-3-8b-8192", + "model_id": "llama3-8b-8192", "params": { "temperature": { "default": PROVIDER_DEFAULT, @@ -488,7 +504,7 @@ class ModelEntry(TypedDict): }, "llama-3-70b": { "groq": { - "model_id": "llama-3-70b-8192", + "model_id": "llama3-70b-8192", "params": { "temperature": { "default": PROVIDER_DEFAULT, @@ -546,6 +562,20 @@ class ModelEntry(TypedDict): }, "extras": {}, }, + "groq": { + "model_id": "llama-3.1-8b-instant", + "params": { + "temperature": { + "default": PROVIDER_DEFAULT, + "max": 2.0, + }, + "max_tokens": { + "default": PROVIDER_DEFAULT, + "max": 8000, + }, + }, + "extras": {}, + }, }, "llama-3.1-70b": { "octoai": { @@ -562,6 +592,20 @@ class ModelEntry(TypedDict): }, "extras": {}, }, + "groq": { + "model_id": "llama-3.1-70b-versatile", + "params": { + "temperature": { + "default": PROVIDER_DEFAULT, + "max": 2.0, + }, + "max_tokens": { + "default": PROVIDER_DEFAULT, + "max": 8000, + }, + }, + "extras": {}, + }, }, "llama-3.1-405b": { "replicate": { @@ -594,4 +638,36 @@ class ModelEntry(TypedDict): "extras": {}, }, }, + "llama-3.2-1b": { + "groq": { + "model_id": "llama-3.2-1b-preview", + "params": { + "temperature": { + "default": PROVIDER_DEFAULT, + "max": 2.0, + }, + "max_tokens": { + "default": PROVIDER_DEFAULT, + "max": 2**13, + }, + }, + "extras": {"preview": True}, + }, + }, + "llama-3.2-3b": { + "groq": { + "model_id": "llama-3.2-3b-preview", + "params": { + "temperature": { + "default": PROVIDER_DEFAULT, + "max": 2.0, + }, + "max_tokens": { + "default": PROVIDER_DEFAULT, + "max": 2**13, + }, + }, + "extras": {"preview": True}, + }, + }, }