From e99b56028a887fcb48b1029b4d70293fc737078a Mon Sep 17 00:00:00 2001 From: BHUVANAN8 Date: Sat, 28 Mar 2026 01:24:17 +0530 Subject: [PATCH 1/2] feat(scripts): add support for more LLM providers in check_llm_key.py --- scripts/check_llm_key.py | 15 +++++++++++++++ 1 file changed, 15 insertions(+) diff --git a/scripts/check_llm_key.py b/scripts/check_llm_key.py index 44c24c0275..fc93b02428 100644 --- a/scripts/check_llm_key.py +++ b/scripts/check_llm_key.py @@ -318,6 +318,21 @@ def check_gemini(api_key: str, **_: str) -> dict: key, "https://api.cerebras.ai/v1/models", "Cerebras" ), "openrouter": lambda key, **kw: check_openrouter(key, **kw), + "deepseek": lambda key, **kw: check_openai_compatible( + key, "https://api.deepseek.com/v1/models", "DeepSeek" + ), + "together": lambda key, **kw: check_openai_compatible( + key, "https://api.together.xyz/v1/models", "Together AI" + ), + "mistral": lambda key, **kw: check_openai_compatible( + key, "https://api.mistral.ai/v1/models", "Mistral" + ), + "xai": lambda key, **kw: check_openai_compatible( + key, "https://api.x.ai/v1/models", "xAI" + ), + "perplexity": lambda key, **kw: check_openai_compatible( + key, "https://api.perplexity.ai/models", "Perplexity" + ), "minimax": lambda key, **kw: check_minimax(key), # Kimi For Coding uses an Anthropic-compatible endpoint; check via /v1/messages # with empty messages (same as check_anthropic, triggers 400 not 401). From 09016903ccc936fe97f62db412045603c7313626 Mon Sep 17 00:00:00 2001 From: BHUVANAN8 Date: Sat, 28 Mar 2026 23:14:52 +0530 Subject: [PATCH 2/2] fix(scripts): correct perplexity endpoint to /v1/models and simplify lambda kwargs to **_ --- scripts/check_llm_key.py | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/scripts/check_llm_key.py b/scripts/check_llm_key.py index fc93b02428..7ee77c81ff 100644 --- a/scripts/check_llm_key.py +++ b/scripts/check_llm_key.py @@ -318,22 +318,22 @@ def check_gemini(api_key: str, **_: str) -> dict: key, "https://api.cerebras.ai/v1/models", "Cerebras" ), "openrouter": lambda key, **kw: check_openrouter(key, **kw), - "deepseek": lambda key, **kw: check_openai_compatible( + "deepseek": lambda key, **_: check_openai_compatible( key, "https://api.deepseek.com/v1/models", "DeepSeek" ), - "together": lambda key, **kw: check_openai_compatible( + "together": lambda key, **_: check_openai_compatible( key, "https://api.together.xyz/v1/models", "Together AI" ), - "mistral": lambda key, **kw: check_openai_compatible( + "mistral": lambda key, **_: check_openai_compatible( key, "https://api.mistral.ai/v1/models", "Mistral" ), - "xai": lambda key, **kw: check_openai_compatible( + "xai": lambda key, **_: check_openai_compatible( key, "https://api.x.ai/v1/models", "xAI" ), - "perplexity": lambda key, **kw: check_openai_compatible( - key, "https://api.perplexity.ai/models", "Perplexity" + "perplexity": lambda key, **_: check_openai_compatible( + key, "https://api.perplexity.ai/v1/models", "Perplexity" ), - "minimax": lambda key, **kw: check_minimax(key), + "minimax": lambda key, **_: check_minimax(key), # Kimi For Coding uses an Anthropic-compatible endpoint; check via /v1/messages # with empty messages (same as check_anthropic, triggers 400 not 401). "kimi": lambda key, **kw: check_anthropic_compatible(