From e680b734b1711f2d7efe38017ea5fd6b8877265c Mon Sep 17 00:00:00 2001 From: mengjiechen Date: Fri, 6 Feb 2026 15:15:15 +0800 Subject: [PATCH 1/2] feat: add Moonshot provider support - Add moonshot to ProvidersConfig schema - Add MOONSHOT_API_BASE environment variable for custom endpoint - Handle kimi-k2.5 model temperature restriction (must be 1.0) - Fix is_vllm detection to exclude moonshot provider Co-Authored-By: Claude Opus 4.6 --- nanobot/config/schema.py | 8 ++++-- nanobot/providers/litellm_provider.py | 40 ++++++++++++++++++++------- 2 files changed, 36 insertions(+), 12 deletions(-) diff --git a/nanobot/config/schema.py b/nanobot/config/schema.py index 7f8c495..5e8e46c 100644 --- a/nanobot/config/schema.py +++ b/nanobot/config/schema.py @@ -77,6 +77,7 @@ class ProvidersConfig(BaseModel): zhipu: ProviderConfig = Field(default_factory=ProviderConfig) vllm: ProviderConfig = Field(default_factory=ProviderConfig) gemini: ProviderConfig = Field(default_factory=ProviderConfig) + moonshot: ProviderConfig = Field(default_factory=ProviderConfig) class GatewayConfig(BaseModel): @@ -122,7 +123,7 @@ class Config(BaseSettings): return Path(self.agents.defaults.workspace).expanduser() def get_api_key(self) -> str | None: - """Get API key in priority order: OpenRouter > DeepSeek > Anthropic > OpenAI > Gemini > Zhipu > Groq > vLLM.""" + """Get API key in priority order: OpenRouter > DeepSeek > Anthropic > OpenAI > Gemini > Zhipu > Groq > Moonshot > vLLM.""" return ( self.providers.openrouter.api_key or self.providers.deepseek.api_key or @@ -131,16 +132,19 @@ class Config(BaseSettings): self.providers.gemini.api_key or self.providers.zhipu.api_key or self.providers.groq.api_key or + self.providers.moonshot.api_key or self.providers.vllm.api_key or None ) def get_api_base(self) -> str | None: - """Get API base URL if using OpenRouter, Zhipu or vLLM.""" + """Get API base URL if using OpenRouter, Zhipu, Moonshot or vLLM.""" if self.providers.openrouter.api_key: return self.providers.openrouter.api_base or "https://openrouter.ai/api/v1" if self.providers.zhipu.api_key: return self.providers.zhipu.api_base + if self.providers.moonshot.api_key: + return self.providers.moonshot.api_base if self.providers.vllm.api_base: return self.providers.vllm.api_base return None diff --git a/nanobot/providers/litellm_provider.py b/nanobot/providers/litellm_provider.py index d010d81..c2cdda7 100644 --- a/nanobot/providers/litellm_provider.py +++ b/nanobot/providers/litellm_provider.py @@ -31,9 +31,15 @@ class LiteLLMProvider(LLMProvider): (api_key and api_key.startswith("sk-or-")) or (api_base and "openrouter" in api_base) ) - + + # Detect Moonshot by api_base or model name + self.is_moonshot = ( + (api_base and "moonshot" in api_base) or + ("moonshot" in default_model or "kimi" in default_model) + ) + # Track if using custom endpoint (vLLM, etc.) - self.is_vllm = bool(api_base) and not self.is_openrouter + self.is_vllm = bool(api_base) and not self.is_openrouter and not self.is_moonshot # Configure LiteLLM based on provider if api_key: @@ -55,8 +61,12 @@ class LiteLLMProvider(LLMProvider): os.environ.setdefault("ZHIPUAI_API_KEY", api_key) elif "groq" in default_model: os.environ.setdefault("GROQ_API_KEY", api_key) - - if api_base: + elif "moonshot" in default_model or "kimi" in default_model: + os.environ.setdefault("MOONSHOT_API_KEY", api_key) + if api_base: + os.environ["MOONSHOT_API_BASE"] = api_base + + if api_base and not self.is_moonshot: litellm.api_base = api_base # Disable LiteLLM logging noise @@ -97,23 +107,33 @@ class LiteLLMProvider(LLMProvider): model.startswith("openrouter/") ): model = f"zai/{model}" - + + # For Moonshot/Kimi, ensure moonshot/ prefix (before vLLM check) + if ("moonshot" in model.lower() or "kimi" in model.lower()) and not ( + model.startswith("moonshot/") or model.startswith("openrouter/") + ): + model = f"moonshot/{model}" + + # For Gemini, ensure gemini/ prefix if not already present + if "gemini" in model.lower() and not model.startswith("gemini/"): + model = f"gemini/{model}" + # For vLLM, use hosted_vllm/ prefix per LiteLLM docs # Convert openai/ prefix to hosted_vllm/ if user specified it if self.is_vllm: model = f"hosted_vllm/{model}" - # For Gemini, ensure gemini/ prefix if not already present - if "gemini" in model.lower() and not model.startswith("gemini/"): - model = f"gemini/{model}" - kwargs: dict[str, Any] = { "model": model, "messages": messages, "max_tokens": max_tokens, "temperature": temperature, } - + + # kimi-k2.5 only supports temperature=1.0 + if "kimi-k2.5" in model.lower(): + kwargs["temperature"] = 1.0 + # Pass api_base directly for custom endpoints (vLLM, etc.) if self.api_base: kwargs["api_base"] = self.api_base From 760a369004dcd22e5468f033deb6d43889551da1 Mon Sep 17 00:00:00 2001 From: Re-bin Date: Fri, 6 Feb 2026 08:01:20 +0000 Subject: [PATCH 2/2] feat: fix API key matching by model name --- nanobot/config/schema.py | 69 +++++++++++++++++++-------- nanobot/providers/litellm_provider.py | 27 ++++------- 2 files changed, 58 insertions(+), 38 deletions(-) diff --git a/nanobot/config/schema.py b/nanobot/config/schema.py index 5e8e46c..353ca4b 100644 --- a/nanobot/config/schema.py +++ b/nanobot/config/schema.py @@ -122,30 +122,57 @@ class Config(BaseSettings): """Get expanded workspace path.""" return Path(self.agents.defaults.workspace).expanduser() - def get_api_key(self) -> str | None: - """Get API key in priority order: OpenRouter > DeepSeek > Anthropic > OpenAI > Gemini > Zhipu > Groq > Moonshot > vLLM.""" - return ( - self.providers.openrouter.api_key or - self.providers.deepseek.api_key or - self.providers.anthropic.api_key or - self.providers.openai.api_key or - self.providers.gemini.api_key or - self.providers.zhipu.api_key or - self.providers.groq.api_key or - self.providers.moonshot.api_key or - self.providers.vllm.api_key or - None - ) + def _match_provider(self, model: str | None = None) -> ProviderConfig | None: + """Match a provider based on model name.""" + model = (model or self.agents.defaults.model).lower() + # Map of keywords to provider configs + providers = { + "openrouter": self.providers.openrouter, + "deepseek": self.providers.deepseek, + "anthropic": self.providers.anthropic, + "claude": self.providers.anthropic, + "openai": self.providers.openai, + "gpt": self.providers.openai, + "gemini": self.providers.gemini, + "zhipu": self.providers.zhipu, + "glm": self.providers.zhipu, + "zai": self.providers.zhipu, + "groq": self.providers.groq, + "moonshot": self.providers.moonshot, + "kimi": self.providers.moonshot, + "vllm": self.providers.vllm, + } + for keyword, provider in providers.items(): + if keyword in model and provider.api_key: + return provider + return None + + def get_api_key(self, model: str | None = None) -> str | None: + """Get API key for the given model (or default model). Falls back to first available key.""" + # Try matching by model name first + matched = self._match_provider(model) + if matched: + return matched.api_key + # Fallback: return first available key + for provider in [ + self.providers.openrouter, self.providers.deepseek, + self.providers.anthropic, self.providers.openai, + self.providers.gemini, self.providers.zhipu, + self.providers.moonshot, self.providers.vllm, + self.providers.groq, + ]: + if provider.api_key: + return provider.api_key + return None - def get_api_base(self) -> str | None: - """Get API base URL if using OpenRouter, Zhipu, Moonshot or vLLM.""" - if self.providers.openrouter.api_key: + def get_api_base(self, model: str | None = None) -> str | None: + """Get API base URL based on model name.""" + model = (model or self.agents.defaults.model).lower() + if "openrouter" in model: return self.providers.openrouter.api_base or "https://openrouter.ai/api/v1" - if self.providers.zhipu.api_key: + if any(k in model for k in ("zhipu", "glm", "zai")): return self.providers.zhipu.api_base - if self.providers.moonshot.api_key: - return self.providers.moonshot.api_base - if self.providers.vllm.api_base: + if "vllm" in model: return self.providers.vllm.api_base return None diff --git a/nanobot/providers/litellm_provider.py b/nanobot/providers/litellm_provider.py index c2cdda7..2125b15 100644 --- a/nanobot/providers/litellm_provider.py +++ b/nanobot/providers/litellm_provider.py @@ -31,15 +31,9 @@ class LiteLLMProvider(LLMProvider): (api_key and api_key.startswith("sk-or-")) or (api_base and "openrouter" in api_base) ) - - # Detect Moonshot by api_base or model name - self.is_moonshot = ( - (api_base and "moonshot" in api_base) or - ("moonshot" in default_model or "kimi" in default_model) - ) - + # Track if using custom endpoint (vLLM, etc.) - self.is_vllm = bool(api_base) and not self.is_openrouter and not self.is_moonshot + self.is_vllm = bool(api_base) and not self.is_openrouter # Configure LiteLLM based on provider if api_key: @@ -63,10 +57,9 @@ class LiteLLMProvider(LLMProvider): os.environ.setdefault("GROQ_API_KEY", api_key) elif "moonshot" in default_model or "kimi" in default_model: os.environ.setdefault("MOONSHOT_API_KEY", api_key) - if api_base: - os.environ["MOONSHOT_API_BASE"] = api_base - - if api_base and not self.is_moonshot: + os.environ.setdefault("MOONSHOT_API_BASE", api_base or "https://api.moonshot.cn/v1") + + if api_base: litellm.api_base = api_base # Disable LiteLLM logging noise @@ -123,17 +116,17 @@ class LiteLLMProvider(LLMProvider): if self.is_vllm: model = f"hosted_vllm/{model}" + # kimi-k2.5 only supports temperature=1.0 + if "kimi-k2.5" in model.lower(): + temperature = 1.0 + kwargs: dict[str, Any] = { "model": model, "messages": messages, "max_tokens": max_tokens, "temperature": temperature, } - - # kimi-k2.5 only supports temperature=1.0 - if "kimi-k2.5" in model.lower(): - kwargs["temperature"] = 1.0 - + # Pass api_base directly for custom endpoints (vLLM, etc.) if self.api_base: kwargs["api_base"] = self.api_base