add dashscope support
This commit is contained in:
parent
ee6c13212c
commit
8499dbf132
@ -52,6 +52,7 @@ class ProvidersConfig(BaseModel):
|
|||||||
openrouter: ProviderConfig = Field(default_factory=ProviderConfig)
|
openrouter: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
groq: ProviderConfig = Field(default_factory=ProviderConfig)
|
groq: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
zhipu: ProviderConfig = Field(default_factory=ProviderConfig)
|
zhipu: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
|
dashscope: ProviderConfig = Field(default_factory=ProviderConfig) # 阿里云通义千问
|
||||||
vllm: ProviderConfig = Field(default_factory=ProviderConfig)
|
vllm: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
gemini: ProviderConfig = Field(default_factory=ProviderConfig)
|
gemini: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
|
|
||||||
@ -92,13 +93,14 @@ class Config(BaseSettings):
|
|||||||
return Path(self.agents.defaults.workspace).expanduser()
|
return Path(self.agents.defaults.workspace).expanduser()
|
||||||
|
|
||||||
def get_api_key(self) -> str | None:
|
def get_api_key(self) -> str | None:
|
||||||
"""Get API key in priority order: OpenRouter > Anthropic > OpenAI > Gemini > Zhipu > Groq > vLLM."""
|
"""Get API key in priority order: OpenRouter > Anthropic > OpenAI > Gemini > Zhipu > DashScope > Groq > vLLM."""
|
||||||
return (
|
return (
|
||||||
self.providers.openrouter.api_key or
|
self.providers.openrouter.api_key or
|
||||||
self.providers.anthropic.api_key or
|
self.providers.anthropic.api_key or
|
||||||
self.providers.openai.api_key or
|
self.providers.openai.api_key or
|
||||||
self.providers.gemini.api_key or
|
self.providers.gemini.api_key or
|
||||||
self.providers.zhipu.api_key or
|
self.providers.zhipu.api_key or
|
||||||
|
self.providers.dashscope.api_key or
|
||||||
self.providers.groq.api_key or
|
self.providers.groq.api_key or
|
||||||
self.providers.vllm.api_key or
|
self.providers.vllm.api_key or
|
||||||
None
|
None
|
||||||
|
|||||||
@ -51,6 +51,8 @@ class LiteLLMProvider(LLMProvider):
|
|||||||
os.environ.setdefault("GEMINI_API_KEY", api_key)
|
os.environ.setdefault("GEMINI_API_KEY", api_key)
|
||||||
elif "zhipu" in default_model or "glm" in default_model or "zai" in default_model:
|
elif "zhipu" in default_model or "glm" in default_model or "zai" in default_model:
|
||||||
os.environ.setdefault("ZHIPUAI_API_KEY", api_key)
|
os.environ.setdefault("ZHIPUAI_API_KEY", api_key)
|
||||||
|
elif "dashscope" in default_model or "qwen" in default_model.lower():
|
||||||
|
os.environ.setdefault("DASHSCOPE_API_KEY", api_key)
|
||||||
elif "groq" in default_model:
|
elif "groq" in default_model:
|
||||||
os.environ.setdefault("GROQ_API_KEY", api_key)
|
os.environ.setdefault("GROQ_API_KEY", api_key)
|
||||||
|
|
||||||
@ -96,6 +98,13 @@ class LiteLLMProvider(LLMProvider):
|
|||||||
):
|
):
|
||||||
model = f"zhipu/{model}"
|
model = f"zhipu/{model}"
|
||||||
|
|
||||||
|
# For DashScope/Qwen, ensure prefix is present
|
||||||
|
if ("qwen" in model.lower() or "dashscope" in model.lower()) and not (
|
||||||
|
model.startswith("dashscope/") or
|
||||||
|
model.startswith("openrouter/")
|
||||||
|
):
|
||||||
|
model = f"dashscope/{model}"
|
||||||
|
|
||||||
# For vLLM, use hosted_vllm/ prefix per LiteLLM docs
|
# For vLLM, use hosted_vllm/ prefix per LiteLLM docs
|
||||||
# Convert openai/ prefix to hosted_vllm/ if user specified it
|
# Convert openai/ prefix to hosted_vllm/ if user specified it
|
||||||
if self.is_vllm:
|
if self.is_vllm:
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user