resolve conflicts: keep both Gemini and Zhipu support
This commit is contained in:
commit
a0950cf6b7
1
.gitignore
vendored
1
.gitignore
vendored
@ -3,6 +3,7 @@
|
|||||||
*.pyc
|
*.pyc
|
||||||
dist/
|
dist/
|
||||||
build/
|
build/
|
||||||
|
docs/
|
||||||
*.egg-info/
|
*.egg-info/
|
||||||
*.egg
|
*.egg
|
||||||
*.pyc
|
*.pyc
|
||||||
|
|||||||
26
README.md
26
README.md
@ -321,7 +321,11 @@ nanobot/
|
|||||||
└── cli/ # 🖥️ Commands
|
└── cli/ # 🖥️ Commands
|
||||||
```
|
```
|
||||||
|
|
||||||
## 🗺️ Roadmap
|
## 🤝 Contribute & Roadmap
|
||||||
|
|
||||||
|
PRs welcome! The codebase is intentionally small and readable. 🤗
|
||||||
|
|
||||||
|
**Roadmap** — Pick an item and [open a PR](https://github.com/HKUDS/nanobot/pulls)!
|
||||||
|
|
||||||
- [ ] **Multi-modal** — See and hear (images, voice, video)
|
- [ ] **Multi-modal** — See and hear (images, voice, video)
|
||||||
- [ ] **Long-term memory** — Never forget important context
|
- [ ] **Long-term memory** — Never forget important context
|
||||||
@ -329,11 +333,25 @@ nanobot/
|
|||||||
- [ ] **More integrations** — Discord, Slack, email, calendar
|
- [ ] **More integrations** — Discord, Slack, email, calendar
|
||||||
- [ ] **Self-improvement** — Learn from feedback and mistakes
|
- [ ] **Self-improvement** — Learn from feedback and mistakes
|
||||||
|
|
||||||
**Want to help?** Pick an item and [open a PR](https://github.com/HKUDS/nanobot/pulls)!
|
### Contributors
|
||||||
|
|
||||||
## 🤝 Contribute
|
<a href="https://github.com/HKUDS/nanobot/graphs/contributors">
|
||||||
|
<img src="https://contrib.rocks/image?repo=HKUDS/nanobot" />
|
||||||
|
</a>
|
||||||
|
|
||||||
PRs welcome! The codebase is intentionally small and readable. 🤗
|
---
|
||||||
|
|
||||||
|
## ⭐ Star History
|
||||||
|
|
||||||
|
<div align="center">
|
||||||
|
<a href="https://star-history.com/#HKUDS/nanobot&Date">
|
||||||
|
<picture>
|
||||||
|
<source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=HKUDS/nanobot&type=Date&theme=dark" />
|
||||||
|
<source media="(prefers-color-scheme: light)" srcset="https://api.star-history.com/svg?repos=HKUDS/nanobot&type=Date" />
|
||||||
|
<img alt="Star History Chart" src="https://api.star-history.com/svg?repos=HKUDS/nanobot&type=Date" style="border-radius: 15px; box-shadow: 0 0 30px rgba(0, 217, 255, 0.3);" />
|
||||||
|
</picture>
|
||||||
|
</a>
|
||||||
|
</div>
|
||||||
|
|
||||||
<p align="center">
|
<p align="center">
|
||||||
<em> Thanks for visiting ✨ nanobot!</em><br><br>
|
<em> Thanks for visiting ✨ nanobot!</em><br><br>
|
||||||
|
|||||||
@ -189,7 +189,8 @@ class AgentLoop:
|
|||||||
|
|
||||||
# Execute tools
|
# Execute tools
|
||||||
for tool_call in response.tool_calls:
|
for tool_call in response.tool_calls:
|
||||||
logger.debug(f"Executing tool: {tool_call.name}")
|
args_str = json.dumps(tool_call.arguments)
|
||||||
|
logger.debug(f"Executing tool: {tool_call.name} with arguments: {args_str}")
|
||||||
result = await self.tools.execute(tool_call.name, tool_call.arguments)
|
result = await self.tools.execute(tool_call.name, tool_call.arguments)
|
||||||
messages = self.context.add_tool_result(
|
messages = self.context.add_tool_result(
|
||||||
messages, tool_call.id, tool_call.name, result
|
messages, tool_call.id, tool_call.name, result
|
||||||
@ -281,7 +282,8 @@ class AgentLoop:
|
|||||||
)
|
)
|
||||||
|
|
||||||
for tool_call in response.tool_calls:
|
for tool_call in response.tool_calls:
|
||||||
logger.debug(f"Executing tool: {tool_call.name}")
|
args_str = json.dumps(tool_call.arguments)
|
||||||
|
logger.debug(f"Executing tool: {tool_call.name} with arguments: {args_str}")
|
||||||
result = await self.tools.execute(tool_call.name, tool_call.arguments)
|
result = await self.tools.execute(tool_call.name, tool_call.arguments)
|
||||||
messages = self.context.add_tool_result(
|
messages = self.context.add_tool_result(
|
||||||
messages, tool_call.id, tool_call.name, result
|
messages, tool_call.id, tool_call.name, result
|
||||||
|
|||||||
@ -154,7 +154,7 @@ This file stores important information that should persist across sessions.
|
|||||||
|
|
||||||
@app.command()
|
@app.command()
|
||||||
def gateway(
|
def gateway(
|
||||||
port: int = typer.Option(18789, "--port", "-p", help="Gateway port"),
|
port: int = typer.Option(18790, "--port", "-p", help="Gateway port"),
|
||||||
verbose: bool = typer.Option(False, "--verbose", "-v", help="Verbose output"),
|
verbose: bool = typer.Option(False, "--verbose", "-v", help="Verbose output"),
|
||||||
):
|
):
|
||||||
"""Start the nanobot gateway."""
|
"""Start the nanobot gateway."""
|
||||||
|
|||||||
@ -50,6 +50,7 @@ class ProvidersConfig(BaseModel):
|
|||||||
anthropic: ProviderConfig = Field(default_factory=ProviderConfig)
|
anthropic: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
openai: ProviderConfig = Field(default_factory=ProviderConfig)
|
openai: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
openrouter: ProviderConfig = Field(default_factory=ProviderConfig)
|
openrouter: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
|
zhipu: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
vllm: ProviderConfig = Field(default_factory=ProviderConfig)
|
vllm: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
gemini: ProviderConfig = Field(default_factory=ProviderConfig)
|
gemini: ProviderConfig = Field(default_factory=ProviderConfig)
|
||||||
|
|
||||||
@ -57,7 +58,7 @@ class ProvidersConfig(BaseModel):
|
|||||||
class GatewayConfig(BaseModel):
|
class GatewayConfig(BaseModel):
|
||||||
"""Gateway/server configuration."""
|
"""Gateway/server configuration."""
|
||||||
host: str = "0.0.0.0"
|
host: str = "0.0.0.0"
|
||||||
port: int = 18789
|
port: int = 18790
|
||||||
|
|
||||||
|
|
||||||
class WebSearchConfig(BaseModel):
|
class WebSearchConfig(BaseModel):
|
||||||
@ -90,20 +91,23 @@ class Config(BaseSettings):
|
|||||||
return Path(self.agents.defaults.workspace).expanduser()
|
return Path(self.agents.defaults.workspace).expanduser()
|
||||||
|
|
||||||
def get_api_key(self) -> str | None:
|
def get_api_key(self) -> str | None:
|
||||||
"""Get API key in priority order: OpenRouter > Anthropic > OpenAI > Gemini > vLLM."""
|
"""Get API key in priority order: OpenRouter > Anthropic > OpenAI > Gemini > Zhipu > vLLM."""
|
||||||
return (
|
return (
|
||||||
self.providers.openrouter.api_key or
|
self.providers.openrouter.api_key or
|
||||||
self.providers.anthropic.api_key or
|
self.providers.anthropic.api_key or
|
||||||
self.providers.openai.api_key or
|
self.providers.openai.api_key or
|
||||||
self.providers.gemini.api_key or
|
self.providers.gemini.api_key or
|
||||||
|
self.providers.zhipu.api_key or
|
||||||
self.providers.vllm.api_key or
|
self.providers.vllm.api_key or
|
||||||
None
|
None
|
||||||
)
|
)
|
||||||
|
|
||||||
def get_api_base(self) -> str | None:
|
def get_api_base(self) -> str | None:
|
||||||
"""Get API base URL if using OpenRouter or vLLM."""
|
"""Get API base URL if using OpenRouter, Zhipu or vLLM."""
|
||||||
if self.providers.openrouter.api_key:
|
if self.providers.openrouter.api_key:
|
||||||
return self.providers.openrouter.api_base or "https://openrouter.ai/api/v1"
|
return self.providers.openrouter.api_base or "https://openrouter.ai/api/v1"
|
||||||
|
if self.providers.zhipu.api_key:
|
||||||
|
return self.providers.zhipu.api_base
|
||||||
if self.providers.vllm.api_base:
|
if self.providers.vllm.api_base:
|
||||||
return self.providers.vllm.api_base
|
return self.providers.vllm.api_base
|
||||||
return None
|
return None
|
||||||
|
|||||||
@ -49,6 +49,8 @@ class LiteLLMProvider(LLMProvider):
|
|||||||
os.environ.setdefault("OPENAI_API_KEY", api_key)
|
os.environ.setdefault("OPENAI_API_KEY", api_key)
|
||||||
elif "gemini" in default_model.lower():
|
elif "gemini" in default_model.lower():
|
||||||
os.environ.setdefault("GEMINI_API_KEY", api_key)
|
os.environ.setdefault("GEMINI_API_KEY", api_key)
|
||||||
|
elif "zhipu" in default_model or "glm" in default_model or "zai" in default_model:
|
||||||
|
os.environ.setdefault("ZHIPUAI_API_KEY", api_key)
|
||||||
|
|
||||||
if api_base:
|
if api_base:
|
||||||
litellm.api_base = api_base
|
litellm.api_base = api_base
|
||||||
@ -83,6 +85,15 @@ class LiteLLMProvider(LLMProvider):
|
|||||||
if self.is_openrouter and not model.startswith("openrouter/"):
|
if self.is_openrouter and not model.startswith("openrouter/"):
|
||||||
model = f"openrouter/{model}"
|
model = f"openrouter/{model}"
|
||||||
|
|
||||||
|
# For Zhipu/Z.ai, ensure prefix is present
|
||||||
|
# Handle cases like "glm-4.7-flash" -> "zhipu/glm-4.7-flash"
|
||||||
|
if ("glm" in model.lower() or "zhipu" in model.lower()) and not (
|
||||||
|
model.startswith("zhipu/") or
|
||||||
|
model.startswith("zai/") or
|
||||||
|
model.startswith("openrouter/")
|
||||||
|
):
|
||||||
|
model = f"zhipu/{model}"
|
||||||
|
|
||||||
# For vLLM, use hosted_vllm/ prefix per LiteLLM docs
|
# For vLLM, use hosted_vllm/ prefix per LiteLLM docs
|
||||||
# Convert openai/ prefix to hosted_vllm/ if user specified it
|
# Convert openai/ prefix to hosted_vllm/ if user specified it
|
||||||
if self.is_vllm:
|
if self.is_vllm:
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user