20 Commits

Author SHA1 Message Date
Xubin Ren
771c918770
Merge pull request #205 from wcmolin/fix/zhipu-api-key
[Fix-204]: use correct ZAI_API_KEY for Zhipu/GLM models #204
2026-02-07 10:20:10 +08:00
Re-bin
2096645ff1 resolve conflicts with main 2026-02-06 16:56:02 +00:00
Re-bin
760a369004 feat: fix API key matching by model name 2026-02-06 08:01:20 +00:00
wcmolin
fea4a6bba8 fix: use correct ZAI_API_KEY for Zhipu/GLM models
LiteLLM's zai provider reads ZAI_API_KEY, not ZHIPUAI_API_KEY.
This fixes authentication errors when using Zhipu/GLM models.
2026-02-06 15:38:25 +08:00
mengjiechen
e680b734b1 feat: add Moonshot provider support
- Add moonshot to ProvidersConfig schema
- Add MOONSHOT_API_BASE environment variable for custom endpoint
- Handle kimi-k2.5 model temperature restriction (must be 1.0)
- Fix is_vllm detection to exclude moonshot provider

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-06 15:25:58 +08:00
Re-bin
301fba568b refactor: remove redundant env var setting, add DeepSeek to docs 2026-02-05 08:55:41 +00:00
Re-bin
ac45630116 resolve conflicts with main 2026-02-05 08:51:11 +00:00
Xubin Ren
b1bce89d4b
Merge pull request #32 from pjperez/main
fix: Use correct 'zai/' prefix for Zhipu AI models in LiteLLM
2026-02-04 17:10:13 +08:00
Re-bin
eb20cea668 resolve conflicts with main 2026-02-03 06:16:17 +00:00
popcell
8cde0b3072 fix: correct API key environment variable for vLLM mode 2026-02-03 12:14:14 +08:00
Kyya Wang
f23548f296 feat: add DeepSeek provider support 2026-02-03 03:09:13 +00:00
Pedro Perez
43fe83755d fix: Use correct 'zai/' prefix for Zhipu AI models in LiteLLM
LiteLLM expects the 'zai/' provider prefix for Zhipu AI (Z.ai) models,
not 'zhipu/'. This was causing 'LLM Provider NOT provided' errors when
users configured models like 'glm-4.7' without an explicit prefix.

According to LiteLLM docs, the correct format is:
- model='zai/glm-4.7' (correct)
- NOT model='zhipu/glm-4.7' (incorrect)

This fix ensures auto-prefixed models use the correct 'zai/' format.

Fixes: Error when using Zhipu AI models with shorthand names like 'glm-4.7'
2026-02-02 22:15:21 +01:00
Re-bin
a0950cf6b7 resolve conflicts: keep both Gemini and Zhipu support 2026-02-02 13:00:24 +00:00
Re-bin
8d834d9b10 restore CLI commands, keep default model 2026-02-02 12:30:56 +00:00
Manus AI
42f62c0c1a feat: add voice transcription support with groq (fixes #13) 2026-02-02 04:33:26 -05:00
Manus
4f01e3ab93 Resolve merge conflicts in PR #3 2026-02-02 04:23:02 -05:00
Anunay Aatipamula
ab45185ed8 feat: add Gemini provider support
- Update configuration schema to include Gemini provider
- Modify API key retrieval priority to include Gemini
- Enhance CLI status command to display Gemini API status
- Update LiteLLMProvider to support Gemini integration
2026-02-02 11:21:41 +05:30
ZhihaoZhang97
2b19dcf9fd feat: add vLLM/local LLM support
- Add vllm provider configuration in config schema
- Auto-detect vLLM endpoints and use hosted_vllm/ prefix for LiteLLM
- Pass api_base directly to acompletion for custom endpoints
- Add vLLM status display in CLI status command
- Add vLLM setup documentation in README
2026-02-02 11:23:04 +11:00
Manus AI
2dd284661d feat: add Zhipu API support and set glm-4.7-flash as default model 2026-02-01 14:36:15 -05:00
Re-bin
d4cc48afd5 🐈nanobot: hello world! 2026-02-01 07:36:42 +00:00