Some checks failed
CI / backend-test (push) Successful in 4m9s
CI / frontend-test (push) Failing after 3m48s
CI / lint-python (push) Successful in 1m41s
CI / secret-scanning (push) Successful in 1m20s
CI / dependency-scan (push) Successful in 10m50s
CI / workflow-summary (push) Successful in 1m11s
## Features Added
### Document Reference System
- Implemented numbered document references (@1, @2, etc.) with autocomplete dropdown
- Added fuzzy filename matching for @filename references
- Document filtering now prioritizes numeric refs > filename refs > all documents
- Autocomplete dropdown appears when typing @ with keyboard navigation (Up/Down, Enter/Tab, Escape)
- Document numbers displayed in UI for easy reference
### Conversation Management
- Added conversation rename functionality with inline editing
- Implemented conversation search (by title and content)
- Search box always visible, even when no conversations exist
- Export reports now replace @N references with actual filenames
### UI/UX Improvements
- Removed debug toggle button
- Improved text contrast in dark mode (better visibility)
- Made input textarea expand to full available width
- Fixed file text color for better readability
- Enhanced document display with numbered badges
### Configuration & Timeouts
- Made HTTP client timeouts configurable (connect, write, pool)
- Added .env.example with all configuration options
- Updated timeout documentation
### Developer Experience
- Added `make test-setup` target for automated test conversation creation
- Test setup script supports TEST_MESSAGE and TEST_DOCS env vars
- Improved Makefile with dev and test-setup targets
### Documentation
- Updated ARCHITECTURE.md with all new features
- Created comprehensive deployment documentation
- Added GPU VM setup guides
- Removed unnecessary markdown files (CLAUDE.md, CONTRIBUTING.md, header.jpg)
- Organized documentation in docs/ directory
### GPU VM / Ollama (Stability + GPU Offload)
- Updated GPU VM docs to reflect the working systemd environment for remote Ollama
- Standardized remote Ollama port to 11434 (and added /v1/models verification)
- Documented required env for GPU offload on this VM:
- `OLLAMA_MODELS=/mnt/data/ollama`, `HOME=/mnt/data/ollama/home`
- `OLLAMA_LLM_LIBRARY=cuda_v12` (not `cuda`)
- `LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12`
## Technical Changes
### Backend
- Enhanced `docs_context.py` with reference parsing (numeric and filename)
- Added `update_conversation_title` to storage.py
- New endpoints: PATCH /api/conversations/{id}/title, GET /api/conversations/search
- Improved report generation with filename substitution
### Frontend
- Removed debugMode state and related code
- Added autocomplete dropdown component
- Implemented search functionality in Sidebar
- Enhanced ChatInterface with autocomplete and improved textarea sizing
- Updated CSS for better contrast and responsive design
## Files Changed
- Backend: config.py, council.py, docs_context.py, main.py, storage.py
- Frontend: App.jsx, ChatInterface.jsx, Sidebar.jsx, and related CSS files
- Documentation: README.md, ARCHITECTURE.md, new docs/ directory
- Configuration: .env.example, Makefile
- Scripts: scripts/test_setup.py
## Breaking Changes
None - all changes are backward compatible
## Testing
- All existing tests pass
- New test-setup script validates conversation creation workflow
- Manual testing of autocomplete, search, and rename features
88 lines
2.9 KiB
Python
88 lines
2.9 KiB
Python
import unittest
|
|
|
|
import httpx
|
|
import json
|
|
|
|
from backend.openai_compat import _resolve_chat_completions_url, _resolve_models_url, query_model, list_models
|
|
|
|
|
|
class TestOpenAICompatUrl(unittest.TestCase):
|
|
def test_resolve_url_when_no_v1(self):
|
|
self.assertEqual(
|
|
_resolve_chat_completions_url("http://gpu:8000"),
|
|
"http://gpu:8000/v1/chat/completions",
|
|
)
|
|
|
|
def test_resolve_url_when_v1(self):
|
|
self.assertEqual(
|
|
_resolve_chat_completions_url("http://gpu:8000/v1"),
|
|
"http://gpu:8000/v1/chat/completions",
|
|
)
|
|
|
|
def test_resolve_url_when_v1_with_trailing_slash(self):
|
|
self.assertEqual(
|
|
_resolve_chat_completions_url("http://gpu:8000/v1/"),
|
|
"http://gpu:8000/v1/chat/completions",
|
|
)
|
|
|
|
def test_resolve_models_url(self):
|
|
self.assertEqual(
|
|
_resolve_models_url("http://gpu:8000"),
|
|
"http://gpu:8000/v1/models",
|
|
)
|
|
|
|
|
|
class TestOpenAICompatRequest(unittest.IsolatedAsyncioTestCase):
|
|
async def test_query_model_builds_payload_and_parses_response(self):
|
|
captured = {}
|
|
|
|
def handler(request: httpx.Request) -> httpx.Response:
|
|
captured["url"] = str(request.url)
|
|
captured["auth"] = request.headers.get("authorization")
|
|
captured["json"] = json.loads(request.content.decode("utf-8"))
|
|
return httpx.Response(
|
|
200,
|
|
json={
|
|
"choices": [
|
|
{
|
|
"message": {"content": "hello", "reasoning_details": None},
|
|
}
|
|
]
|
|
},
|
|
)
|
|
|
|
transport = httpx.MockTransport(handler)
|
|
async with httpx.AsyncClient(transport=transport, timeout=10.0) as client:
|
|
out = await query_model(
|
|
"my-model",
|
|
[{"role": "user", "content": "hi"}],
|
|
base_url="http://gpu:8000",
|
|
api_key="secret",
|
|
max_tokens=123,
|
|
timeout=10.0,
|
|
client=client,
|
|
)
|
|
|
|
self.assertEqual(captured["url"], "http://gpu:8000/v1/chat/completions")
|
|
self.assertEqual(captured["auth"], "Bearer secret")
|
|
self.assertEqual(captured["json"]["model"], "my-model")
|
|
self.assertEqual(captured["json"]["max_tokens"], 123)
|
|
self.assertEqual(out["content"], "hello")
|
|
|
|
async def test_list_models_parses_ids(self):
|
|
def handler(request: httpx.Request) -> httpx.Response:
|
|
return httpx.Response(
|
|
200,
|
|
json={"data": [{"id": "a"}, {"id": "b"}, {"nope": "c"}]},
|
|
)
|
|
|
|
transport = httpx.MockTransport(handler)
|
|
async with httpx.AsyncClient(transport=transport, timeout=10.0) as client:
|
|
ids = await list_models(
|
|
base_url="http://gpu:8000",
|
|
client=client,
|
|
)
|
|
self.assertEqual(ids, ["a", "b"])
|
|
|
|
|