ilia 4b9ffb5ddf docs: Update architecture and add new documentation for LLM and MCP
- Enhanced `ARCHITECTURE.md` with details on LLM models for work (Llama 3.1 70B Q4) and family agents (Phi-3 Mini 3.8B Q4).
- Introduced new documents:
  - `ASR_EVALUATION.md` for ASR engine evaluation and selection.
  - `HARDWARE.md` outlining hardware requirements and purchase plans.
  - `IMPLEMENTATION_GUIDE.md` for Milestone 2 implementation steps.
  - `LLM_CAPACITY.md` assessing VRAM and context window limits.
  - `LLM_MODEL_SURVEY.md` surveying open-weight LLM models.
  - `LLM_USAGE_AND_COSTS.md` detailing LLM usage and operational costs.
  - `MCP_ARCHITECTURE.md` describing the Model Context Protocol architecture.
  - `MCP_IMPLEMENTATION_SUMMARY.md` summarizing MCP implementation status.

These updates provide comprehensive guidance for the next phases of development and ensure clarity in project documentation.
2026-01-05 23:44:16 -05:00

28 lines
608 B
Bash
Executable File

#!/bin/bash
# Setup script for 4080 LLM Server
set -e
echo "Setting up 4080 LLM Server (Work Agent)..."
# Check if Ollama is installed
if ! command -v ollama &> /dev/null; then
echo "Installing Ollama..."
curl -fsSL https://ollama.com/install.sh | sh
else
echo "Ollama is already installed"
fi
# Download model
echo "Downloading Llama 3.1 70B Q4 model..."
ollama pull llama3.1:70b-q4_0
echo "Setup complete!"
echo ""
echo "To start the server:"
echo " ollama serve"
echo ""
echo "Or use systemd service:"
echo " sudo systemctl enable ollama-4080"
echo " sudo systemctl start ollama-4080"