3 Commits

Author SHA1 Message Date
70dc396fe3 fix: remove hardcoded secrets from docker-compose, use env vars + profiles
- Remove hardcoded OpenRouter API key and URL from docker-compose.yml
- App service now reads OPENAI_* vars from .env file (env_file) and
  falls back to http://ollama:11434/v1 defaults
- Ollama and model-init moved to 'ollama' Docker Compose profile,
  so they only start when explicitly requested:
    docker compose --profile ollama up      # with local Ollama
    docker compose up                         # cloud provider only
- Port mapping uses 5656 from .env
- .env.docker updated with documented options for Ollama vs OpenRouter
2026-04-13 01:06:23 -04:00
86d399a04b fix: use npm i instead of npm ci in Docker build, expose on port 5656
- Dockerfile: drop package-lock.json copy, use npm i instead of npm ci
  so install works even if lock file is slightly out of sync
- docker-compose: map host port 5656 → container port 3000
2026-04-13 00:43:19 -04:00
792fafc661 feat: containerize with Docker Compose + Ollama
Add full Docker setup so the app runs with a single 'docker compose up':

- Dockerfile: multi-stage build (node:22-alpine) for the SvelteKit app
- docker-compose.yml: three services:
  1. ollama: runs Ollama server with persistent volume for models
  2. model-init: one-shot container that pulls the configured model
     after Ollama is healthy, then exits
  3. app: the SvelteKit app, starts only after model-init succeeds
- .env.docker: set OLLAMA_MODEL to control which model is pulled
- .dockerignore: keeps image lean
- Switched adapter-auto to adapter-node (required for Docker/Node hosting)
- Updated README with Docker and local dev instructions

Usage:
  docker compose up              # default: llama3
  OLLAMA_MODEL=gemma2 docker compose up  # any Ollama model
2026-04-13 00:22:19 -04:00