--- name: docker-infra description: Docker and Docker Compose configuration and troubleshooting. Use for containers, services, volumes, networking, health checks, and deployment patterns. --- # Docker & Infrastructure Skill **Activation:** Docker, Docker Compose, containers, services, deployment ## Development Environment ### Architecture **Pre-bundled frontend**: `astro/dist/` is committed to git. No Vite dev server at runtime. **Runtime containers**: db, backend, nginx **Build-only container**: frontend (used for `just build-astro`) ### Services Overview | Service | Port | Image | Purpose | |---------|------|-------|---------| | nginx | 9000 | nginx:alpine | Routes static files + API proxy | | backend | 8000 | python:3.12-slim | FastAPI with uvicorn --reload | | astro | - | node:24-alpine | Build-only (Astro build) | | worker | - | python:3.12-slim | TaskIQ workers | | db | 5432 | postgres:18-alpine | PostgreSQL database | | redis | 6379 | redis:8.4.0-alpine | Job queue + cache | **Note:** Frontend container is only used for builds (`just build-astro`). At runtime, nginx serves pre-built files from the bind-mounted `astro/dist/` directory. ### Quick Commands ```bash # Start all services docker compose up # Start specific service docker compose up backend # Rebuild after dependency changes docker compose build backend docker compose up backend # View logs docker compose logs -f backend worker # Shell into container docker compose exec backend bash docker compose exec astro sh # Stop all docker compose down # Stop and remove volumes (reset DB) docker compose down -v ``` ## Dockerfile Patterns ### Backend (Python) ```dockerfile # backend/Dockerfile.dev FROM python:3.12-slim WORKDIR /app # System dependencies RUN apt-get update && apt-get install -y --no-install-recommends \ build-essential \ curl \ libsndfile1 \ ffmpeg \ && rm -rf /var/lib/apt/lists/* # Install uv for fast dependency management COPY --from=ghcr.io/astral-sh/uv:latest /uv /usr/local/bin/uv # Copy dependency files COPY backend/pyproject.toml ./ # Install dependencies RUN uv pip install --system -e ".[dev]" # Copy source (volumes override in dev) COPY backend/ . # Run with hot-reload (via entrypoint.sh for migrations) ENTRYPOINT ["/bin/bash", "/app/entrypoint.sh"] CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000", "--reload"] ``` ### Frontend (Node.js) - Build Only The frontend container is used for builds only. No dev server runs at runtime. ```dockerfile # astro/Dockerfile.dev FROM node:24-alpine WORKDIR /app # Install pnpm RUN corepack enable && corepack prepare pnpm@latest --activate # Copy dependency files COPY package.json pnpm-lock.yaml ./ # Install dependencies RUN pnpm install --frozen-lockfile # Copy source (volumes override in dev) COPY . . # Build command (not a running server) # Used via: docker compose exec astro pnpm build CMD ["sleep", "infinity"] ``` **Note:** Frontend builds to `astro/dist/` which is committed to git and bind-mounted into nginx. ### Production (Multi-stage) ```dockerfile # backend/Dockerfile FROM python:3.12-slim AS builder WORKDIR /app COPY --from=ghcr.io/astral-sh/uv:latest /uv /usr/local/bin/uv COPY backend/pyproject.toml ./ RUN uv pip install --system -e "." FROM python:3.12-slim WORKDIR /app RUN apt-get update && apt-get install -y --no-install-recommends \ libsndfile1 ffmpeg \ && rm -rf /var/lib/apt/lists/* COPY --from=builder /usr/local/lib/python3.12/site-packages /usr/local/lib/python3.12/site-packages COPY backend/ . CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000"] ``` ## Docker Compose Patterns ### Development Compose ```yaml # docker-compose.yml services: backend: build: context: ./backend dockerfile: Dockerfile.dev ports: - "8000:8000" volumes: - ./backend:/app - backend-cache:/app/.venv environment: - DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout - REDIS_URL=redis://redis:6379 depends_on: db: condition: service_healthy redis: condition: service_started frontend: build: context: ./frontend dockerfile: Dockerfile.dev # No port exposed - access via nginx at port 9000 volumes: - ./frontend:/app - frontend-node-modules:/app/node_modules environment: - PUBLIC_API_URL=http://localhost:9000 worker: build: context: ./backend dockerfile: Dockerfile.dev command: uv run taskiq worker app.tasks:broker volumes: - ./backend:/app - ./pipeline:/pipeline:ro environment: - DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout - REDIS_URL=redis://redis:6379 depends_on: - db - redis db: image: postgres:18-alpine environment: POSTGRES_USER: shootout POSTGRES_PASSWORD: ${DB_PASSWORD} POSTGRES_DB: shootout ports: - "5432:5432" volumes: - postgres-data:/var/lib/postgresql/data healthcheck: test: ["CMD-SHELL", "pg_isready -U shootout -d shootout"] interval: 5s timeout: 5s retries: 5 redis: image: redis:8.4.0-alpine ports: - "6379:6379" volumes: - redis-data:/data healthcheck: test: ["CMD", "redis-cli", "ping"] interval: 5s timeout: 5s retries: 5 volumes: postgres-data: redis-data: backend-cache: frontend-node-modules: ``` ### Environment Variables ```bash # .env.example # Database DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout POSTGRES_USER=postgres POSTGRES_PASSWORD=postgres POSTGRES_DB=shootout # Redis REDIS_URL=redis://redis:6379 # Frontend PUBLIC_API_URL=http://localhost:8000 # Note: T3K auth uses passwordless OAuth - no client credentials needed. # Tokens saved to .gts-auth.json via ./worktree.py auth-login ``` ## Networking ### Service Discovery Services communicate using service names as hostnames: - Backend → DB: `postgresql://postgres:5432` - Backend → Redis: `redis://redis:6379` - Frontend → Backend: `http://backend:8000` (internal) - Browser → Backend: `http://localhost:8000` (external) ### Port Mapping ```yaml ports: - "host:container" - "8000:8000" # Accessible from host at localhost:8000 ``` ### Expose vs Ports ```yaml # Expose: Internal only (other containers) expose: - "5432" # Ports: External (host machine) ports: - "5432:5432" ``` ## Volume Patterns ### Named Volumes (Persistent) ```yaml volumes: postgres-data: # Database files redis-data: # Redis persistence ``` ### Bind Mounts (Development) ```yaml volumes: - ./backend:/app # Source code (hot-reload) - ./pipeline:/pipeline:ro # Read-only access ``` ### Anonymous Volumes (Cache) ```yaml volumes: - /app/node_modules # Preserve installed dependencies - /app/.venv # Preserve Python venv ``` ## Health Checks ```yaml db: healthcheck: test: ["CMD-SHELL", "pg_isready -U postgres"] interval: 5s timeout: 5s retries: 5 start_period: 10s backend: healthcheck: test: ["CMD", "curl", "-f", "http://localhost:8000/health"] interval: 30s timeout: 10s retries: 3 ``` ## Troubleshooting ### Container Won't Start ```bash # Check logs docker compose logs backend # Check container status docker compose ps # Rebuild from scratch docker compose build --no-cache backend ``` ### Hot-reload Not Working (Backend Python) ```bash # Verify volume mount docker compose exec backend ls -la /app # Check file permissions docker compose exec backend stat /app/app/main.py # Restart service (only if uvicorn --reload isn't picking up changes) docker compose restart backend ``` ### Frontend Templates Not Updating **Pre-bundled architecture**: `astro/dist/` is committed to git. nginx serves files directly from bind mount. ```bash # Rebuild templates (one-time) just build-astro # Watch mode (run in separate terminal) - auto-rebuilds on file changes just watch-templates # Verify dist/ is updated ls -la astro/dist/layouts/ ls -la astro/dist/_astro/ # Check nginx can see the files docker compose exec nginx ls -la /usr/share/nginx/html/ ``` **How it works:** 1. `astro/dist/` is bind-mounted into nginx container 2. Changes to `astro/dist/` are immediately visible to nginx 3. No container restart needed 4. Backend reads Jinja2 wrapper from same `astro/dist/` directory **Note:** Backend reads templates fresh from disk on each request (`auto_reload=True`). If changes still don't appear, check that `astro/dist/` was updated by the build. ### Database Connection Issues ```bash # Check if DB is healthy docker compose ps db # Check DB logs docker compose logs db # Verify network docker compose exec backend ping db # Test connection docker compose exec backend python -c " import asyncpg import asyncio asyncio.run(asyncpg.connect('postgresql://postgres:postgres@db:5432/shootout')) " ``` ### Reset Everything ```bash # Stop all containers docker compose down # Remove all volumes docker compose down -v # Remove all images docker compose down --rmi all # Fresh start docker compose build --no-cache docker compose up ``` ## Production Patterns ### Multi-stage Builds See Dockerfile examples above for smaller production images. ### Resource Limits ```yaml services: backend: deploy: resources: limits: cpus: '1.0' memory: 512M reservations: cpus: '0.25' memory: 128M ``` ### Logging ```yaml services: backend: logging: driver: "json-file" options: max-size: "10m" max-file: "3" ```