| name | docker-infra |
| description | Docker and Docker Compose configuration and troubleshooting. Use for containers, services, volumes, networking, health checks, and deployment patterns. |
Docker & Infrastructure Skill
Activation: Docker, Docker Compose, containers, services, deployment
Development Environment
Architecture
Pre-bundled frontend: astro/dist/ is committed to git. No Vite dev server at runtime.
Runtime containers: db, backend, nginx
Build-only container: frontend (used for just build-astro)
Services Overview
| Service | Port | Image | Purpose |
|---|---|---|---|
| nginx | 9000 | nginx:alpine | Routes static files + API proxy |
| backend | 8000 | python:3.12-slim | FastAPI with uvicorn --reload |
| astro | - | node:24-alpine | Build-only (Astro build) |
| worker | - | python:3.12-slim | TaskIQ workers |
| db | 5432 | postgres:18-alpine | PostgreSQL database |
| redis | 6379 | redis:8.4.0-alpine | Job queue + cache |
Note: Frontend container is only used for builds (just build-astro). At runtime, nginx serves pre-built files from the bind-mounted astro/dist/ directory.
Quick Commands
# Start all services
docker compose up
# Start specific service
docker compose up backend
# Rebuild after dependency changes
docker compose build backend
docker compose up backend
# View logs
docker compose logs -f backend worker
# Shell into container
docker compose exec backend bash
docker compose exec astro sh
# Stop all
docker compose down
# Stop and remove volumes (reset DB)
docker compose down -v
Dockerfile Patterns
Backend (Python)
# backend/Dockerfile.dev
FROM python:3.12-slim
WORKDIR /app
# System dependencies
RUN apt-get update && apt-get install -y --no-install-recommends \
build-essential \
curl \
libsndfile1 \
ffmpeg \
&& rm -rf /var/lib/apt/lists/*
# Install uv for fast dependency management
COPY --from=ghcr.io/astral-sh/uv:latest /uv /usr/local/bin/uv
# Copy dependency files
COPY backend/pyproject.toml ./
# Install dependencies
RUN uv pip install --system -e ".[dev]"
# Copy source (volumes override in dev)
COPY backend/ .
# Run with hot-reload (via entrypoint.sh for migrations)
ENTRYPOINT ["/bin/bash", "/app/entrypoint.sh"]
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000", "--reload"]
Frontend (Node.js) - Build Only
The frontend container is used for builds only. No dev server runs at runtime.
# astro/Dockerfile.dev
FROM node:24-alpine
WORKDIR /app
# Install pnpm
RUN corepack enable && corepack prepare pnpm@latest --activate
# Copy dependency files
COPY package.json pnpm-lock.yaml ./
# Install dependencies
RUN pnpm install --frozen-lockfile
# Copy source (volumes override in dev)
COPY . .
# Build command (not a running server)
# Used via: docker compose exec astro pnpm build
CMD ["sleep", "infinity"]
Note: Frontend builds to astro/dist/ which is committed to git and bind-mounted into nginx.
Production (Multi-stage)
# backend/Dockerfile
FROM python:3.12-slim AS builder
WORKDIR /app
COPY --from=ghcr.io/astral-sh/uv:latest /uv /usr/local/bin/uv
COPY backend/pyproject.toml ./
RUN uv pip install --system -e "."
FROM python:3.12-slim
WORKDIR /app
RUN apt-get update && apt-get install -y --no-install-recommends \
libsndfile1 ffmpeg \
&& rm -rf /var/lib/apt/lists/*
COPY --from=builder /usr/local/lib/python3.12/site-packages /usr/local/lib/python3.12/site-packages
COPY backend/ .
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "8000"]
Docker Compose Patterns
Development Compose
# docker-compose.yml
services:
backend:
build:
context: ./backend
dockerfile: Dockerfile.dev
ports:
- "8000:8000"
volumes:
- ./backend:/app
- backend-cache:/app/.venv
environment:
- DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout
- REDIS_URL=redis://redis:6379
depends_on:
db:
condition: service_healthy
redis:
condition: service_started
frontend:
build:
context: ./frontend
dockerfile: Dockerfile.dev
# No port exposed - access via nginx at port 9000
volumes:
- ./frontend:/app
- frontend-node-modules:/app/node_modules
environment:
- PUBLIC_API_URL=http://localhost:9000
worker:
build:
context: ./backend
dockerfile: Dockerfile.dev
command: uv run taskiq worker app.tasks:broker
volumes:
- ./backend:/app
- ./pipeline:/pipeline:ro
environment:
- DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout
- REDIS_URL=redis://redis:6379
depends_on:
- db
- redis
db:
image: postgres:18-alpine
environment:
POSTGRES_USER: shootout
POSTGRES_PASSWORD: ${DB_PASSWORD}
POSTGRES_DB: shootout
ports:
- "5432:5432"
volumes:
- postgres-data:/var/lib/postgresql/data
healthcheck:
test: ["CMD-SHELL", "pg_isready -U shootout -d shootout"]
interval: 5s
timeout: 5s
retries: 5
redis:
image: redis:8.4.0-alpine
ports:
- "6379:6379"
volumes:
- redis-data:/data
healthcheck:
test: ["CMD", "redis-cli", "ping"]
interval: 5s
timeout: 5s
retries: 5
volumes:
postgres-data:
redis-data:
backend-cache:
frontend-node-modules:
Environment Variables
# .env.example
# Database
DATABASE_URL=postgresql+asyncpg://postgres:postgres@db:5432/shootout
POSTGRES_USER=postgres
POSTGRES_PASSWORD=postgres
POSTGRES_DB=shootout
# Redis
REDIS_URL=redis://redis:6379
# Frontend
PUBLIC_API_URL=http://localhost:8000
# Note: T3K auth uses passwordless OAuth - no client credentials needed.
# Tokens saved to .gts-auth.json via ./worktree.py auth-login
Networking
Service Discovery
Services communicate using service names as hostnames:
- Backend → DB:
postgresql://postgres:5432 - Backend → Redis:
redis://redis:6379 - Frontend → Backend:
http://backend:8000(internal) - Browser → Backend:
http://localhost:8000(external)
Port Mapping
ports:
- "host:container"
- "8000:8000" # Accessible from host at localhost:8000
Expose vs Ports
# Expose: Internal only (other containers)
expose:
- "5432"
# Ports: External (host machine)
ports:
- "5432:5432"
Volume Patterns
Named Volumes (Persistent)
volumes:
postgres-data: # Database files
redis-data: # Redis persistence
Bind Mounts (Development)
volumes:
- ./backend:/app # Source code (hot-reload)
- ./pipeline:/pipeline:ro # Read-only access
Anonymous Volumes (Cache)
volumes:
- /app/node_modules # Preserve installed dependencies
- /app/.venv # Preserve Python venv
Health Checks
db:
healthcheck:
test: ["CMD-SHELL", "pg_isready -U postgres"]
interval: 5s
timeout: 5s
retries: 5
start_period: 10s
backend:
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8000/health"]
interval: 30s
timeout: 10s
retries: 3
Troubleshooting
Container Won't Start
# Check logs
docker compose logs backend
# Check container status
docker compose ps
# Rebuild from scratch
docker compose build --no-cache backend
Hot-reload Not Working (Backend Python)
# Verify volume mount
docker compose exec backend ls -la /app
# Check file permissions
docker compose exec backend stat /app/app/main.py
# Restart service (only if uvicorn --reload isn't picking up changes)
docker compose restart backend
Frontend Templates Not Updating
Pre-bundled architecture: astro/dist/ is committed to git. nginx serves files directly from bind mount.
# Rebuild templates (one-time)
just build-astro
# Watch mode (run in separate terminal) - auto-rebuilds on file changes
just watch-templates
# Verify dist/ is updated
ls -la astro/dist/layouts/
ls -la astro/dist/_astro/
# Check nginx can see the files
docker compose exec nginx ls -la /usr/share/nginx/html/
How it works:
astro/dist/is bind-mounted into nginx container- Changes to
astro/dist/are immediately visible to nginx - No container restart needed
- Backend reads Jinja2 wrapper from same
astro/dist/directory
Note: Backend reads templates fresh from disk on each request (auto_reload=True). If changes still don't appear, check that astro/dist/ was updated by the build.
Database Connection Issues
# Check if DB is healthy
docker compose ps db
# Check DB logs
docker compose logs db
# Verify network
docker compose exec backend ping db
# Test connection
docker compose exec backend python -c "
import asyncpg
import asyncio
asyncio.run(asyncpg.connect('postgresql://postgres:postgres@db:5432/shootout'))
"
Reset Everything
# Stop all containers
docker compose down
# Remove all volumes
docker compose down -v
# Remove all images
docker compose down --rmi all
# Fresh start
docker compose build --no-cache
docker compose up
Production Patterns
Multi-stage Builds
See Dockerfile examples above for smaller production images.
Resource Limits
services:
backend:
deploy:
resources:
limits:
cpus: '1.0'
memory: 512M
reservations:
cpus: '0.25'
memory: 128M
Logging
services:
backend:
logging:
driver: "json-file"
options:
max-size: "10m"
max-file: "3"