version: '3.8' services: app: build: . ports: - 8000:8000 - 5678:5678 volumes: - .:/code command: uvicorn src.main:app --host 0.0.0.0 --port 8000 --reload restart: always depends_on: - ollama - llama-webui networks: - ollama-docker ollama: image: ollama/ollama:latest ports: - 11434:11434 volumes: - .:/code - ./ollama/ollama:/root/.ollama container_name: ollama pull_policy: always tty: true restart: always networks: - ollama-docker llama-webui: image: ghcr.io/ollama-webui/ollama-webui:main container_name: ollama-webui volumes: - ./ollama/ollama-webui:/app/backend/data depends_on: - ollama ports: - 3000:8080 environment: - '/ollama/api=http://ollama:11434/api' extra_hosts: - host.docker.internal:host-gateway restart: unless-stopped networks: - ollama-docker networks: ollama-docker: external: false