diff --git a/docker-compose-ollama-gpu.yaml b/docker-compose-ollama-gpu.yaml index e4de6b6..97e7a22 100644 --- a/docker-compose-ollama-gpu.yaml +++ b/docker-compose-ollama-gpu.yaml @@ -13,6 +13,8 @@ services: depends_on: - ollama - ollama-webui + networks: + - ollama-docker ollama: volumes: @@ -24,6 +26,8 @@ services: image: ollama/ollama:latest ports: - 11434:11434 + networks: + - ollama-docker deploy: resources: reservations: @@ -40,9 +44,15 @@ services: depends_on: - ollama ports: - - 3000:8080 + - 8080:8080 environment: - '/ollama/api=http://ollama:11434/api' extra_hosts: - host.docker.internal:host-gateway - restart: unless-stopped \ No newline at end of file + restart: unless-stopped + networks: + - ollama-docker + +networks: + ollama-docker: + external: false \ No newline at end of file diff --git a/docker-compose.yml b/docker-compose.yml index da7ddd6..f7c66ab 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -11,7 +11,7 @@ services: restart: always depends_on: - ollama - - llama-webui + - ollama-webui networks: - ollama-docker @@ -29,7 +29,7 @@ services: networks: - ollama-docker - llama-webui: + ollama-webui: image: ghcr.io/ollama-webui/ollama-webui:main container_name: ollama-webui volumes: @@ -37,7 +37,7 @@ services: depends_on: - ollama ports: - - 3000:8080 + - 8080:8080 environment: - '/ollama/api=http://ollama:11434/api' extra_hosts: diff --git a/src/index.html b/src/index.html index a0bafe6..7a92d47 100644 --- a/src/index.html +++ b/src/index.html @@ -32,7 +32,7 @@

Simplify Your Development with Ollama Docker

Streamline your Ollama deployments using Docker Compose. Dive into a containerized environment designed for simplicity and efficiency.

- Ollama Web UI + Ollama Web UI Ollama App Container