My compose file to run ollama and ollama-webui (original) (raw)
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.Learn more about bidirectional Unicode characters
[ Show hidden characters]({{ revealButtonHref }})
services: |
---|
# ollama and API |
ollama: |
image: ollama/ollama:latest |
container_name: ollama |
pull_policy: missing |
tty: true |
restart: unless-stopped |
# Expose Ollama API outside the container stack (but only on the same computer; |
# remove 127.0.0.1: to make Ollama available on your network) |
ports: |
- 127.0.0.1:11434:11434 |
volumes: |
- ollama:/root/.ollama |
# GPU support (turn off by commenting with # if you don't have an nvidia gpu) |
deploy: |
resources: |
reservations: |
devices: |
- driver: nvidia |
count: 1 |
capabilities: |
- gpu |
# webui, nagivate to http://localhost:3000/ to use |
open-webui: |
image: ghcr.io/open-webui/open-webui:main |
container_name: open-webui |
pull_policy: missing |
volumes: |
- open-webui:/app/backend/data |
depends_on: |
- ollama |
ports: |
- 3000:8080 |
environment: |
- "OLLAMA_API_BASE_URL=http://ollama:11434/api" |
extra_hosts: |
- host.docker.internal:host-gateway |
restart: unless-stopped |
volumes: |
ollama: {} |
open-webui: {} |