• pezhore@infosec.pub
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 days ago

          I’m doing that with docker compose in my homelab, it’s pretty neat!

          services:
            ollama:
              volumes:
                - /etc/ollama-docker/ollama:/root/.ollama
              container_name: ollama
              pull_policy: always
              tty: true
              restart: unless-stopped
              image: ollama/ollama
              ports:
                - 11434:11434
              deploy:
                resources:
                  reservations:
                    devices:
                      - driver: nvidia
                        device_ids: ['0']
                        capabilities:
                          - gpu
          
            open-webui:
              build:
                context: .
                args:
                  OLLAMA_BASE_URL: '/ollama'
                dockerfile: Dockerfile
              image: ghcr.io/open-webui/open-webui:main
              container_name: open-webui
              volumes:
                - /etc/ollama-docker/open-webui:/app/backend/data
              depends_on:
                - ollama
              ports:
                - 3000:8080
              environment:
                - 'OLLAMA_BASE_URL=http://ollama:11434/'
                - 'WEBUI_SECRET_KEY='
              extra_hosts:
                - host.docker.internal:host-gateway
              restart: unless-stopped
          
          volumes:
            ollama: {}
            open-webui: {}