• pezhore@infosec.pub
            link
            fedilink
            English
            arrow-up
            1
            ·
            5 days ago

            I’m doing that with docker compose in my homelab, it’s pretty neat!

            services:
              ollama:
                volumes:
                  - /etc/ollama-docker/ollama:/root/.ollama
                container_name: ollama
                pull_policy: always
                tty: true
                restart: unless-stopped
                image: ollama/ollama
                ports:
                  - 11434:11434
                deploy:
                  resources:
                    reservations:
                      devices:
                        - driver: nvidia
                          device_ids: ['0']
                          capabilities:
                            - gpu
            
              open-webui:
                build:
                  context: .
                  args:
                    OLLAMA_BASE_URL: '/ollama'
                  dockerfile: Dockerfile
                image: ghcr.io/open-webui/open-webui:main
                container_name: open-webui
                volumes:
                  - /etc/ollama-docker/open-webui:/app/backend/data
                depends_on:
                  - ollama
                ports:
                  - 3000:8080
                environment:
                  - 'OLLAMA_BASE_URL=http://ollama:11434/'
                  - 'WEBUI_SECRET_KEY='
                extra_hosts:
                  - host.docker.internal:host-gateway
                restart: unless-stopped
            
            volumes:
              ollama: {}
              open-webui: {}