Guest User

Untitled

a guest
Feb 1st, 2025
35
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
YAML 1.29 KB | None | 0 0
  1. version: '3.8'
  2.  
  3. services:
  4.   ollama-intel-gpu:
  5.     container_name: ollama-intel-gpu
  6.     image: intelanalytics/ipex-llm-inference-cpp-xpu:latest
  7.     restart: unless-stopped
  8.     ports:
  9.      - "11434:11434"
  10.     devices:
  11.      - /dev/dri:/dev/dri
  12.     volumes:
  13.      - ./ollama:/root/.ollama
  14.     environment:
  15.      - no_proxy=localhost,127.0.0.1
  16.       - DEVICE=Arc
  17.       - OLLAMA_HOST=0.0.0.0:11434
  18.     entrypoint: /bin/bash /root/.ollama/start.sh
  19.     #command: /bin/bash source ipex-llm-init --gpu --device $DEVICE && chmod +X /llm/scripts/start-ollama.sh && bash /llm/scripts/start-ollama.sh
  20.  
  21.   #ollama:
  22.     #image: ollama/ollama
  23.     #container_name: ollama
  24.     #ports:
  25.     #  - "11434:11434"
  26.     #volumes:
  27.     #  - ./ollama:/root/.ollama
  28.     #restart: unless-stopped
  29.     #deploy:
  30.     #  resources:
  31.     #    reservations:
  32.     #      devices:
  33.     #        - capabilities: ["gpu"]
  34.     # For older versions of Docker, you might need to use the environment method
  35.     # environment:
  36.     #   - NVIDIA_VISIBLE_DEVICES=all
  37.  
  38.   open-webui:
  39.     image: ghcr.io/open-webui/open-webui:main
  40.     container_name: open-webui
  41.     ports:
  42.      - "8282:8080"
  43.     extra_hosts:
  44.      - "host.docker.internal:host-gateway"
  45.     volumes:
  46.      - ./open-webui:/app/backend/data
  47.     restart: unless-stopped
  48.  
Advertisement
Add Comment
Please, Sign In to add comment