Advertisement
Guest User

Untitled

a guest
May 26th, 2025
27
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.91 KB | None | 0 0
  1. services:
  2. llm-server:
  3. stdin_open: true
  4. tty: true
  5. environment:
  6. - OLLAMA_MODEL=https://ollama.com/PetrosStav/gemma3-tools
  7. - OLLAMA_MODELS=/ssd/.ollama
  8. - OLLAMA_HOST=0.0.0.0:9000
  9. - OLLAMA_CONTEXT_LEN=4096
  10. - OLLAMA_LOGS=/root/.ollama/ollama.log
  11. - DOCKER_PULL=always
  12. - HF_TOKEN=${HF_TOKEN}
  13. - HF_HUB_CACHE=/root/.cache/huggingface
  14. volumes:
  15. - /ssd/cache/ollama:/root/.ollama
  16. - /ssd/cache:/root/.cache
  17. deploy:
  18. resources:
  19. reservations:
  20. devices:
  21. - driver: nvidia
  22. count: all
  23. capabilities:
  24. - gpu
  25. ports:
  26. - 9000:9000
  27. pull_policy: always
  28. image: dustynv/ollama:main-r36.4.0
  29. healthcheck:
  30. test: ["CMD", "curl", "-f", "http://0.0.0.0:9000/v1/models"]
  31. interval: 20s
  32. timeout: 60s
  33. retries: 45
  34. start_period: 15s
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement