Advertisement
Guest User

a

a guest
Jul 20th, 2019
83
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 1.12 KB | None | 0 0
  1. version: "3.7"
  2. services:
  3.  
  4. spark-master:
  5. image: michaelcozzolino/apache-spark
  6. networks:
  7. - outside
  8.  
  9. volumes:
  10. - ./spark_data:/spark_data
  11. - ./spark_data/conf:/usr/local/spark/conf
  12. environment:
  13. - "SPARK_MASTER_HOST=192.168.1.15"
  14. - "SPARK_ROLE=master"
  15. - "SPARK_WORKER_OPTS=-Dspark.worker.cleanup.enabled=false"
  16.  
  17.  
  18. deploy:
  19. replicas: 1
  20. placement:
  21. constraints: [node.labels.type == sparkmaster]
  22. resources:
  23. limits:
  24. memory: 900m
  25.  
  26. spark-worker:
  27. image: michaelcozzolino/apache-spark
  28. networks:
  29. outside:
  30. entrypoint: dockerize -wait tcp://192.168.1.15:7077 -timeout 240s /sbin/my_init
  31.  
  32. environment:
  33. - "SPARK_ROLE=slave"
  34. - "SPARK_MASTER=192.168.1.15"
  35. - "SPARK_WORKER_MEMORY=2000m"
  36. - "SPARK_WORKER_CORES=1"
  37. deploy:
  38. replicas: 1
  39. placement:
  40. constraints: [node.labels.type != sparkmaster]
  41.  
  42. volumes:
  43. data:
  44. app-data:
  45.  
  46. networks:
  47. outside:
  48. external: true
  49. name: "host"
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement