Guest User

flink-conf.yaml

a guest
Jul 21st, 2017
278
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
YAML 8.92 KB | None | 0 0
  1. ################################################################################
  2. #  Licensed to the Apache Software Foundation (ASF) under one
  3. #  or more contributor license agreements.  See the NOTICE file
  4. #  distributed with this work for additional information
  5. #  regarding copyright ownership.  The ASF licenses this file
  6. #  to you under the Apache License, Version 2.0 (the
  7. #  "License"); you may not use this file except in compliance
  8. #  with the License.  You may obtain a copy of the License at
  9. #
  10. #      http://www.apache.org/licenses/LICENSE-2.0
  11. #
  12. #  Unless required by applicable law or agreed to in writing, software
  13. #  distributed under the License is distributed on an "AS IS" BASIS,
  14. #  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  15. #  See the License for the specific language governing permissions and
  16. # limitations under the License.
  17. ################################################################################
  18.  
  19.  
  20. #==============================================================================
  21. # Common
  22. #==============================================================================
  23.  
  24. # The external address of the host on which the JobManager runs and can be
  25. # reached by the TaskManagers and any clients which want to connect. This setting
  26. # is only used in Standalone mode and may be overwritten on the JobManager side
  27. # by specifying the --host <hostname> parameter of the bin/jobmanager.sh executable.
  28. # In high availability mode, if you use the bin/start-cluster.sh script and setup
  29. # the conf/masters file, this will be taken care of automatically. Yarn/Mesos
  30. # automatically configure the host name based on the hostname of the node where the
  31. # JobManager runs.
  32.  
  33. jobmanager.rpc.address: 10.0.20.148
  34.  
  35. # The RPC port where the JobManager is reachable.
  36.  
  37. jobmanager.rpc.port: 6123
  38.  
  39.  
  40. # The heap size for the JobManager JVM
  41.  
  42. jobmanager.heap.mb: 1024
  43.  
  44.  
  45. # The heap size for the TaskManager JVM
  46.  
  47. taskmanager.heap.mb: 1024
  48.  
  49.  
  50. # The number of task slots that each TaskManager offers. Each slot runs one parallel pipeline.
  51.  
  52. taskmanager.numberOfTaskSlots: 1
  53.  
  54. # Specify whether TaskManager memory should be allocated when starting up (true) or when
  55. # memory is required in the memory manager (false)
  56. # Important Note: For pure streaming setups, we highly recommend to set this value to `false`
  57. # as the default state backends currently do not use the managed memory.
  58.  
  59. taskmanager.memory.preallocate: false
  60.  
  61. # The parallelism used for programs that did not specify and other parallelism.
  62.  
  63. parallelism.default: 1
  64.  
  65.  
  66. #==============================================================================
  67. # Web Frontend
  68. #==============================================================================
  69.  
  70. # The address under which the web-based runtime monitor listens.
  71. #
  72. #jobmanager.web.address: 0.0.0.0
  73.  
  74. # The port under which the web-based runtime monitor listens.
  75. # A value of -1 deactivates the web server.
  76.  
  77. jobmanager.web.port: 8081
  78.  
  79. # Flag to specify whether job submission is enabled from the web-based
  80. # runtime monitor. Uncomment to disable.
  81.  
  82. #jobmanager.web.submit.enable: false
  83.  
  84. #==============================================================================
  85. # HistoryServer
  86. #==============================================================================
  87.  
  88. # The HistoryServer is started and stopped via bin/historyserver.sh (start|stop)
  89.  
  90. # Directory to upload completed jobs to. Add this directory to the list of
  91. # monitored directories of the HistoryServer as well (see below).
  92. jobmanager.archive.fs.dir: hdfs:///viktor/flink/
  93.  
  94. # The address under which the web-based HistoryServer listens.
  95. #historyserver.web.address: 0.0.0.0
  96.  
  97. # The port under which the web-based HistoryServer listens.
  98. #historyserver.web.port: 8082
  99.  
  100. # Comma separated list of directories to monitor for completed jobs.
  101. historyserver.archive.fs.dir: hdfs:///viktor/flink/
  102.  
  103. # Interval in milliseconds for refreshing the monitored directories.
  104. #historyserver.archive.fs.refresh-interval: 10000
  105.  
  106. #==============================================================================
  107. # Streaming state checkpointing
  108. #==============================================================================
  109.  
  110. # The backend that will be used to store operator state checkpoints if
  111. # checkpointing is enabled.
  112. #
  113. # Supported backends: jobmanager, filesystem, rocksdb, <class-name-of-factory>
  114. #
  115. # state.backend: filesystem
  116.  
  117.  
  118. # Directory for storing checkpoints in a Flink-supported filesystem
  119. # Note: State backend must be accessible from the JobManager and all TaskManagers.
  120. # Use "hdfs://" for HDFS setups, "file://" for UNIX/POSIX-compliant file systems,
  121. # (or any local file system under Windows), or "S3://" for S3 file system.
  122. #
  123. state.backend.fs.checkpointdir: hdfs://cdh01.sbgdinc.com:8020/viktor/flink
  124.  
  125.  
  126. #==============================================================================
  127. # Advanced
  128. #==============================================================================
  129.  
  130. # The number of buffers for the network stack.
  131. #
  132. # taskmanager.network.numberOfBuffers: 2048
  133.  
  134.  
  135. # Directories for temporary files.
  136. #
  137. # Add a delimited list for multiple directories, using the system directory
  138. # delimiter (colon ':' on unix) or a comma, e.g.:
  139. #     /data1/tmp:/data2/tmp:/data3/tmp
  140. #
  141. # Note: Each directory entry is read from and written to by a different I/O
  142. # thread. You can include the same directory multiple times in order to create
  143. # multiple I/O threads against that directory. This is for example relevant for
  144. # high-throughput RAIDs.
  145. #
  146. # If not specified, the system-specific Java temporary directory (java.io.tmpdir
  147. # property) is taken.
  148. #
  149. taskmanager.tmp.dirs: /data/tmp/flink
  150.  
  151.  
  152. # Path to the Hadoop configuration directory.
  153. #
  154. # This configuration is used when writing into HDFS. Unless specified otherwise,
  155. # HDFS file creation will use HDFS default settings with respect to block-size,
  156. # replication factor, etc.
  157. #
  158. # You can also directly specify the paths to hdfs-default.xml and hdfs-site.xml
  159. # via keys 'fs.hdfs.hdfsdefault' and 'fs.hdfs.hdfssite'.
  160. #
  161.  
  162. fs.hdfs.hadoopconf: /etc/hadoop/conf/
  163.  
  164.  
  165. #==============================================================================
  166. # High Availability
  167. #==============================================================================
  168.  
  169. # The high-availability mode. Possible options are 'NONE' or 'zookeeper'.
  170. #
  171. # high-availability: zookeeper
  172.  
  173. # The path where metadata for master recovery is persisted. While ZooKeeper stored
  174. # the small ground truth for checkpoint and leader election, this location stores
  175. # the larger objects, like persisted dataflow graphs.
  176. #
  177. # Must be a durable file system that is accessible from all nodes
  178. # (like HDFS, S3, Ceph, nfs, ...)
  179. #
  180. high-availability.storageDir: hdfs://cdh01.sbgdinc.com:8020/viktor/flink
  181.  
  182. # The list of ZooKeeper quorum peers that coordinate the high-availability
  183. # setup. This must be a list of the form:
  184. # "host1:clientPort,host2:clientPort,..." (default clientPort: 2181)
  185. #
  186. # high-availability.zookeeper.quorum: localhost:2181
  187.  
  188.  
  189. # ACL options are based on https://zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes
  190. # It can be either "creator" (ZOO_CREATE_ALL_ACL) or "open" (ZOO_OPEN_ACL_UNSAFE)
  191. # The default value is "open" and it can be changed to "creator" if ZK security is enabled
  192. #
  193. # high-availability.zookeeper.client.acl: open
  194.  
  195. #==============================================================================
  196. # Flink Cluster Security Configuration (optional configuration)
  197. #==============================================================================
  198.  
  199. # Kerberos authentication for various components - Hadoop, ZooKeeper, and connectors -
  200. # may be enabled in four steps:
  201. # 1. configure the local krb5.conf file
  202. # 2. provide Kerberos credentials (either a keytab or a ticket cache w/ kinit)
  203. # 3. make the credentials available to various JAAS login contexts
  204. # 4. configure the connector to use JAAS/SASL
  205.  
  206. # The below configure how Kerberos credentials are provided. A keytab will be used instead of
  207. # a ticket cache if the keytab path and principal are set.
  208.  
  209. # security.kerberos.login.use-ticket-cache: true
  210. # security.kerberos.login.keytab: /path/to/kerberos/keytab
  211. # security.kerberos.login.principal: flink-user
  212.  
  213. # The configuration below defines which JAAS login contexts
  214.  
  215. # security.kerberos.login.contexts: Client,KafkaClient
  216.  
  217. #==============================================================================
  218. # ZK Security Configuration (optional configuration)
  219. #==============================================================================
  220.  
  221. # Below configurations are applicable if ZK ensemble is configured for security
  222.  
  223. # Override below configuration to provide custom ZK service name if configured
  224. # zookeeper.sasl.service-name: zookeeper
  225.  
  226. # The configuration below must match one of the values set in "security.kerberos.login.contexts"
  227. # zookeeper.sasl.login-context-name: Client
Add Comment
Please, Sign In to add comment