Guest User

Untitled

a guest
Feb 9th, 2014
126
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 9.29 KB | None | 0 0
  1. /* This configuration is as close to 2.5.x default behavior as possible
  2. The values closely match ./gmond/metric.h definitions in 2.5.x */
  3. globals {
  4. daemonize = yes
  5. setuid = yes
  6. user = ganglia
  7. debug_level = 0
  8. max_udp_msg_len = 1472
  9. mute = no
  10. deaf = no
  11. allow_extra_data = yes
  12. host_dmax = 86400 /*secs. Expires (removes from web interface) hosts in 1 day */
  13. host_tmax = 20 /*secs */
  14. cleanup_threshold = 300 /*secs */
  15. gexec = no
  16. # By default gmond will use reverse DNS resolution when displaying your hostname
  17. # Uncommeting following value will override that value.
  18. # override_hostname = "mywebserver.domain.com"
  19. # If you are not using multicast this value should be set to something other than 0.
  20. # Otherwise if you restart aggregator gmond you will get empty graphs. 60 seconds is reasonable
  21. send_metadata_interval = 0 /*secs */
  22.  
  23. }
  24.  
  25. /*
  26. * The cluster attributes specified will be used as part of the <CLUSTER>
  27. * tag that will wrap all hosts collected by this instance.
  28. */
  29. cluster {
  30. name = "Ganglia Test Setup"
  31. owner = "Ops"
  32. latlong = "unspecified"
  33. url = "unspecified"
  34. }
  35.  
  36. /* The host section describes attributes of the host, like the location */
  37. host {
  38. location = "unspecified"
  39. }
  40.  
  41. /* Feel free to specify as many udp_send_channels as you like. Gmond
  42. used to only support having a single channel */
  43. udp_send_channel {
  44. #bind_hostname = yes # Highly recommended, soon to be default.
  45. # This option tells gmond to use a source address
  46. # that resolves to the machine's hostname. Without
  47. # this, the metrics may appear to come from any
  48. # interface and the DNS names associated with
  49. # those IPs will be used to create the RRDs.
  50. #host = test4
  51. mcast_join = 239.2.11.71
  52. port = 8649
  53. ttl = 1
  54. }
  55.  
  56. /* You can specify as many udp_recv_channels as you like as well. */
  57. udp_recv_channel {
  58. mcast_join = 239.2.11.71
  59. port = 8649
  60. bind = 239.2.11.71
  61. retry_bind = true
  62. # Size of the UDP buffer. If you are handling lots of metrics you really
  63. # should bump it up to e.g. 10MB or even higher.
  64. # buffer = 10485760
  65. }
  66.  
  67. /* You can specify as many tcp_accept_channels as you like to share
  68. an xml description of the state of the cluster */
  69. tcp_accept_channel {
  70. port = 8649
  71. # If you want to gzip XML output
  72. #gzip_output = no
  73. }
  74.  
  75. /* Channel to receive sFlow datagrams */
  76. #udp_recv_channel {
  77. # port = 6343
  78. #}
  79.  
  80. /* Optional sFlow settings */
  81. #sflow {
  82. # udp_port = 6343
  83. # accept_vm_metrics = yes
  84. # accept_jvm_metrics = yes
  85. # multiple_jvm_instances = no
  86. # accept_http_metrics = yes
  87. # multiple_http_instances = no
  88. # accept_memcache_metrics = yes
  89. # multiple_memcache_instances = no
  90. #}
  91.  
  92. /* Each metrics module that is referenced by gmond must be specified and
  93. loaded. If the module has been statically linked with gmond, it does
  94. not require a load path. However all dynamically loadable modules must
  95. include a load path. */
  96. modules {
  97. module {
  98. name = "core_metrics"
  99. }
  100. module {
  101. name = "python_module"
  102. path = "/usr/local/lib64/ganglia/modpython.so"
  103. params = "/usr/local/lib64/ganglia/python_modules/"
  104. }
  105. module {
  106. name = "cpu_module"
  107. path = "modcpu.so"
  108. }
  109. module {
  110. name = "disk_module"
  111. path = "moddisk.so"
  112. }
  113. module {
  114. name = "load_module"
  115. path = "modload.so"
  116. }
  117. module {
  118. name = "mem_module"
  119. path = "modmem.so"
  120. }
  121. module {
  122. name = "net_module"
  123. path = "modnet.so"
  124. }
  125. module {
  126. name = "proc_module"
  127. path = "modproc.so"
  128. }
  129. module {
  130. name = "sys_module"
  131. path = "modsys.so"
  132. }
  133. }
  134.  
  135. /* The old internal 2.5.x metric array has been replaced by the following
  136. collection_group directives. What follows is the default behavior for
  137. collecting and sending metrics that is as close to 2.5.x behavior as
  138. possible. */
  139.  
  140. /* This collection group will cause a heartbeat (or beacon) to be sent every
  141. 20 seconds. In the heartbeat is the GMOND_STARTED data which expresses
  142. the age of the running gmond. */
  143. collection_group {
  144. collect_once = yes
  145. time_threshold = 20
  146. metric {
  147. name = "heartbeat"
  148. }
  149. }
  150.  
  151. /* This collection group will send general info about this host every
  152. 1200 secs.
  153. This information doesn't change between reboots and is only collected
  154. once. */
  155. collection_group {
  156. collect_once = yes
  157. time_threshold = 1200
  158. metric {
  159. name = "cpu_num"
  160. title = "CPU Count"
  161. }
  162. metric {
  163. name = "cpu_speed"
  164. title = "CPU Speed"
  165. }
  166. metric {
  167. name = "mem_total"
  168. title = "Memory Total"
  169. }
  170. /* Should this be here? Swap can be added/removed between reboots. */
  171. metric {
  172. name = "swap_total"
  173. title = "Swap Space Total"
  174. }
  175. metric {
  176. name = "boottime"
  177. title = "Last Boot Time"
  178. }
  179. metric {
  180. name = "machine_type"
  181. title = "Machine Type"
  182. }
  183. metric {
  184. name = "os_name"
  185. title = "Operating System"
  186. }
  187. metric {
  188. name = "os_release"
  189. title = "Operating System Release"
  190. }
  191. metric {
  192. name = "location"
  193. title = "Location"
  194. }
  195. }
  196.  
  197. /* This collection group will send the status of gexecd for this host
  198. every 300 secs.*/
  199. /* Unlike 2.5.x the default behavior is to report gexecd OFF. */
  200. collection_group {
  201. collect_once = yes
  202. time_threshold = 300
  203. metric {
  204. name = "gexec"
  205. title = "Gexec Status"
  206. }
  207. }
  208.  
  209. /* This collection group will collect the CPU status info every 20 secs.
  210. The time threshold is set to 90 seconds. In honesty, this
  211. time_threshold could be set significantly higher to reduce
  212. unneccessary network chatter. */
  213. collection_group {
  214. collect_every = 20
  215. time_threshold = 90
  216. /* CPU status */
  217. metric {
  218. name = "cpu_user"
  219. value_threshold = "1.0"
  220. title = "CPU User"
  221. }
  222. metric {
  223. name = "cpu_system"
  224. value_threshold = "1.0"
  225. title = "CPU System"
  226. }
  227. metric {
  228. name = "cpu_idle"
  229. value_threshold = "5.0"
  230. title = "CPU Idle"
  231. }
  232. metric {
  233. name = "cpu_nice"
  234. value_threshold = "1.0"
  235. title = "CPU Nice"
  236. }
  237. metric {
  238. name = "cpu_aidle"
  239. value_threshold = "5.0"
  240. title = "CPU aidle"
  241. }
  242. metric {
  243. name = "cpu_wio"
  244. value_threshold = "1.0"
  245. title = "CPU wio"
  246. }
  247. metric {
  248. name = "cpu_steal"
  249. value_threshold = "1.0"
  250. title = "CPU steal"
  251. }
  252. /* The next two metrics are optional if you want more detail...
  253. ... since they are accounted for in cpu_system.
  254. metric {
  255. name = "cpu_intr"
  256. value_threshold = "1.0"
  257. title = "CPU intr"
  258. }
  259. metric {
  260. name = "cpu_sintr"
  261. value_threshold = "1.0"
  262. title = "CPU sintr"
  263. }
  264. */
  265. }
  266.  
  267. collection_group {
  268. collect_every = 20
  269. time_threshold = 90
  270. /* Load Averages */
  271. metric {
  272. name = "load_one"
  273. value_threshold = "1.0"
  274. title = "One Minute Load Average"
  275. }
  276. metric {
  277. name = "load_five"
  278. value_threshold = "1.0"
  279. title = "Five Minute Load Average"
  280. }
  281. metric {
  282. name = "load_fifteen"
  283. value_threshold = "1.0"
  284. title = "Fifteen Minute Load Average"
  285. }
  286. }
  287.  
  288. /* This group collects the number of running and total processes */
  289. collection_group {
  290. collect_every = 80
  291. time_threshold = 950
  292. metric {
  293. name = "proc_run"
  294. value_threshold = "1.0"
  295. title = "Total Running Processes"
  296. }
  297. metric {
  298. name = "proc_total"
  299. value_threshold = "1.0"
  300. title = "Total Processes"
  301. }
  302. }
  303.  
  304. /* This collection group grabs the volatile memory metrics every 40 secs and
  305. sends them at least every 180 secs. This time_threshold can be increased
  306. significantly to reduce unneeded network traffic. */
  307. collection_group {
  308. collect_every = 40
  309. time_threshold = 180
  310. metric {
  311. name = "mem_free"
  312. value_threshold = "1024.0"
  313. title = "Free Memory"
  314. }
  315. metric {
  316. name = "mem_shared"
  317. value_threshold = "1024.0"
  318. title = "Shared Memory"
  319. }
  320. metric {
  321. name = "mem_buffers"
  322. value_threshold = "1024.0"
  323. title = "Memory Buffers"
  324. }
  325. metric {
  326. name = "mem_cached"
  327. value_threshold = "1024.0"
  328. title = "Cached Memory"
  329. }
  330. metric {
  331. name = "swap_free"
  332. value_threshold = "1024.0"
  333. title = "Free Swap Space"
  334. }
  335. }
  336.  
  337. collection_group {
  338. collect_every = 40
  339. time_threshold = 300
  340. metric {
  341. name = "bytes_out"
  342. value_threshold = 4096
  343. title = "Bytes Sent"
  344. }
  345. metric {
  346. name = "bytes_in"
  347. value_threshold = 4096
  348. title = "Bytes Received"
  349. }
  350. metric {
  351. name = "pkts_in"
  352. value_threshold = 256
  353. title = "Packets Received"
  354. }
  355. metric {
  356. name = "pkts_out"
  357. value_threshold = 256
  358. title = "Packets Sent"
  359. }
  360. }
  361.  
  362. /* Different than 2.5.x default since the old config made no sense */
  363. collection_group {
  364. collect_every = 1800
  365. time_threshold = 3600
  366. metric {
  367. name = "disk_total"
  368. value_threshold = 1.0
  369. title = "Total Disk Space"
  370. }
  371. }
  372.  
  373. collection_group {
  374. collect_every = 40
  375. time_threshold = 180
  376. metric {
  377. name = "disk_free"
  378. value_threshold = 1.0
  379. title = "Disk Space Available"
  380. }
  381. metric {
  382. name = "part_max_used"
  383. value_threshold = 1.0
  384. title = "Maximum Disk Space Used"
  385. }
  386. }
  387.  
  388. include ("/usr/local/etc/conf.d/*.conf")
  389.  
  390. include('/etc/ganglia/conf.d/*.pyconf')
Advertisement
Add Comment
Please, Sign In to add comment