Advertisement
CamiloSan

/var/log/logstash/logstash-plain.log

Feb 10th, 2017
152
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
  1. [2017-02-10T22:11:22,086][DEBUG][logstash.runner          ] -------- Logstash Settings (* means modified) ---------
  2. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] node.name: "logstash-indexer1-fqdn"
  3. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] *path.config: "/etc/logstash/conf.d"
  4. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] *path.data: "/var/lib/logstash" (default: "/usr/share/logstash/data")
  5. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] config.test_and_exit: false
  6. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] config.reload.automatic: false
  7. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] config.reload.interval: 3
  8. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] metric.collect: true
  9. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] pipeline.id: "main"
  10. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] *pipeline.workers: 6 (default: 8)
  11. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] pipeline.output.workers: 1
  12. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] pipeline.batch.size: 125
  13. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] pipeline.batch.delay: 5
  14. [2017-02-10T22:11:22,088][DEBUG][logstash.runner          ] pipeline.unsafe_shutdown: false
  15. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] path.plugins: []
  16. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] config.debug: false
  17. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] *log.level: "debug" (default: "info")
  18. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] version: false
  19. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] help: false
  20. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] log.format: "plain"
  21. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] http.host: "127.0.0.1"
  22. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] http.port: 9600..9700
  23. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] http.environment: "production"
  24. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.type: "memory"
  25. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.page_capacity: 262144000
  26. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.max_bytes: 1073741824
  27. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.max_events: 0
  28. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.checkpoint.acks: 1024
  29. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.checkpoint.writes: 1024
  30. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] queue.checkpoint.interval: 1000
  31. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] slowlog.threshold.warn: -1
  32. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] slowlog.threshold.info: -1
  33. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] slowlog.threshold.debug: -1
  34. [2017-02-10T22:11:22,089][DEBUG][logstash.runner          ] slowlog.threshold.trace: -1
  35. [2017-02-10T22:11:22,090][DEBUG][logstash.runner          ] path.queue: "/usr/share/logstash/data/queue"
  36. [2017-02-10T22:11:22,090][DEBUG][logstash.runner          ] *path.settings: "/etc/logstash" (default: "/usr/share/logstash/config")
  37. [2017-02-10T22:11:22,090][DEBUG][logstash.runner          ] *path.logs: "/var/log/logstash" (default: "/usr/share/logstash/logs")
  38. [2017-02-10T22:11:22,090][DEBUG][logstash.runner          ] --------------- Logstash Settings -------------------
  39. [2017-02-10T22:11:22,097][DEBUG][logstash.agent           ] Agent: Configuring metric collection
  40. [2017-02-10T22:11:22,098][DEBUG][logstash.instrument.periodicpoller.os] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
  41. [2017-02-10T22:11:22,115][DEBUG][logstash.instrument.periodicpoller.jvm] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
  42. [2017-02-10T22:11:22,144][DEBUG][logstash.instrument.periodicpoller.persistentqueue] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
  43. [2017-02-10T22:11:22,148][DEBUG][logstash.agent           ] Reading config file {:config_file=>"/etc/logstash/conf.d/09_input_kafka_cluster.yml"}
  44. [2017-02-10T22:11:22,148][DEBUG][logstash.agent           ] Reading config file {:config_file=>"/etc/logstash/conf.d/50_filter_pending.yml"}
  45. [2017-02-10T22:11:22,149][DEBUG][logstash.agent           ] Reading config file {:config_file=>"/etc/logstash/conf.d/80_output_elastic.yml"}
  46. [2017-02-10T22:11:22,187][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"kafka", :type=>"input", :class=>LogStash::Inputs::Kafka}
  47. [2017-02-10T22:11:22,195][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"plain", :type=>"codec", :class=>LogStash::Codecs::Plain}
  48. [2017-02-10T22:11:22,197][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@id = "plain_c6eb5a12-536e-4442-9c98-5bc73f1f31da"
  49. [2017-02-10T22:11:22,198][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@enable_metric = true
  50. [2017-02-10T22:11:22,198][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@charset = "UTF-8"
  51. [2017-02-10T22:11:22,199][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@bootstrap_servers = "kafka3-fqdn:9092,kafka1-fqdn:9092,kafka2-fqdn:9092"
  52. [2017-02-10T22:11:22,199][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@topics = ["clusterlogs"]
  53. [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@client_id = "logstash-indexer1-fqdn"
  54. [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@group_id = "logstash_indexer"
  55. [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@add_field = {"log_origin"=>"kafka"}
  56. [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@id = "cd8451de5cf0159c42b614a3e21258137a6e2e5e-1"
  57. [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@enable_metric = true
  58. [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@codec = <LogStash::Codecs::Plain id=>"plain_c6eb5a12-536e-4442-9c98-5bc73f1f31da", enable_metric=>true, charset=>"UTF-8">
  59. [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@auto_commit_interval_ms = "5000"
  60. [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@consumer_threads = 1
  61. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@enable_auto_commit = "true"
  62. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@key_deserializer_class = "org.apache.kafka.common.serialization.StringDeserializer"
  63. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@value_deserializer_class = "org.apache.kafka.common.serialization.StringDeserializer"
  64. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@poll_timeout_ms = 100
  65. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@ssl = false
  66. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@security_protocol = "PLAINTEXT"
  67. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@sasl_mechanism = "GSSAPI"
  68. [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka    ] config LogStash::Inputs::Kafka/@decorate_events = false
  69. [2017-02-10T22:11:22,281][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"elasticsearch", :type=>"output", :class=>LogStash::Outputs::ElasticSearch}
  70. [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@id = "plain_506a9778-3738-4f33-bb90-9b058de6eb7f"
  71. [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@enable_metric = true
  72. [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain    ] config LogStash::Codecs::Plain/@charset = "UTF-8"
  73. [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@hosts = [//10.86.97.128:9200]
  74. [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@manage_template = false
  75. [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@index = "%{[@metadata][beat]}-%{+YYYY.MM.dd}"
  76. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@document_type = "%{[@metadata][type]}"
  77. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@id = "cd8451de5cf0159c42b614a3e21258137a6e2e5e-2"
  78. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@enable_metric = true
  79. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@codec = <LogStash::Codecs::Plain id=>"plain_506a9778-3738-4f33-bb90-9b058de6eb7f", enable_metric=>true, charset=>"UTF-8">
  80. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@workers = 1
  81. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_name = "logstash"
  82. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_overwrite = false
  83. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@parent = nil
  84. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@idle_flush_time = 1
  85. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@upsert = ""
  86. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@doc_as_upsert = false
  87. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script = ""
  88. [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_type = "inline"
  89. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_lang = "painless"
  90. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_var_name = "event"
  91. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@scripted_upsert = false
  92. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_initial_interval = 2
  93. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_max_interval = 64
  94. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_on_conflict = 1
  95. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pipeline = nil
  96. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@action = "index"
  97. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@ssl_certificate_verification = true
  98. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing = false
  99. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing_delay = 5
  100. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@timeout = 60
  101. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@failure_type_logging_whitelist = []
  102. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max = 1000
  103. [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max_per_route = 100
  104. [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@healthcheck_path = "/"
  105. [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@absolute_healthcheck_path = false
  106. [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@resurrect_delay = 5
  107. [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@validate_after_inactivity = 10000
  108. [2017-02-10T22:11:22,298][DEBUG][logstash.agent           ] starting agent
  109. [2017-02-10T22:11:22,299][DEBUG][logstash.agent           ] starting pipeline {:id=>"main"}
  110. [2017-02-10T22:11:22,304][DEBUG][logstash.outputs.elasticsearch] Normalizing http path {:path=>nil, :normalized=>nil}
  111. [2017-02-10T22:11:22,501][INFO ][logstash.outputs.elasticsearch] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://10.86.97.128:9200/]}}
  112. [2017-02-10T22:11:22,501][INFO ][logstash.outputs.elasticsearch] Running health check to see if an Elasticsearch connection is working {:healthcheck_url=>http://10.86.97.128:9200/, :path=>"/"}
  113. [2017-02-10T22:11:22,550][WARN ][logstash.outputs.elasticsearch] Restored connection to ES instance {:url=>#<URI::HTTP:0x5e1e5983 URL:http://10.86.97.128:9200/>}
  114. [2017-02-10T22:11:22,551][INFO ][logstash.outputs.elasticsearch] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>[#<URI::Generic:0x3eb5a14 URL://10.86.97.128:9200>]}
  115. [2017-02-10T22:11:22,573][INFO ][logstash.pipeline        ] Starting pipeline {"id"=>"main", "pipeline.workers"=>6, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>5, "pipeline.max_inflight"=>750}
  116. [2017-02-10T22:11:22,575][INFO ][logstash.pipeline        ] Pipeline main started
  117. [2017-02-10T22:11:22,579][DEBUG][logstash.agent           ] Starting puma
  118. [2017-02-10T22:11:22,579][DEBUG][logstash.agent           ] Trying to start WebServer {:port=>9600}
  119. [2017-02-10T22:11:22,580][DEBUG][logstash.api.service     ] [api-service] start
  120. [2017-02-10T22:11:22,593][INFO ][logstash.agent           ] Successfully started Logstash API endpoint {:port=>9600}
  121. [2017-02-10T22:11:27,576][DEBUG][logstash.pipeline        ] Pushing flush onto pipeline
  122. [2017-02-10T22:11:30,157][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  123. ......repeats n times ........
  124. [2017-02-10T22:11:30,198][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  125. [2017-02-10T22:11:30,197][DEBUG][logstash.pipeline        ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.155Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [ScheduledTasks:1] 2017-02-10 16:11:15,484 GCInspector.java (line 119) GC for ConcurrentMarkSweep: 833 ms for 1 collections, 3763451976 used; max is 4211081216\",\"offset\":9776849,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
  126. [2017-02-10T22:11:30,198][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  127. ......repeats n times ........
  128. [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  129. [2017-02-10T22:11:30,199][DEBUG][logstash.pipeline        ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.158Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [FlushWriter:14] 2017-02-10 16:11:15,486 Memtable.java (line 398) Writing Memtable-Memberships@804510876(12986/207700 serialized/live bytes, 110 ops)\",\"offset\":9777706,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
  130. [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  131. [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  132. [2017-02-10T22:11:30,199][DEBUG][logstash.pipeline        ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.159Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" WARN [ScheduledTasks:1] 2017-02-10 16:11:16,486 GCInspector.java (line 142) Heap is 0.981591859186788 full.  You may need to reduce memtable and/or cache sizes.  app2 will now flush up to the two largest memtables to free up memory.  Adjust flush_largest_memtables_at threshold in app2.yaml if you don't want app2 to do this automatically\",\"offset\":9778327,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
  133. [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  134. [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  135. ......repeats n times ........
  136. [2017-02-10T22:11:30,673][DEBUG][logstash.pipeline        ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.666Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:28.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:28,684 FailureDetector.java:456 - Ignoring interval time of 2254218147 for /10.173.25.126\",\"offset\":17510710,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
  137. [2017-02-10T22:11:30,673][DEBUG][logstash.pipeline        ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.666Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:29.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:29,271 FailureDetector.java:456 - Ignoring interval time of 2841082158 for /10.86.97.31\",\"offset\":17510837,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
  138. [2017-02-10T22:11:30,674][DEBUG][logstash.pipeline        ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.667Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:27.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:27,684 FailureDetector.java:456 - Ignoring interval time of 2745074644 for /10.86.97.16\",\"offset\":17510325,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
  139. [2017-02-10T22:11:30,706][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  140. [2017-02-10T22:11:30,707][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
  141. ......repeats n times ........
  142. [2017-02-10T22:12:13,140][DEBUG][logstash.pipeline        ] output received {"event"=>{"@timestamp"=>2017-02-10T22:12:13.133Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:12:11.291Z\",\"beat\":{\"hostname\":\"applicationbhv29.\",\"name\":\"applicationbhv29.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [FlushWriter:64] 2017-02-10 22:12:10,799 Memtable.java (line 398) Writing Memtable-Memberships@341287112(9030/103368 serialized/live bytes, 81 ops)\",\"offset\":11712605,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
  143. [2017-02-10T22:12:13,521][WARN ][logstash.runner          ] SIGTERM received. Shutting down the agent.
  144. [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.os] PeriodicPoller: Stopping
  145. [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.jvm] PeriodicPoller: Stopping
  146. [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.persistentqueue] PeriodicPoller: Stopping
  147. [2017-02-10T22:12:13,524][WARN ][logstash.agent           ] stopping pipeline {:id=>"main"}
  148. [2017-02-10T22:12:13,524][DEBUG][logstash.pipeline        ] Closing inputs
  149. [2017-02-10T22:12:13,524][DEBUG][logstash.inputs.kafka    ] stopping {:plugin=>"LogStash::Inputs::Kafka"}
  150. [2017-02-10T22:12:13,524][DEBUG][logstash.pipeline        ] Closed inputs
  151. [2017-02-10T22:12:13,531][DEBUG][logstash.inputs.kafka    ] closing {:plugin=>"LogStash::Inputs::Kafka"}
  152. [2017-02-10T22:12:13,531][DEBUG][logstash.pipeline        ] Input plugins stopped! Will shutdown filter/output workers.
  153. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing flush onto pipeline
  154. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x7bd0f99c sleep>"}
  155. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x7fceb418 sleep>"}
  156. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x4d84c1ca sleep>"}
  157. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x7e3ca6d6 sleep>"}
  158. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x2bd9248 run>"}
  159. [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline        ] Pushing shutdown {:thread=>"#<Thread:0x28d4eb3c sleep>"}
  160. [2017-02-10T22:12:13,571][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x7bd0f99c>
  161. [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x7fceb418>
  162. [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x4d84c1ca>
  163. [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x7e3ca6d6>
  164. [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x2bd9248>
  165. [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline        ] Shutdown waiting for worker thread #<Thread:0x28d4eb3c>
  166. [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] closing {:plugin=>"LogStash::Outputs::ElasticSearch"}
  167. [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] Stopping sniffer
  168. [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] Stopping resurrectionist
  169. [2017-02-10T22:12:14,557][DEBUG][logstash.outputs.elasticsearch] Waiting for in use manticore connections
  170. [2017-02-10T22:12:14,557][DEBUG][logstash.outputs.elasticsearch] Closing adapter #<LogStash::Outputs::ElasticSearch::HttpClient::ManticoreAdapter:0x57b02f5d>
  171. [2017-02-10T22:12:14,557][DEBUG][logstash.pipeline        ] Pipeline main has been shutdown
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement