Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [2017-02-10T22:11:22,086][DEBUG][logstash.runner ] -------- Logstash Settings (* means modified) ---------
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] node.name: "logstash-indexer1-fqdn"
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] *path.config: "/etc/logstash/conf.d"
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] *path.data: "/var/lib/logstash" (default: "/usr/share/logstash/data")
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] config.test_and_exit: false
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] config.reload.automatic: false
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] config.reload.interval: 3
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] metric.collect: true
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] pipeline.id: "main"
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] *pipeline.workers: 6 (default: 8)
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] pipeline.output.workers: 1
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] pipeline.batch.size: 125
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] pipeline.batch.delay: 5
- [2017-02-10T22:11:22,088][DEBUG][logstash.runner ] pipeline.unsafe_shutdown: false
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] path.plugins: []
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] config.debug: false
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] *log.level: "debug" (default: "info")
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] version: false
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] help: false
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] log.format: "plain"
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] http.host: "127.0.0.1"
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] http.port: 9600..9700
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] http.environment: "production"
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.type: "memory"
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.page_capacity: 262144000
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.max_bytes: 1073741824
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.max_events: 0
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.checkpoint.acks: 1024
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.checkpoint.writes: 1024
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] queue.checkpoint.interval: 1000
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] slowlog.threshold.warn: -1
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] slowlog.threshold.info: -1
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] slowlog.threshold.debug: -1
- [2017-02-10T22:11:22,089][DEBUG][logstash.runner ] slowlog.threshold.trace: -1
- [2017-02-10T22:11:22,090][DEBUG][logstash.runner ] path.queue: "/usr/share/logstash/data/queue"
- [2017-02-10T22:11:22,090][DEBUG][logstash.runner ] *path.settings: "/etc/logstash" (default: "/usr/share/logstash/config")
- [2017-02-10T22:11:22,090][DEBUG][logstash.runner ] *path.logs: "/var/log/logstash" (default: "/usr/share/logstash/logs")
- [2017-02-10T22:11:22,090][DEBUG][logstash.runner ] --------------- Logstash Settings -------------------
- [2017-02-10T22:11:22,097][DEBUG][logstash.agent ] Agent: Configuring metric collection
- [2017-02-10T22:11:22,098][DEBUG][logstash.instrument.periodicpoller.os] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
- [2017-02-10T22:11:22,115][DEBUG][logstash.instrument.periodicpoller.jvm] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
- [2017-02-10T22:11:22,144][DEBUG][logstash.instrument.periodicpoller.persistentqueue] PeriodicPoller: Starting {:polling_interval=>5, :polling_timeout=>120}
- [2017-02-10T22:11:22,148][DEBUG][logstash.agent ] Reading config file {:config_file=>"/etc/logstash/conf.d/09_input_kafka_cluster.yml"}
- [2017-02-10T22:11:22,148][DEBUG][logstash.agent ] Reading config file {:config_file=>"/etc/logstash/conf.d/50_filter_pending.yml"}
- [2017-02-10T22:11:22,149][DEBUG][logstash.agent ] Reading config file {:config_file=>"/etc/logstash/conf.d/80_output_elastic.yml"}
- [2017-02-10T22:11:22,187][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"kafka", :type=>"input", :class=>LogStash::Inputs::Kafka}
- [2017-02-10T22:11:22,195][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"plain", :type=>"codec", :class=>LogStash::Codecs::Plain}
- [2017-02-10T22:11:22,197][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@id = "plain_c6eb5a12-536e-4442-9c98-5bc73f1f31da"
- [2017-02-10T22:11:22,198][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@enable_metric = true
- [2017-02-10T22:11:22,198][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@charset = "UTF-8"
- [2017-02-10T22:11:22,199][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@bootstrap_servers = "kafka3-fqdn:9092,kafka1-fqdn:9092,kafka2-fqdn:9092"
- [2017-02-10T22:11:22,199][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@topics = ["clusterlogs"]
- [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@client_id = "logstash-indexer1-fqdn"
- [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@group_id = "logstash_indexer"
- [2017-02-10T22:11:22,200][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@add_field = {"log_origin"=>"kafka"}
- [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@id = "cd8451de5cf0159c42b614a3e21258137a6e2e5e-1"
- [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@enable_metric = true
- [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@codec = <LogStash::Codecs::Plain id=>"plain_c6eb5a12-536e-4442-9c98-5bc73f1f31da", enable_metric=>true, charset=>"UTF-8">
- [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@auto_commit_interval_ms = "5000"
- [2017-02-10T22:11:22,201][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@consumer_threads = 1
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@enable_auto_commit = "true"
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@key_deserializer_class = "org.apache.kafka.common.serialization.StringDeserializer"
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@value_deserializer_class = "org.apache.kafka.common.serialization.StringDeserializer"
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@poll_timeout_ms = 100
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@ssl = false
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@security_protocol = "PLAINTEXT"
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@sasl_mechanism = "GSSAPI"
- [2017-02-10T22:11:22,202][DEBUG][logstash.inputs.kafka ] config LogStash::Inputs::Kafka/@decorate_events = false
- [2017-02-10T22:11:22,281][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"elasticsearch", :type=>"output", :class=>LogStash::Outputs::ElasticSearch}
- [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@id = "plain_506a9778-3738-4f33-bb90-9b058de6eb7f"
- [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@enable_metric = true
- [2017-02-10T22:11:22,285][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@charset = "UTF-8"
- [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@hosts = [//10.86.97.128:9200]
- [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@manage_template = false
- [2017-02-10T22:11:22,287][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@index = "%{[@metadata][beat]}-%{+YYYY.MM.dd}"
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@document_type = "%{[@metadata][type]}"
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@id = "cd8451de5cf0159c42b614a3e21258137a6e2e5e-2"
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@enable_metric = true
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@codec = <LogStash::Codecs::Plain id=>"plain_506a9778-3738-4f33-bb90-9b058de6eb7f", enable_metric=>true, charset=>"UTF-8">
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@workers = 1
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_name = "logstash"
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_overwrite = false
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@parent = nil
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@idle_flush_time = 1
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@upsert = ""
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@doc_as_upsert = false
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script = ""
- [2017-02-10T22:11:22,288][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_type = "inline"
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_lang = "painless"
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_var_name = "event"
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@scripted_upsert = false
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_initial_interval = 2
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_max_interval = 64
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_on_conflict = 1
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pipeline = nil
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@action = "index"
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@ssl_certificate_verification = true
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing = false
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing_delay = 5
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@timeout = 60
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@failure_type_logging_whitelist = []
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max = 1000
- [2017-02-10T22:11:22,289][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max_per_route = 100
- [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@healthcheck_path = "/"
- [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@absolute_healthcheck_path = false
- [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@resurrect_delay = 5
- [2017-02-10T22:11:22,290][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@validate_after_inactivity = 10000
- [2017-02-10T22:11:22,298][DEBUG][logstash.agent ] starting agent
- [2017-02-10T22:11:22,299][DEBUG][logstash.agent ] starting pipeline {:id=>"main"}
- [2017-02-10T22:11:22,304][DEBUG][logstash.outputs.elasticsearch] Normalizing http path {:path=>nil, :normalized=>nil}
- [2017-02-10T22:11:22,501][INFO ][logstash.outputs.elasticsearch] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://10.86.97.128:9200/]}}
- [2017-02-10T22:11:22,501][INFO ][logstash.outputs.elasticsearch] Running health check to see if an Elasticsearch connection is working {:healthcheck_url=>http://10.86.97.128:9200/, :path=>"/"}
- [2017-02-10T22:11:22,550][WARN ][logstash.outputs.elasticsearch] Restored connection to ES instance {:url=>#<URI::HTTP:0x5e1e5983 URL:http://10.86.97.128:9200/>}
- [2017-02-10T22:11:22,551][INFO ][logstash.outputs.elasticsearch] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>[#<URI::Generic:0x3eb5a14 URL://10.86.97.128:9200>]}
- [2017-02-10T22:11:22,573][INFO ][logstash.pipeline ] Starting pipeline {"id"=>"main", "pipeline.workers"=>6, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>5, "pipeline.max_inflight"=>750}
- [2017-02-10T22:11:22,575][INFO ][logstash.pipeline ] Pipeline main started
- [2017-02-10T22:11:22,579][DEBUG][logstash.agent ] Starting puma
- [2017-02-10T22:11:22,579][DEBUG][logstash.agent ] Trying to start WebServer {:port=>9600}
- [2017-02-10T22:11:22,580][DEBUG][logstash.api.service ] [api-service] start
- [2017-02-10T22:11:22,593][INFO ][logstash.agent ] Successfully started Logstash API endpoint {:port=>9600}
- [2017-02-10T22:11:27,576][DEBUG][logstash.pipeline ] Pushing flush onto pipeline
- [2017-02-10T22:11:30,157][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- ......repeats n times ........
- [2017-02-10T22:11:30,198][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,197][DEBUG][logstash.pipeline ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.155Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [ScheduledTasks:1] 2017-02-10 16:11:15,484 GCInspector.java (line 119) GC for ConcurrentMarkSweep: 833 ms for 1 collections, 3763451976 used; max is 4211081216\",\"offset\":9776849,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,198][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- ......repeats n times ........
- [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,199][DEBUG][logstash.pipeline ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.158Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [FlushWriter:14] 2017-02-10 16:11:15,486 Memtable.java (line 398) Writing Memtable-Memberships@804510876(12986/207700 serialized/live bytes, 110 ops)\",\"offset\":9777706,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,199][DEBUG][logstash.pipeline ] filter received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.159Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:16.991Z\",\"beat\":{\"hostname\":\"applicationbhv6.\",\"name\":\"applicationbhv6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" WARN [ScheduledTasks:1] 2017-02-10 16:11:16,486 GCInspector.java (line 142) Heap is 0.981591859186788 full. You may need to reduce memtable and/or cache sizes. app2 will now flush up to the two largest memtables to free up memory. Adjust flush_largest_memtables_at threshold in app2.yaml if you don't want app2 to do this automatically\",\"offset\":9778327,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,199][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- ......repeats n times ........
- [2017-02-10T22:11:30,673][DEBUG][logstash.pipeline ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.666Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:28.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:28,684 FailureDetector.java:456 - Ignoring interval time of 2254218147 for /10.173.25.126\",\"offset\":17510710,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,673][DEBUG][logstash.pipeline ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.666Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:29.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:29,271 FailureDetector.java:456 - Ignoring interval time of 2841082158 for /10.86.97.31\",\"offset\":17510837,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,674][DEBUG][logstash.pipeline ] output received {"event"=>{"@timestamp"=>2017-02-10T22:11:30.667Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:11:27.810Z\",\"beat\":{\"hostname\":\"cache6.\",\"name\":\"cache6.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\"DEBUG [GossipStage:1] 2017-02-10 22:11:27,684 FailureDetector.java:456 - Ignoring interval time of 2745074644 for /10.86.97.16\",\"offset\":17510325,\"source\":\"/var/log/app2/debug.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:11:30,706][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- [2017-02-10T22:11:30,707][DEBUG][logstash.util.decorators ] inputs/LogStash::Inputs::Kafka: adding value to field {"field"=>"log_origin", "value"=>["kafka"]}
- ......repeats n times ........
- [2017-02-10T22:12:13,140][DEBUG][logstash.pipeline ] output received {"event"=>{"@timestamp"=>2017-02-10T22:12:13.133Z, "log_origin"=>"kafka", "@version"=>"1", "message"=>"{\"@timestamp\":\"2017-02-10T22:12:11.291Z\",\"beat\":{\"hostname\":\"applicationbhv29.\",\"name\":\"applicationbhv29.\",\"version\":\"5.2.0\"},\"input_type\":\"log\",\"message\":\" INFO [FlushWriter:64] 2017-02-10 22:12:10,799 Memtable.java (line 398) Writing Memtable-Memberships@341287112(9030/103368 serialized/live bytes, 81 ops)\",\"offset\":11712605,\"source\":\"/var/log/app2/system.log\",\"type\":\"clusterlogs\"}"}}
- [2017-02-10T22:12:13,521][WARN ][logstash.runner ] SIGTERM received. Shutting down the agent.
- [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.os] PeriodicPoller: Stopping
- [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.jvm] PeriodicPoller: Stopping
- [2017-02-10T22:12:13,522][DEBUG][logstash.instrument.periodicpoller.persistentqueue] PeriodicPoller: Stopping
- [2017-02-10T22:12:13,524][WARN ][logstash.agent ] stopping pipeline {:id=>"main"}
- [2017-02-10T22:12:13,524][DEBUG][logstash.pipeline ] Closing inputs
- [2017-02-10T22:12:13,524][DEBUG][logstash.inputs.kafka ] stopping {:plugin=>"LogStash::Inputs::Kafka"}
- [2017-02-10T22:12:13,524][DEBUG][logstash.pipeline ] Closed inputs
- [2017-02-10T22:12:13,531][DEBUG][logstash.inputs.kafka ] closing {:plugin=>"LogStash::Inputs::Kafka"}
- [2017-02-10T22:12:13,531][DEBUG][logstash.pipeline ] Input plugins stopped! Will shutdown filter/output workers.
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing flush onto pipeline
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x7bd0f99c sleep>"}
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x7fceb418 sleep>"}
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x4d84c1ca sleep>"}
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x7e3ca6d6 sleep>"}
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x2bd9248 run>"}
- [2017-02-10T22:12:13,570][DEBUG][logstash.pipeline ] Pushing shutdown {:thread=>"#<Thread:0x28d4eb3c sleep>"}
- [2017-02-10T22:12:13,571][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x7bd0f99c>
- [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x7fceb418>
- [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x4d84c1ca>
- [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x7e3ca6d6>
- [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x2bd9248>
- [2017-02-10T22:12:13,604][DEBUG][logstash.pipeline ] Shutdown waiting for worker thread #<Thread:0x28d4eb3c>
- [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] closing {:plugin=>"LogStash::Outputs::ElasticSearch"}
- [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] Stopping sniffer
- [2017-02-10T22:12:13,604][DEBUG][logstash.outputs.elasticsearch] Stopping resurrectionist
- [2017-02-10T22:12:14,557][DEBUG][logstash.outputs.elasticsearch] Waiting for in use manticore connections
- [2017-02-10T22:12:14,557][DEBUG][logstash.outputs.elasticsearch] Closing adapter #<LogStash::Outputs::ElasticSearch::HttpClient::ManticoreAdapter:0x57b02f5d>
- [2017-02-10T22:12:14,557][DEBUG][logstash.pipeline ] Pipeline main has been shutdown
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement