Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [2018-03-16T16:14:06,633][INFO ][logstash.modules.scaffold] Initializing module {:module_name=>"fb_apache", :directory=>"/usr/share/logstash/modules/fb_apache/configuration"}
- [2018-03-16T16:14:06,638][DEBUG][logstash.plugins.registry] Adding plugin to the registry {:name=>"fb_apache", :type=>:modules, :class=>#<LogStash::Modules::Scaffold:0x3d271b2 @module_name="fb_apache", @directory="/usr/share/logstash/modules/fb_apache/configuration", @kibana_version_parts=["6", "0", "0"]>}
- [2018-03-16T16:14:06,639][INFO ][logstash.modules.scaffold] Initializing module {:module_name=>"netflow", :directory=>"/usr/share/logstash/modules/netflow/configuration"}
- [2018-03-16T16:14:06,639][DEBUG][logstash.plugins.registry] Adding plugin to the registry {:name=>"netflow", :type=>:modules, :class=>#<LogStash::Modules::Scaffold:0x2f48752d @module_name="netflow", @directory="/usr/share/logstash/modules/netflow/configuration", @kibana_version_parts=["6", "0", "0"]>}
- [2018-03-16T16:14:06,826][DEBUG][logstash.runner ] -------- Logstash Settings (* means modified) ---------
- [2018-03-16T16:14:06,826][DEBUG][logstash.runner ] node.name: "desktop-20"
- [2018-03-16T16:14:06,826][DEBUG][logstash.runner ] *path.data: "/var/lib/logstash" (default: "/usr/share/logstash/data")
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] modules.cli: []
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] modules: []
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] modules_setup: false
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] config.test_and_exit: false
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] config.reload.automatic: false
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] config.reload.interval: 3000000000
- [2018-03-16T16:14:06,827][DEBUG][logstash.runner ] config.support_escapes: false
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] metric.collect: true
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] pipeline.id: "main"
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] pipeline.system: false
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] pipeline.workers: 4
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] pipeline.output.workers: 1
- [2018-03-16T16:14:06,828][DEBUG][logstash.runner ] pipeline.batch.size: 125
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] pipeline.batch.delay: 50
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] pipeline.unsafe_shutdown: false
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] pipeline.java_execution: false
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] pipeline.reloadable: true
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] path.plugins: []
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] config.debug: false
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] *log.level: "debug" (default: "info")
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] version: false
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] help: false
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] log.format: "plain"
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] http.host: "127.0.0.1"
- [2018-03-16T16:14:06,829][DEBUG][logstash.runner ] http.port: 9600..9700
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] http.environment: "production"
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.type: "memory"
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.drain: false
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.page_capacity: 67108864
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.max_bytes: 1073741824
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.max_events: 0
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.checkpoint.acks: 1024
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.checkpoint.writes: 1024
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] queue.checkpoint.interval: 1000
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] dead_letter_queue.enable: false
- [2018-03-16T16:14:06,830][DEBUG][logstash.runner ] dead_letter_queue.max_bytes: 1073741824
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] slowlog.threshold.warn: -1
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] slowlog.threshold.info: -1
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] slowlog.threshold.debug: -1
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] slowlog.threshold.trace: -1
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] keystore.classname: "org.logstash.secret.store.backend.JavaKeyStore"
- [2018-03-16T16:14:06,831][DEBUG][logstash.runner ] keystore.file: "/usr/share/logstash/config/logstash.keystore"
- [2018-03-16T16:14:06,832][DEBUG][logstash.runner ] *path.queue: "/var/lib/logstash/queue" (default: "/usr/share/logstash/data/queue")
- [2018-03-16T16:14:06,832][DEBUG][logstash.runner ] *path.dead_letter_queue: "/var/lib/logstash/dead_letter_queue" (default: "/usr/share/logstash/data/dead_letter_queue")
- [2018-03-16T16:14:06,832][DEBUG][logstash.runner ] path.settings: "/usr/share/logstash/config"
- [2018-03-16T16:14:06,832][DEBUG][logstash.runner ] *path.logs: "/var/log/logstash" (default: "/usr/share/logstash/logs")
- [2018-03-16T16:14:06,832][DEBUG][logstash.runner ] --------------- Logstash Settings -------------------
- [2018-03-16T16:14:06,834][DEBUG][logstash.config.source.multilocal] Reading pipeline configurations from YAML {:location=>"/usr/share/logstash/config/pipelines.yml"}
- [2018-03-16T16:14:06,855][DEBUG][logstash.agent ] Setting up metric collection
- [2018-03-16T16:14:06,861][DEBUG][logstash.instrument.periodicpoller.os] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-16T16:14:06,895][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:06,917][DEBUG][logstash.instrument.periodicpoller.jvm] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-16T16:14:06,958][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:06,959][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:06,973][DEBUG][logstash.instrument.periodicpoller.persistentqueue] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-16T16:14:06,977][DEBUG][logstash.instrument.periodicpoller.deadletterqueue] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-16T16:14:06,988][INFO ][logstash.runner ] Starting Logstash {"logstash.version"=>"6.2.2"}
- [2018-03-16T16:14:06,993][DEBUG][logstash.agent ] Starting agent
- [2018-03-16T16:14:07,001][DEBUG][logstash.config.source.multilocal] Reading pipeline configurations from YAML {:location=>"/usr/share/logstash/config/pipelines.yml"}
- [2018-03-16T16:14:07,031][DEBUG][logstash.agent ] Starting puma
- [2018-03-16T16:14:07,037][DEBUG][logstash.agent ] Trying to start WebServer {:port=>9600}
- [2018-03-16T16:14:07,046][DEBUG][logstash.config.source.local.configpathloader] Skipping the following files while reading config since they don't match the specified glob pattern {:files=>[]}
- [2018-03-16T16:14:07,050][DEBUG][logstash.config.source.local.configpathloader] Reading config file {:config_file=>"/etc/logstash/conf.d/test.conf"}
- [2018-03-16T16:14:07,077][DEBUG][logstash.api.service ] [api-service] start
- [2018-03-16T16:14:07,075][DEBUG][logstash.agent ] Converging pipelines state {:actions_count=>1}
- [2018-03-16T16:14:07,081][DEBUG][logstash.agent ] Executing action {:action=>LogStash::PipelineAction::Create/pipeline_id:main}
- [2018-03-16T16:14:07,143][INFO ][logstash.agent ] Successfully started Logstash API endpoint {:port=>9600}
- [2018-03-16T16:14:07,386][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"file", :type=>"input", :class=>LogStash::Inputs::File}
- [2018-03-16T16:14:07,419][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"json", :type=>"codec", :class=>LogStash::Codecs::JSON}
- [2018-03-16T16:14:07,423][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@id = "json_430b7978-1549-4715-8a4d-64a9a96533d3"
- [2018-03-16T16:14:07,423][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@enable_metric = true
- [2018-03-16T16:14:07,423][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@charset = "UTF-8"
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@codec = <LogStash::Codecs::JSON id=>"json_430b7978-1549-4715-8a4d-64a9a96533d3", enable_metric=>true, charset=>"UTF-8">
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@path = ["/tmp/sleep.json"]
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@sincedb_path = "/dev/null"
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@id = "64ad4add776cbefaa7da6e09a6f9ff164f8ce03638fbdba0c46979acf6399e72"
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@enable_metric = true
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@add_field = {}
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@stat_interval = 1
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@discover_interval = 15
- [2018-03-16T16:14:07,426][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@sincedb_write_interval = 15
- [2018-03-16T16:14:07,427][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@start_position = "end"
- [2018-03-16T16:14:07,427][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@delimiter = "\n"
- [2018-03-16T16:14:07,427][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@close_older = 3600
- [2018-03-16T16:14:07,444][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"date", :type=>"filter", :class=>LogStash::Filters::Date}
- [2018-03-16T16:14:07,450][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@match = ["dateOfSleep", "yyyy-dd-MM"]
- [2018-03-16T16:14:07,450][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@id = "7549648263495dd9081f3a436b0dc54413aa84229a75babf834fd4628c7ce75c"
- [2018-03-16T16:14:07,450][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@enable_metric = true
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@add_tag = []
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@remove_tag = []
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@add_field = {}
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@remove_field = []
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@periodic_flush = false
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@target = "@timestamp"
- [2018-03-16T16:14:07,451][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@tag_on_failure = ["_dateparsefailure"]
- [2018-03-16T16:14:07,464][DEBUG][org.logstash.filters.DateFilter] Date filter with format=yyyy-dd-MM, locale=null, timezone=null built as org.logstash.filters.parser.JodaParser
- [2018-03-16T16:14:07,537][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"elasticsearch", :type=>"output", :class=>LogStash::Outputs::ElasticSearch}
- [2018-03-16T16:14:07,572][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"plain", :type=>"codec", :class=>LogStash::Codecs::Plain}
- [2018-03-16T16:14:07,583][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@id = "plain_b39c8a7f-e482-469c-9218-ff1772715334"
- [2018-03-16T16:14:07,583][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@enable_metric = true
- [2018-03-16T16:14:07,583][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@charset = "UTF-8"
- [2018-03-16T16:14:07,591][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@hosts = [//localhost]
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@index = "sleep_quality_two"
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@id = "296b9c76ed195876db4a5100159b06d92269e4dbd4c96b32f45aa11cf85821b0"
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@enable_metric = true
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@codec = <LogStash::Codecs::Plain id=>"plain_b39c8a7f-e482-469c-9218-ff1772715334", enable_metric=>true, charset=>"UTF-8">
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@workers = 1
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@manage_template = true
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_name = "logstash"
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_overwrite = false
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@parent = nil
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@join_field = nil
- [2018-03-16T16:14:07,592][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@upsert = ""
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@doc_as_upsert = false
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script = ""
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_type = "inline"
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_lang = "painless"
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_var_name = "event"
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@scripted_upsert = false
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_initial_interval = 2
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_max_interval = 64
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_on_conflict = 1
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pipeline = nil
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@action = "index"
- [2018-03-16T16:14:07,593][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@ssl_certificate_verification = true
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing = false
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing_delay = 5
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@timeout = 60
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@failure_type_logging_whitelist = []
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max = 1000
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max_per_route = 100
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@resurrect_delay = 5
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@validate_after_inactivity = 10000
- [2018-03-16T16:14:07,594][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@http_compression = false
- [2018-03-16T16:14:07,630][INFO ][logstash.pipeline ] Starting pipeline {:pipeline_id=>"main", "pipeline.workers"=>4, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>50}
- [2018-03-16T16:14:07,656][DEBUG][logstash.outputs.elasticsearch] Normalizing http path {:path=>nil, :normalized=>nil}
- [2018-03-16T16:14:07,811][INFO ][logstash.outputs.elasticsearch] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://localhost:9200/]}}
- [2018-03-16T16:14:07,813][INFO ][logstash.outputs.elasticsearch] Running health check to see if an Elasticsearch connection is working {:healthcheck_url=>http://localhost:9200/, :path=>"/"}
- [2018-03-16T16:14:07,875][WARN ][logstash.outputs.elasticsearch] Restored connection to ES instance {:url=>"http://localhost:9200/"}
- [2018-03-16T16:14:07,905][INFO ][logstash.outputs.elasticsearch] ES Output version determined {:es_version=>nil}
- [2018-03-16T16:14:07,906][WARN ][logstash.outputs.elasticsearch] Detected a 6.x and above cluster: the `type` event field won't be used to determine the document _type {:es_version=>6}
- [2018-03-16T16:14:07,915][INFO ][logstash.outputs.elasticsearch] Using mapping template from {:path=>nil}
- [2018-03-16T16:14:07,918][INFO ][logstash.outputs.elasticsearch] Attempting to install template {:manage_template=>{"template"=>"logstash-*", "version"=>60001, "settings"=>{"index.refresh_interval"=>"5s"}, "mappings"=>{"_default_"=>{"dynamic_templates"=>[{"message_field"=>{"path_match"=>"message", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false}}}, {"string_fields"=>{"match"=>"*", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false, "fields"=>{"keyword"=>{"type"=>"keyword", "ignore_above"=>256}}}}}], "properties"=>{"@timestamp"=>{"type"=>"date"}, "@version"=>{"type"=>"keyword"}, "geoip"=>{"dynamic"=>true, "properties"=>{"ip"=>{"type"=>"ip"}, "location"=>{"type"=>"geo_point"}, "latitude"=>{"type"=>"half_float"}, "longitude"=>{"type"=>"half_float"}}}}}}}}
- [2018-03-16T16:14:07,926][DEBUG][logstash.outputs.elasticsearch] Found existing Elasticsearch template. Skipping template management {:name=>"logstash"}
- [2018-03-16T16:14:07,927][INFO ][logstash.outputs.elasticsearch] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>["//localhost"]}
- [2018-03-16T16:14:08,140][INFO ][logstash.pipeline ] Pipeline started succesfully {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:08,144][DEBUG][logstash.inputs.file ] _globbed_files: /tmp/sleep.json: glob is: ["/tmp/sleep.json"]
- [2018-03-16T16:14:08,146][DEBUG][logstash.inputs.file ] _discover_file: /tmp/sleep.json: new: /tmp/sleep.json (exclude is [])
- [2018-03-16T16:14:08,155][DEBUG][logstash.inputs.file ] _open_file: /tmp/sleep.json: opening
- [2018-03-16T16:14:08,156][DEBUG][logstash.inputs.file ] /tmp/sleep.json: initial create, no sincedb, seeking to end 65104
- [2018-03-16T16:14:08,158][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:08,158][INFO ][logstash.agent ] Pipelines running {:count=>1, :pipelines=>["main"]}
- [2018-03-16T16:14:09,160][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:10,162][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:11,164][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:11,937][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:11,974][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:11,975][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:12,166][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:13,134][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:13,168][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:14,170][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:15,192][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:16,195][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:16,952][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:16,992][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:16,993][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:17,197][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:18,136][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:18,198][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:19,199][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:20,201][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:21,202][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:21,966][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:22,003][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:22,003][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:22,204][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:22,207][DEBUG][logstash.inputs.file ] _globbed_files: /tmp/sleep.json: glob is: ["/tmp/sleep.json"]
- [2018-03-16T16:14:23,136][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:23,208][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:24,210][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:25,212][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:26,214][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:26,981][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:27,010][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:27,011][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:27,216][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:28,136][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:28,218][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:29,220][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:30,221][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:31,223][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:31,995][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:32,018][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:32,018][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-16T16:14:32,225][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:33,137][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x71eb58aa@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-16T16:14:33,227][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:34,229][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:35,231][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:36,233][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-16T16:14:37,006][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-16T16:14:37,025][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-16T16:14:37,026][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement