Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [2018-03-14T15:50:33,494][INFO ][logstash.modules.scaffold] Initializing module {:module_name=>"fb_apache", :directory=>"/usr/share/logstash/modules/fb_apache/configuration"}
- [2018-03-14T15:50:33,498][DEBUG][logstash.plugins.registry] Adding plugin to the registry {:name=>"fb_apache", :type=>:modules, :class=>#<LogStash::Modules::Scaffold:0x485ba178 @module_name="fb_apache", @directory="/usr/share/logstash/modules/fb_apache/configuration", @kibana_version_parts=["6", "0", "0"]>}
- [2018-03-14T15:50:33,499][INFO ][logstash.modules.scaffold] Initializing module {:module_name=>"netflow", :directory=>"/usr/share/logstash/modules/netflow/configuration"}
- [2018-03-14T15:50:33,499][DEBUG][logstash.plugins.registry] Adding plugin to the registry {:name=>"netflow", :type=>:modules, :class=>#<LogStash::Modules::Scaffold:0x714b883e @module_name="netflow", @directory="/usr/share/logstash/modules/netflow/configuration", @kibana_version_parts=["6", "0", "0"]>}
- [2018-03-14T15:50:33,749][DEBUG][logstash.runner ] -------- Logstash Settings (* means modified) ---------
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] node.name: "desktop-20"
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] *path.data: "/var/lib/logstash" (default: "/usr/share/logstash/data")
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] modules.cli: []
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] modules: []
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] modules_setup: false
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] config.test_and_exit: false
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] config.reload.automatic: false
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] config.reload.interval: 3000000000
- [2018-03-14T15:50:33,750][DEBUG][logstash.runner ] config.support_escapes: false
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] metric.collect: true
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] pipeline.id: "main"
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] pipeline.system: false
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] pipeline.workers: 4
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] pipeline.output.workers: 1
- [2018-03-14T15:50:33,753][DEBUG][logstash.runner ] pipeline.batch.size: 125
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] pipeline.batch.delay: 50
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] pipeline.unsafe_shutdown: false
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] pipeline.java_execution: false
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] pipeline.reloadable: true
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] path.plugins: []
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] config.debug: false
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] *log.level: "debug" (default: "info")
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] version: false
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] help: false
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] log.format: "plain"
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] http.host: "127.0.0.1"
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] http.port: 9600..9700
- [2018-03-14T15:50:33,754][DEBUG][logstash.runner ] http.environment: "production"
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.type: "memory"
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.drain: false
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.page_capacity: 67108864
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.max_bytes: 1073741824
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.max_events: 0
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.checkpoint.acks: 1024
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.checkpoint.writes: 1024
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] queue.checkpoint.interval: 1000
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] dead_letter_queue.enable: false
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] dead_letter_queue.max_bytes: 1073741824
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] slowlog.threshold.warn: -1
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] slowlog.threshold.info: -1
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] slowlog.threshold.debug: -1
- [2018-03-14T15:50:33,755][DEBUG][logstash.runner ] slowlog.threshold.trace: -1
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] keystore.classname: "org.logstash.secret.store.backend.JavaKeyStore"
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] keystore.file: "/usr/share/logstash/config/logstash.keystore"
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] *path.queue: "/var/lib/logstash/queue" (default: "/usr/share/logstash/data/queue")
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] *path.dead_letter_queue: "/var/lib/logstash/dead_letter_queue" (default: "/usr/share/logstash/data/dead_letter_queue")
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] path.settings: "/usr/share/logstash/config"
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] *path.logs: "/var/log/logstash" (default: "/usr/share/logstash/logs")
- [2018-03-14T15:50:33,756][DEBUG][logstash.runner ] --------------- Logstash Settings -------------------
- [2018-03-14T15:50:33,758][DEBUG][logstash.config.source.multilocal] Reading pipeline configurations from YAML {:location=>"/usr/share/logstash/config/pipelines.yml"}
- [2018-03-14T15:50:33,787][DEBUG][logstash.agent ] Setting up metric collection
- [2018-03-14T15:50:33,793][DEBUG][logstash.instrument.periodicpoller.os] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-14T15:50:33,822][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:33,845][DEBUG][logstash.instrument.periodicpoller.jvm] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-14T15:50:33,894][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:33,895][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:33,901][DEBUG][logstash.instrument.periodicpoller.persistentqueue] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-14T15:50:33,903][DEBUG][logstash.instrument.periodicpoller.deadletterqueue] Starting {:polling_interval=>5, :polling_timeout=>120}
- [2018-03-14T15:50:33,912][INFO ][logstash.runner ] Starting Logstash {"logstash.version"=>"6.2.2"}
- [2018-03-14T15:50:33,918][DEBUG][logstash.agent ] Starting agent
- [2018-03-14T15:50:33,930][DEBUG][logstash.agent ] Starting puma
- [2018-03-14T15:50:33,932][DEBUG][logstash.agent ] Trying to start WebServer {:port=>9600}
- [2018-03-14T15:50:33,951][DEBUG][logstash.config.source.multilocal] Reading pipeline configurations from YAML {:location=>"/usr/share/logstash/config/pipelines.yml"}
- [2018-03-14T15:50:33,960][DEBUG][logstash.api.service ] [api-service] start
- [2018-03-14T15:50:33,998][DEBUG][logstash.config.source.local.configpathloader] Skipping the following files while reading config since they don't match the specified glob pattern {:files=>[]}
- [2018-03-14T15:50:33,999][DEBUG][logstash.config.source.local.configpathloader] Reading config file {:config_file=>"/etc/logstash/conf.d/test.conf"}
- [2018-03-14T15:50:34,016][DEBUG][logstash.agent ] Converging pipelines state {:actions_count=>1}
- [2018-03-14T15:50:34,022][DEBUG][logstash.agent ] Executing action {:action=>LogStash::PipelineAction::Create/pipeline_id:main}
- [2018-03-14T15:50:34,064][INFO ][logstash.agent ] Successfully started Logstash API endpoint {:port=>9600}
- [2018-03-14T15:50:34,326][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"file", :type=>"input", :class=>LogStash::Inputs::File}
- [2018-03-14T15:50:34,357][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"json", :type=>"codec", :class=>LogStash::Codecs::JSON}
- [2018-03-14T15:50:34,362][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@id = "json_6fa275b6-2b17-4bc7-90ec-947f00ebb6dc"
- [2018-03-14T15:50:34,362][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@enable_metric = true
- [2018-03-14T15:50:34,363][DEBUG][logstash.codecs.json ] config LogStash::Codecs::JSON/@charset = "UTF-8"
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@codec = <LogStash::Codecs::JSON id=>"json_6fa275b6-2b17-4bc7-90ec-947f00ebb6dc", enable_metric=>true, charset=>"UTF-8">
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@path = ["/tmp/sleep.json"]
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@id = "8649245bb637ed7420a391cc534c86bd5da448699b940d4d615e60a2dd4eab68"
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@enable_metric = true
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@add_field = {}
- [2018-03-14T15:50:34,370][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@stat_interval = 1
- [2018-03-14T15:50:34,371][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@discover_interval = 15
- [2018-03-14T15:50:34,371][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@sincedb_write_interval = 15
- [2018-03-14T15:50:34,371][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@start_position = "end"
- [2018-03-14T15:50:34,371][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@delimiter = "\n"
- [2018-03-14T15:50:34,371][DEBUG][logstash.inputs.file ] config LogStash::Inputs::File/@close_older = 3600
- [2018-03-14T15:50:34,381][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"date", :type=>"filter", :class=>LogStash::Filters::Date}
- [2018-03-14T15:50:34,386][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@match = ["dateOfSleep", "yyyy-dd-MM"]
- [2018-03-14T15:50:34,386][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@id = "14a7c23f569930a5ac9f4ddc9f4eff3ae9474355228a1f86d36de87aeaa27b15"
- [2018-03-14T15:50:34,386][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@enable_metric = true
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@add_tag = []
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@remove_tag = []
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@add_field = {}
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@remove_field = []
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@periodic_flush = false
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@target = "@timestamp"
- [2018-03-14T15:50:34,387][DEBUG][logstash.filters.date ] config LogStash::Filters::Date/@tag_on_failure = ["_dateparsefailure"]
- [2018-03-14T15:50:34,401][DEBUG][org.logstash.filters.DateFilter] Date filter with format=yyyy-dd-MM, locale=null, timezone=null built as org.logstash.filters.parser.JodaParser
- [2018-03-14T15:50:34,474][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"elasticsearch", :type=>"output", :class=>LogStash::Outputs::ElasticSearch}
- [2018-03-14T15:50:34,507][DEBUG][logstash.plugins.registry] On demand adding plugin to the registry {:name=>"plain", :type=>"codec", :class=>LogStash::Codecs::Plain}
- [2018-03-14T15:50:34,516][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@id = "plain_81c0a274-74e0-4ae6-8267-3760746358a4"
- [2018-03-14T15:50:34,516][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@enable_metric = true
- [2018-03-14T15:50:34,516][DEBUG][logstash.codecs.plain ] config LogStash::Codecs::Plain/@charset = "UTF-8"
- [2018-03-14T15:50:34,520][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@hosts = [//localhost]
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@index = "sleep-quality"
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@id = "7355511a8bd532c2a8dbb909ae0f643259563cd8216cfbd1fa1fe7696b78fc0e"
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@enable_metric = true
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@codec = <LogStash::Codecs::Plain id=>"plain_81c0a274-74e0-4ae6-8267-3760746358a4", enable_metric=>true, charset=>"UTF-8">
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@workers = 1
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@manage_template = true
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_name = "logstash"
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@template_overwrite = false
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@parent = nil
- [2018-03-14T15:50:34,521][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@join_field = nil
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@upsert = ""
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@doc_as_upsert = false
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script = ""
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_type = "inline"
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_lang = "painless"
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@script_var_name = "event"
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@scripted_upsert = false
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_initial_interval = 2
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_max_interval = 64
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@retry_on_conflict = 1
- [2018-03-14T15:50:34,522][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pipeline = nil
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@action = "index"
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@ssl_certificate_verification = true
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing = false
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@sniffing_delay = 5
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@timeout = 60
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@failure_type_logging_whitelist = []
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max = 1000
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@pool_max_per_route = 100
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@resurrect_delay = 5
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@validate_after_inactivity = 10000
- [2018-03-14T15:50:34,523][DEBUG][logstash.outputs.elasticsearch] config LogStash::Outputs::ElasticSearch/@http_compression = false
- [2018-03-14T15:50:34,544][INFO ][logstash.pipeline ] Starting pipeline {:pipeline_id=>"main", "pipeline.workers"=>4, "pipeline.batch.size"=>125, "pipeline.batch.delay"=>50}
- [2018-03-14T15:50:34,570][DEBUG][logstash.outputs.elasticsearch] Normalizing http path {:path=>nil, :normalized=>nil}
- [2018-03-14T15:50:34,724][INFO ][logstash.outputs.elasticsearch] Elasticsearch pool URLs updated {:changes=>{:removed=>[], :added=>[http://localhost:9200/]}}
- [2018-03-14T15:50:34,726][INFO ][logstash.outputs.elasticsearch] Running health check to see if an Elasticsearch connection is working {:healthcheck_url=>http://localhost:9200/, :path=>"/"}
- [2018-03-14T15:50:34,805][WARN ][logstash.outputs.elasticsearch] Restored connection to ES instance {:url=>"http://localhost:9200/"}
- [2018-03-14T15:50:34,851][INFO ][logstash.outputs.elasticsearch] ES Output version determined {:es_version=>nil}
- [2018-03-14T15:50:34,851][WARN ][logstash.outputs.elasticsearch] Detected a 6.x and above cluster: the `type` event field won't be used to determine the document _type {:es_version=>6}
- [2018-03-14T15:50:34,853][INFO ][logstash.outputs.elasticsearch] Using mapping template from {:path=>nil}
- [2018-03-14T15:50:34,858][INFO ][logstash.outputs.elasticsearch] Attempting to install template {:manage_template=>{"template"=>"logstash-*", "version"=>60001, "settings"=>{"index.refresh_interval"=>"5s"}, "mappings"=>{"_default_"=>{"dynamic_templates"=>[{"message_field"=>{"path_match"=>"message", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false}}}, {"string_fields"=>{"match"=>"*", "match_mapping_type"=>"string", "mapping"=>{"type"=>"text", "norms"=>false, "fields"=>{"keyword"=>{"type"=>"keyword", "ignore_above"=>256}}}}}], "properties"=>{"@timestamp"=>{"type"=>"date"}, "@version"=>{"type"=>"keyword"}, "geoip"=>{"dynamic"=>true, "properties"=>{"ip"=>{"type"=>"ip"}, "location"=>{"type"=>"geo_point"}, "latitude"=>{"type"=>"half_float"}, "longitude"=>{"type"=>"half_float"}}}}}}}}
- [2018-03-14T15:50:34,865][DEBUG][logstash.outputs.elasticsearch] Found existing Elasticsearch template. Skipping template management {:name=>"logstash"}
- [2018-03-14T15:50:34,866][INFO ][logstash.outputs.elasticsearch] New Elasticsearch output {:class=>"LogStash::Outputs::ElasticSearch", :hosts=>["//localhost"]}
- [2018-03-14T15:50:35,116][INFO ][logstash.pipeline ] Pipeline started succesfully {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:50:35,143][INFO ][logstash.agent ] Pipelines running {:count=>1, :pipelines=>["main"]}
- [2018-03-14T15:50:35,164][DEBUG][logstash.inputs.file ] _globbed_files: /tmp/sleep.json: glob is: ["/tmp/sleep.json"]
- [2018-03-14T15:50:35,166][DEBUG][logstash.inputs.file ] _discover_file: /tmp/sleep.json: new: /tmp/sleep.json (exclude is [])
- [2018-03-14T15:50:35,176][DEBUG][logstash.inputs.file ] _open_file: /tmp/sleep.json: opening
- [2018-03-14T15:50:35,177][DEBUG][logstash.inputs.file ] /tmp/sleep.json: initial create, no sincedb, seeking to end 65104
- [2018-03-14T15:50:35,180][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:36,181][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:37,182][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:38,184][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:38,863][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:38,927][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:38,937][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:39,186][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:40,151][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:50:40,188][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:41,190][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:42,192][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:43,195][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:43,881][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:43,945][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:43,954][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:44,197][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:45,151][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:50:45,199][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:46,201][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:47,202][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:48,204][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:48,895][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:48,971][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:48,972][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:49,206][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:49,209][DEBUG][logstash.inputs.file ] _globbed_files: /tmp/sleep.json: glob is: ["/tmp/sleep.json"]
- [2018-03-14T15:50:50,151][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:50:50,210][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:51,212][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:52,213][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:53,215][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:53,910][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:53,979][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:53,979][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:54,217][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:55,152][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:50:55,219][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:56,221][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:57,223][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:58,225][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:50:58,922][DEBUG][logstash.instrument.periodicpoller.cgroup] Error, cannot retrieve cgroups information {:exception=>"NoMethodError", :message=>"undefined method `[]' for nil:NilClass"}
- [2018-03-14T15:50:58,987][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ParNew"}
- [2018-03-14T15:50:58,988][DEBUG][logstash.instrument.periodicpoller.jvm] collector name {:name=>"ConcurrentMarkSweep"}
- [2018-03-14T15:50:59,227][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:51:00,152][DEBUG][logstash.pipeline ] Pushing flush onto pipeline {:pipeline_id=>"main", :thread=>"#<Thread:0x596b1d87@/usr/share/logstash/logstash-core/lib/logstash/pipeline.rb:246 sleep>"}
- [2018-03-14T15:51:00,228][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- [2018-03-14T15:51:01,231][DEBUG][logstash.inputs.file ] each: file grew: /tmp/sleep.json: old size 0, new size 65104
- ^C
- ➜ logstash
Advertisement
Add Comment
Please, Sign In to add comment