Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- :message=>"Reading config file", :file=>"logstash/agent.rb", :level=>:debug, :line=>"301"}
- :message=>"Compiled pipeline code:\n@inputs = []\n@filters = []\n@outputs = []\n@input_file_1 = plugin(\"input\", \"file\", LogStash::Util.hash_merge_many({ \"path\" => (\"/var/log/apache2/access.log\".force_encoding(\"UTF-8\")) }))\n\n@inputs << @input_file_1\n@filter_grok_2 = plugin(\"filter\", \"grok\", LogStash::Util.hash_merge_many({ \"match\" => {(\"message\".force_encoding(\"UTF-8\")) => (\"%{COMBINEDAPACHELOG}\".force_encoding(\"UTF-8\"))} }))\n\n@filters << @filter_grok_2\n@filter_date_3 = plugin(\"filter\", \"date\", LogStash::Util.hash_merge_many({ \"match\" => [(\"timestamp\".force_encoding(\"UTF-8\")), (\"dd/MMM/yyyy:HH:mm:ss Z\".force_encoding(\"UTF-8\"))] }))\n\n@filters << @filter_date_3\n@output_elasticsearch_4 = plugin(\"output\", \"elasticsearch\", LogStash::Util.hash_merge_many({ \"host\" => (\"localhost\".force_encoding(\"UTF-8\")) }))\n\n@outputs << @output_elasticsearch_4\n@output_stdout_5 = plugin(\"output\", \"stdout\", LogStash::Util.hash_merge_many({ \"codec\" => (\"rubydebug\".force_encoding(\"UTF-8\")) }))\n\n@outputs << @output_stdout_5\n @filter_func = lambda do |event, &block|\n extra_events = []\n @logger.debug? && @logger.debug(\"filter received\", :event => event.to_hash)\n newevents = []\n extra_events.each do |event|\n @filter_grok_2.filter(event) do |newevent|\n newevents << newevent\n end\n end\n extra_events += newevents\n @filter_grok_2.filter(event) do |newevent|\n extra_events << newevent\n end\n if event.cancelled?\n extra_events.each(&block)\n return\n end\n newevents = []\n extra_events.each do |event|\n @filter_date_3.filter(event) do |newevent|\n newevents << newevent\n end\n end\n extra_events += newevents\n @filter_date_3.filter(event) do |newevent|\n extra_events << newevent\n end\n if event.cancelled?\n extra_events.each(&block)\n return\n end\n \n extra_events.each(&block)\n end\n @output_func = lambda do |event, &block|\n @logger.debug? && @logger.debug(\"output received\", :event => event.to_hash)\n @output_elasticsearch_4.handle(event)\n @output_stdout_5.handle(event)\n \n end", :level=>:debug, :file=>"logstash/pipeline.rb", :line=>"26"}
- :message=>"Using milestone 2 input plugin 'file'. This plugin should be stable, but if you see strange behavior, please let us know! For more information on plugin milestones, see http://logstash.net/docs/1.4.2-modified/plugin-milestones", :level=>:warn, :file=>"logstash/config/mixin.rb", :line=>"209"}
- :message=>"config LogStash::Codecs::Plain/@charset = \"UTF-8\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@path = [\"/var/log/apache2/access.log\"]", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@debug = false", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@codec = <LogStash::Codecs::Plain charset=>\"UTF-8\">", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@add_field = {}", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@stat_interval = 1", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@discover_interval = 15", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@sincedb_write_interval = 15", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Inputs::File/@start_position = \"end\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@match = {\"message\"=>\"%{COMBINEDAPACHELOG}\"}", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@type = \"\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@exclude_tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@add_tag = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@remove_tag = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@add_field = {}", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@remove_field = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@patterns_dir = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@drop_if_match = false", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@break_on_match = true", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@named_captures_only = true", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@keep_empty_captures = false", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@singles = true", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@tag_on_failure = [\"_grokparsefailure\"]", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Grok/@overwrite = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@match = [\"timestamp\", \"dd/MMM/yyyy:HH:mm:ss Z\"]", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@type = \"\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@exclude_tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@add_tag = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@remove_tag = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@add_field = {}", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@remove_field = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Filters::Date/@target = \"@timestamp\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Codecs::Plain/@charset = \"UTF-8\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@host = \"localhost\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@type = \"\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@exclude_tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@codec = <LogStash::Codecs::Plain charset=>\"UTF-8\">", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@workers = 1", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@index = \"logstash-%{+YYYY.MM.dd}\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@manage_template = true", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@template_name = \"logstash\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@template_overwrite = false", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@document_id = nil", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@embedded = false", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@embedded_http_port = \"9200-9300\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@max_inflight_requests = 50", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@flush_size = 5000", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@idle_flush_time = 1", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::ElasticSearch/@action = \"index\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::Stdout/@codec = <LogStash::Codecs::RubyDebug >", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::Stdout/@type = \"\"", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::Stdout/@tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::Stdout/@exclude_tags = []", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"config LogStash::Outputs::Stdout/@workers = 1", :level=>:debug, :file=>"logstash/config/mixin.rb", :line=>"105"}
- :message=>"Registering file input", :path=>["/var/log/apache2/access.log"], :level=>:info, :file=>"logstash/inputs/file.rb", :line=>"74"}
- :message=>"No sincedb_path set, generating one based on the file path", :sincedb_path=>"/var/lib/logstash/.sincedb_8636a19711465cc96926000984eb4005", :path=>["/var/log/apache2/access.log"], :level=>:info, :file=>"logstash/inputs/file.rb", :line=>"115"}
- :message=>"_sincedb_open: reading from /var/lib/logstash/.sincedb_8636a19711465cc96926000984eb4005", :level=>:debug, :file=>"filewatch/tail.rb", :line=>"199"}
- :message=>"_discover_file_glob: /var/log/apache2/access.log: glob is: []", :level=>:debug, :file=>"filewatch/watch.rb", :line=>"117"}
- :message=>"Grok patterns path", :patterns_dir=>["/opt/logstash/patterns/*"], :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"240"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/mongodb", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/java", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/linux-syslog", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/redis", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/postgresql", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/nagios", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/mcollective-patterns", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/ruby", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- org.elasticsearch.node: [logstash-linux01-6158-4008] version[1.1.1], pid[6158], build[f1585f0/2014-04-16T14:27:12Z]
- org.elasticsearch.node: [logstash-linux01-6158-4008] initializing ...
- org.elasticsearch.node: [logstash-linux01-6158-4008] using home [/var/lib/logstash], config [/var/lib/logstash/config], data [[/var/lib/logstash/data]], logs [/var/lib/logstash/logs], work [/var/lib/logstash/work], plugins [/var/lib/logstash/plugins]
- org.elasticsearch.plugins: [logstash-linux01-6158-4008] [/var/lib/logstash/plugins] directory does not exist.
- org.elasticsearch.plugins: [logstash-linux01-6158-4008] loaded [], sites []
- org.elasticsearch.common.compress.lzf: using [UnsafeChunkDecoder] decoder
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [generic], type [cached], keep_alive [30s]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [index], type [fixed], size [1], queue_size [200]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [bulk], type [fixed], size [1], queue_size [50]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [get], type [fixed], size [1], queue_size [1k]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [search], type [fixed], size [3], queue_size [1k]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [suggest], type [fixed], size [1], queue_size [1k]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [percolate], type [fixed], size [1], queue_size [1k]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [management], type [scaling], min [1], size [5], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [flush], type [scaling], min [1], size [1], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [merge], type [scaling], min [1], size [1], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [refresh], type [scaling], min [1], size [1], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [warmer], type [scaling], min [1], size [1], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [snapshot], type [scaling], min [1], size [1], keep_alive [5m]
- org.elasticsearch.threadpool: [logstash-linux01-6158-4008] creating thread_pool [optimize], type [fixed], size [1], queue_size [null]
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] using worker_count[2], port[9300-9400], bind_host[null], publish_host[null], compress[false], connect_timeout[30s], connections_per_node[2/3/6/1/1], receive_predictor[512kb->512kb]
- org.elasticsearch.discovery.zen.ping.unicast: [logstash-linux01-6158-4008] using initial hosts [localhost:9300, localhost:9301, localhost:9302, localhost:9303, localhost:9304, localhost:9305], with concurrent_connects [10]
- org.elasticsearch.discovery.zen: [logstash-linux01-6158-4008] using ping.timeout [3s], master_election.filter_client [true], master_election.filter_data [false]
- org.elasticsearch.discovery.zen.elect: [logstash-linux01-6158-4008] using minimum_master_nodes [-1]
- org.elasticsearch.discovery.zen.fd: [logstash-linux01-6158-4008] [master] uses ping_interval [1s], ping_timeout [30s], ping_retries [3]
- org.elasticsearch.discovery.zen.fd: [logstash-linux01-6158-4008] [node ] uses ping_interval [1s], ping_timeout [30s], ping_retries [3]
- org.elasticsearch.monitor.jvm: [logstash-linux01-6158-4008] enabled [true], last_gc_enabled [false], interval [1s], gc_threshold [{old=GcThreshold{name='old', warnThreshold=10000, infoThreshold=5000, debugThreshold=2000}, default=GcThreshold{name='default', warnThreshold=10000, infoThreshold=5000, debugThreshold=2000}, young=GcThreshold{name='young', warnThreshold=1000, infoThreshold=700, debugThreshold=400}}]
- org.elasticsearch.monitor.os: [logstash-linux01-6158-4008] Using probe [org.elasticsearch.monitor.os.JmxOsProbe@6edad23a] with refresh_interval [1s]
- org.elasticsearch.monitor.process: [logstash-linux01-6158-4008] Using probe [org.elasticsearch.monitor.process.JmxProcessProbe@72e4c181] with refresh_interval [1s]
- org.elasticsearch.monitor.jvm: [logstash-linux01-6158-4008] Using refresh_interval [1s]
- org.elasticsearch.monitor.network: [logstash-linux01-6158-4008] Using probe [org.elasticsearch.monitor.network.JmxNetworkProbe@1dfc8b68] with refresh_interval [5s]
- org.elasticsearch.monitor.network: [logstash-linux01-6158-4008] net_info
- host [linux01]
- eth0 display_name [eth0]
- address [/fe80:0:0:0:20c:29ff:fed0:5f66%2] [/192.168.8.19]
- mtu [1500] multicast [true] ptp [false] loopback [false] up [true] virtual [false]
- lo display_name [lo]
- address [/0:0:0:0:0:0:0:1%1] [/127.0.0.1]
- mtu [65536] multicast [false] ptp [false] loopback [true] up [true] virtual [false]
- org.elasticsearch.monitor.fs: [logstash-linux01-6158-4008] Using probe [org.elasticsearch.monitor.fs.JmxFsProbe@6ce629e8] with refresh_interval [1s]
- org.elasticsearch.indices.store: [logstash-linux01-6158-4008] using indices.store.throttle.type [MERGE], with index.store.throttle.max_bytes_per_sec [20mb]
- org.elasticsearch.script: [logstash-linux01-6158-4008] using script cache with max_size [500], expire [null]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using node_concurrent_recoveries [2], node_initial_primaries_recoveries [4]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster.routing.allocation.allow_rebalance] with [indices_all_active]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster_concurrent_rebalance] with [2]
- org.elasticsearch.gateway.local: [logstash-linux01-6158-4008] using initial_shards [quorum], list_timeout [30s]
- org.elasticsearch.indices.recovery: [logstash-linux01-6158-4008] using max_bytes_per_sec[20mb], concurrent_streams [3], file_chunk_size [512kb], translog_size [512kb], translog_ops [1000], and compress [true]
- org.elasticsearch.indices.memory: [logstash-linux01-6158-4008] using index_buffer_size [49.1mb], with min_shard_index_buffer_size [4mb], max_shard_index_buffer_size [512mb], shard_inactive_time [30m]
- org.elasticsearch.indices.cache.filter: [logstash-linux01-6158-4008] using [node] weighted filter cache with size [20%], actual_size [98.3mb], expire [null], clean_interval [1m]
- org.elasticsearch.indices.fielddata.cache: [logstash-linux01-6158-4008] using size [-1] [-1b], expire [null]
- org.elasticsearch.gateway.local.state.meta: [logstash-linux01-6158-4008] using gateway.local.auto_import_dangled [YES], with gateway.local.dangling_timeout [2h]
- org.elasticsearch.bulk.udp: [logstash-linux01-6158-4008] using enabled [false], host [null], port [9700-9800], bulk_actions [1000], bulk_size [5mb], flush_interval [5s], concurrent_requests [4]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using node_concurrent_recoveries [2], node_initial_primaries_recoveries [4]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster.routing.allocation.allow_rebalance] with [indices_all_active]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster_concurrent_rebalance] with [2]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using node_concurrent_recoveries [2], node_initial_primaries_recoveries [4]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster.routing.allocation.allow_rebalance] with [indices_all_active]
- org.elasticsearch.cluster.routing.allocation.decider: [logstash-linux01-6158-4008] using [cluster_concurrent_rebalance] with [2]
- org.elasticsearch.node: [logstash-linux01-6158-4008] initialized
- org.elasticsearch.node: [logstash-linux01-6158-4008] starting ...
- org.elasticsearch.netty.channel.socket.nio.SelectorUtil: Using select timeout of 500
- org.elasticsearch.netty.channel.socket.nio.SelectorUtil: Epoll-bug workaround enabled = false
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] Bound to address [/0:0:0:0:0:0:0:0:9301]
- org.elasticsearch.transport: [logstash-linux01-6158-4008] bound_address {inet[/0:0:0:0:0:0:0:0:9301]}, publish_address {inet[/192.168.8.19:9301]}
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] connected to node [[#zen_unicast_1#][linux01][inet[localhost/127.0.0.1:9300]]]
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] connected to node [[#zen_unicast_2#][linux01][inet[localhost/127.0.0.1:9301]]]
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] disconnected from [[#zen_unicast_1#][linux01][inet[localhost/127.0.0.1:9300]]]
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] disconnected from [[#zen_unicast_2#][linux01][inet[localhost/127.0.0.1:9301]]]
- org.elasticsearch.discovery.zen: [logstash-linux01-6158-4008] filtered ping responses: (filter_client[true], filter_data[false])
- --> target [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]], master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]]
- org.elasticsearch.transport.netty: [logstash-linux01-6158-4008] connected to node [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]]
- org.elasticsearch.discovery.zen.publish: [logstash-linux01-6158-4008] received cluster state version 173
- org.elasticsearch.discovery.zen: [logstash-linux01-6158-4008] received cluster state from [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]] which is also master but with cluster name [Cluster [elasticsearch]]
- org.elasticsearch.discovery.zen: [logstash-linux01-6158-4008] got a new state from master node, though we are already trying to rejoin the cluster
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] processing [zen-disco-receive(from master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]])]: execute
- org.elasticsearch.discovery.zen.fd: [logstash-linux01-6158-4008] [master] restarting fault detection against master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]], reason [new cluster state received and we are monitoring the wrong master [null]]
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] got first state from fresh master [g_MjDNTAQW-uuTX3UnPf2g]
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] cluster state updated, version [173], source [zen-disco-receive(from master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]])]
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] detected_master [Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]], added {[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]],}, reason: zen-disco-receive(from master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]])
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] set local cluster state to version 173
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] processing [zen-disco-receive(from master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]])]: done applying updated cluster_state (version: 173)
- org.elasticsearch.discovery: [logstash-linux01-6158-4008] elasticsearch/drgo1vl-RmCYphYJ8hoatg
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] processing [updating local node id]: execute
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] cluster state updated, version [173], source [updating local node id]
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] set local cluster state to version 173
- org.elasticsearch.cluster.service: [logstash-linux01-6158-4008] processing [updating local node id]: done applying updated cluster_state (version: 173)
- org.elasticsearch.node: [logstash-linux01-6158-4008] started
- org.elasticsearch.discovery.zen.fd: [logstash-linux01-6158-4008] [master] starting fault detection against master [[Graydon Creed][g_MjDNTAQW-uuTX3UnPf2g][linux01][inet[/192.168.8.19:9300]]], reason [initial_join]
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/grok-patterns", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/haproxy", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/junos", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/mcollective", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Grok loading patterns from file", :path=>"/opt/logstash/patterns/firewalls", :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"247"}
- :message=>"Match data", :match=>{"message"=>"%{COMBINEDAPACHELOG}"}, :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"254"}
- :message=>"Grok compile", :field=>"message", :patterns=>["%{COMBINEDAPACHELOG}"], :level=>:info, :file=>"logstash/filters/grok.rb", :line=>"265"}
- :message=>"regexp: /message", :pattern=>"%{COMBINEDAPACHELOG}", :level=>:debug, :file=>"logstash/filters/grok.rb", :line=>"267"}
- :message=>"Adding type with date config", :type=>"", :field=>"timestamp", :format=>"dd/MMM/yyyy:HH:mm:ss Z", :level=>:debug, :file=>"logstash/filters/date.rb", :line=>"165"}
- :message=>"Pipeline started", :level=>:info, :file=>"logstash/pipeline.rb", :line=>"78"}
- :message=>"log4j java properties setup", :log4j_level=>"DEBUG", :level=>:debug, :file=>"logstash/logging.rb", :line=>"87"}
- :message=>"New Elasticsearch output", :cluster=>nil, :host=>"localhost", :port=>"9300-9305", :embedded=>false, :protocol=>"node", :level=>:info, :file=>"logstash/outputs/elasticsearch.rb", :line=>"252"}
- :message=>"Automatic template management enabled", :manage_template=>"true", :level=>:info, :file=>"logstash/outputs/elasticsearch.rb", :line=>"258"}
- :message=>"Using mapping template", :template=>"{ \"template\" : \"logstash-*\", \"settings\" : { \"index.refresh_interval\" : \"5s\" }, \"mappings\" : { \"_default_\" : { \"_all\" : {\"enabled\" : true}, \"dynamic_templates\" : [ { \"string_fields\" : { \"match\" : \"*\", \"match_mapping_type\" : \"string\", \"mapping\" : { \"type\" : \"string\", \"index\" : \"analyzed\", \"omit_norms\" : true, \"fields\" : { \"raw\" : {\"type\": \"string\", \"index\" : \"not_analyzed\", \"ignore_above\" : 256} } } } } ], \"properties\" : { \"@version\": { \"type\": \"string\", \"index\": \"not_analyzed\" }, \"geoip\" : { \"type\" : \"object\", \"dynamic\": true, \"path\": \"full\", \"properties\" : { \"location\" : { \"type\" : \"geo_point\" } } } } } }}", :level=>:info, :file=>"logstash/outputs/elasticsearch.rb", :line=>"278"}
- :message=>"_discover_file_glob: /var/log/apache2/access.log: glob is: []", :level=>:debug, :file=>"filewatch/watch.rb", :line=>"117"}
- :message=>"_discover_file_glob: /var/log/apache2/access.log: glob is: []", :level=>:debug, :file=>"filewatch/watch.rb", :line=>"117"}
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement