Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,866+0000", "level": "INFO", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "bex", "message": "master node [{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}] failed, restarting discovery" ,
- elasticsearch | "stacktrace": ["org.elasticsearch.ElasticsearchException: node [{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}] failed [3] consecutive checks",
- elasticsearch | "at org.elasticsearch.cluster.coordination.LeaderChecker$CheckScheduler$1.handleException(LeaderChecker.java:278) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleException(TransportService.java:1111) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleException(TransportService.java:1111) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.lambda$handleException$2(InboundHandler.java:246) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.handleException(InboundHandler.java:244) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.handlerResponseError(InboundHandler.java:236) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:139) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323) ~[?:?]",
- elasticsearch | "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) ~[?:?]",
- elasticsearch | "at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906) ~[?:?]",
- elasticsearch | "at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[?:?]",
- elasticsearch | "at java.lang.Thread.run(Thread.java:835) [?:?]",
- elasticsearch | "Caused by: org.elasticsearch.transport.RemoteTransportException: [el3][172.18.0.2:9300][internal:coordination/fault_detection/leader_check]",
- elasticsearch | "Caused by: org.elasticsearch.cluster.coordination.CoordinationStateRejectedException: non-leader rejecting leader check",
- elasticsearch | "at org.elasticsearch.cluster.coordination.LeaderChecker.handleLeaderCheck(LeaderChecker.java:178) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.cluster.coordination.LeaderChecker.lambda$new$0(LeaderChecker.java:105) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.xpack.security.transport.SecurityServerTransportInterceptor$ProfileSecuredRequestHandler$1.doRun(SecurityServerTransportInterceptor.java:257) ~[?:?]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.xpack.security.transport.SecurityServerTransportInterceptor$ProfileSecuredRequestHandler.messageReceived(SecurityServerTransportInterceptor.java:315) ~[?:?]",
- elasticsearch | "at org.elasticsearch.transport.RequestHandlerRegistry.processMessageReceived(RequestHandlerRegistry.java:63) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler$RequestHandler.doRun(InboundHandler.java:267) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.handleRequest(InboundHandler.java:188) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:121) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323) ~[?:?]",
- elasticsearch | "at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[?:?]",
- elasticsearch | "at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[?:?]",
- elasticsearch | "at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[?:?]",
- elasticsearch | "at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536) ~[?:?]",
- elasticsearch | "at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) ~[?:?]",
- elasticsearch | "at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906) ~[?:?]",
- elasticsearch | "at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[?:?]",
- elasticsearch | "at java.lang.Thread.run(Thread.java:835) ~[?:?]"] }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,870+0000", "level": "DEBUG", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "bex", "message": "onLeaderFailure: coordinator becoming CANDIDATE in term 299 (was FOLLOWER, lastKnownLeader was [Optional[{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}]])" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,870+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "setCurrentNodes: nodes: \n" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,870+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "already closed, doing nothing" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,870+0000", "level": "TRACE", "component": "o.e.c.c.PreVoteCollector", "cluster.name": "docker-cluster", "node.name": "bex", "message": "updating with preVoteResponse=PreVoteResponse{currentTerm=299, lastAcceptedTerm=299, lastAcceptedVersion=118}, leader=null" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,870+0000", "level": "DEBUG", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "bex", "message": "joinLeaderInTerm: for [{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}] with term 300" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,871+0000", "level": "DEBUG", "component": "o.e.c.c.CoordinationState", "cluster.name": "docker-cluster", "node.name": "bex", "message": "handleStartJoin: leaving term [299] due to StartJoinRequest{term=300,node={el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}}" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,972+0000", "level": "TRACE", "component": "o.e.c.c.PreVoteCollector", "cluster.name": "docker-cluster", "node.name": "bex", "message": "updating with preVoteResponse=PreVoteResponse{currentTerm=300, lastAcceptedTerm=299, lastAcceptedVersion=118}, leader=null" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,972+0000", "level": "DEBUG", "component": "o.e.c.c.JoinHelper", "cluster.name": "docker-cluster", "node.name": "bex", "message": "attempting to join {el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true} with JoinRequest{sourceNode={bex}{jPAH12MSRBCPusanG_hcIw}{1_6RVTvbQWKkMq7w6N502Q}{10.230.0.1}{10.230.0.1:9300}{dim}{ml.machine_memory=66956541952, xpack.installed=true, ml.max_open_jobs=20}, optionalJoin=Optional[Join{term=300, lastAcceptedTerm=299, lastAcceptedVersion=118, sourceNode={bex}{jPAH12MSRBCPusanG_hcIw}{1_6RVTvbQWKkMq7w6N502Q}{10.230.0.1}{10.230.0.1:9300}{dim}{ml.machine_memory=66956541952, xpack.installed=true, ml.max_open_jobs=20}, targetNode={el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}}]}" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,972+0000", "level": "DEBUG", "component": "o.e.c.c.ElectionSchedulerFactory", "cluster.name": "docker-cluster", "node.name": "bex", "message": "scheduling scheduleNextElection{gracePeriod=0s, thisAttempt=0, maxDelayMillis=100, delayMillis=69, ElectionScheduler{attempt=1, ElectionSchedulerFactory{initialTimeout=100ms, backoffTime=100ms, maxTimeout=10s}}}" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,985+0000", "level": "TRACE", "component": "o.e.c.c.FollowersChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "responding to FollowerCheckRequest{term=300, sender={el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}} on slow path" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,986+0000", "level": "DEBUG", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "bex", "message": "onFollowerCheckRequest: coordinator becoming FOLLOWER of [{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}] in term 300 (was CANDIDATE, lastKnownLeader was [Optional[{el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}]])" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,986+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "setCurrentNodes: nodes: \n" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,986+0000", "level": "TRACE", "component": "o.e.c.c.PreVoteCollector", "cluster.name": "docker-cluster", "node.name": "bex", "message": "updating with preVoteResponse=PreVoteResponse{currentTerm=300, lastAcceptedTerm=299, lastAcceptedVersion=118}, leader={el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,986+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "checking {el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true} with [cluster.fault_detection.leader_check.timeout] = 10s" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:55:59,990+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "scheduling next check of {el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true} for [cluster.fault_detection.leader_check.interval] = 1s" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,001+0000", "level": "DEBUG", "component": "o.e.c.c.PublicationTransportHandler", "cluster.name": "docker-cluster", "node.name": "bex", "message": "received full cluster state version [119] with size [75158]" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,003+0000", "level": "TRACE", "component": "o.e.c.c.Coordinator", "cluster.name": "docker-cluster", "node.name": "bex", "message": "handlePublishRequest: handling [PublishRequest{term=300, version=119, state=cluster uuid: UohcIpmXS6y-I_tb__DP1Q [committed: false]\nversion: 119\nstate uuid: eUVL-QamRu652iWZ9o4-LA\nfrom_diff: false\nmeta data version: 230154\n coordination_metadata:\n term: 300\n last_committed_config: VotingConfiguration{jPAH12MSRBCPusanG_hcIw,JZiS21uQRo23OpLpEmeCHw,{bootstrap-placeholder}-10.230.0.4}\n last_accepted_config: VotingConfiguration{jPAH12MSRBCPusanG_hcIw,JZiS21uQRo23OpLpEmeCHw,{bootstrap-placeholder}-10.230.0.4}\n voting tombstones: []\n [clients-20190612/REHh_nGlQ4KScS1ZaNgnyA]: v[17], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [YtsSRR1ARACDIQpBPTrvtg, YEPfQZ9MT7uAOH06MA3rPg]\n [clients-20190822/_g3N1OhKQbGoTpWOqFp4nA]: v[5], mv[2], sv[1], av[1]\n 0: p_term [1], isa_ids [1feHj1QwTkiQadQe-l8zeQ, 2tB1V22HT6yJHiShhlbcwA]\n [clients-20190604/ZF3YijXNTP62YNP-PhXhwA]: v[13], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [eGG7-tIkTBmiuG9RiVu0yA, 2uEdxNNJRkqOfNsCzmID6w]\n [exittraffic-20190724/oIil2onSSbOZsM8q8blf2Q]: v[11], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [XEYtcLoLTT-eBKbsH2K6ig, di28hWGCRRm0UomKpphaMg]\n [clients-20190223/-qb4YWPlT4iAkbLOvxxpmw]: v[26], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [Y5ckqT4DTVeruHwP1ZGwXw, QI8-g09BSKyCXaE3iQF6dg]\n [clients-20190105/t-YltESrQxGLjGYtjs6rZA]: v[33], mv[1], sv[1], av[1]\n 0: p_term [11], isa_ids [eBnfO6DUQMeOJQ-4T5bx8Q, Qp07cNkRQTe01qaqupz3zQ]\n [exittraffic-20190122/up1692_oSPGLG5NzTSs6hw]: v[37], mv[1], sv[1], av[1]\n 0: p_term [12], isa_ids [MGr5hZM5Tk6dMyLePc0PZA, jGtF6jNfRCGLn7zegdfqUg]\n [exittraffic-20190613/ZDq3W_7XQrOWpyT1VGsnng]: v[10], mv[1], sv[1], av[1]\n 0: p_term [1], isa_ids [bZe6DJYaTyq3o1VChk1Bqw, SNkpaUUKRBuM2oFNdn1n_g]\n [exittraffic-20190530/mksNwj02TG6heOgK4yrEZQ]: v[12], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [aK2kIguXT76Lf51R896DGA, U4mXLXssSZKXn3icgcYzeQ]\n [clients-20190718/SVEf7waLSPOs7MW3QJGQVQ]: v[12], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [EFV1NaZlTD6buATyrtXecw, T5sxG3kHRqmjS5cH0qpYPA]\n [clients-20190816/1LwZRVacS_CiKpR_Q_-AEw]: v[10], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [rmnMunobS0Wp0qMIM5Bg8g, O4Kb5XsYS4ynhmgkalVHAw]\n [clients-20190603/eycNrF6pRBScgPKZbuL1xA]: v[14], mv[1], sv[1], av[1]\n 0: p_term [4], isa_ids [9i8jrZhPTUS4EzYONmfxAA, ME7BQwMkRFCbqltkSvDT-w]\n [exittraffic-20190308/d5-h2r93QEOfxNIMcNTWkg]: v[23], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [eKxsDjdsQkiMZD0ITuTCSg, pZ8GfBUXSvSfL2h9Rupo2g]\n [clients-20190308/QdYkX3FDRdSutBImviuMbw]: v[27], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [lySQ6-fqTM-rgx5ae9_Vjw, 9nrPr3VES92llbI-qtCKag]\n [clients-20190413/P8RsbX_ESbuH3qLyFCdshQ]: v[25], mv[1], sv[1], av[1]\n 0: p_term [10], isa_ids [ok95Qao7RfiwWXBChXasMg, vCW_lqb8QhWXFTrJJKl2Kg]\n [exittraffic-20190214/_EQmYz7lTSuiMBjgJibxMA]: v[24], mv[1], sv[1], av[1]\n 0: p_term [9], isa_ids [I6Dy0bSZQHKF_1Yx5bC8dg, zPDk7sjYTCSQ5_KrGQe49A]\n [consensus/fGwN_b66Rk6uXDBpftZGVA]: v[102], mv[1], sv[1], av[1]\n 0: p_term [13], isa_ids [oJJUvbQLTKuOUAxsnATtBQ, ii9BBOP0REicE0fNFcCi7w]\n 1: p_term [13], isa_ids [vegnlitNQvSTs6b77_Tn-Q, lL9xKN-bQgaPqcwvmPOR0Q]\n 2: p_term [13], isa_ids [ipY33rqNTMqiXNxOH1Cdwg, qfHQgyyMQBCfec6cHnswjg]\n 3: p_term [10], isa_ids [vlRCfitCTXO7fd5v-DXUhg, cIqprpnzSjiP63gki7tC3w]\n 4: p_term [9], isa_ids [Y6jKcaj8TfWLqgML80ppfQ, RAPYmzHdQKq8-Yq1TM_VXQ]\n [clients-20190618/6MtVWumzTgCdx_KALfg7qA]: v[13], mv[1], sv[1], av[1]\n 0: p_term [4], isa_ids [2J4jrrR8QWu23OboVof2hw, tX7IfQqCQTqL8-kWOLut1A]\n [exittraffic-20190503/o4aKg40EQX6QOqV_43XnWw]: v[19], mv[1], sv[1], av[1]\n 0: p_term [5], isa_ids [X6RapzK2Qg-BEqHphZvzIw, -3FzzxzaSjyigR5Qb81HTQ]\n [search/-ePpz-ZqQGeFkzpAUXnB-g]: v[82], mv[1], sv[1], av[1]\n 0: p_term [9], isa_ids [6lD1CPJwRH-VjaCby7u3Cg, 1LR4rzEnSrauH7A5AFUWEQ]\n 1: p_term [8], isa_ids [IFFwLe6_SCuH5O2mzEc1PQ, i404AaJOTammxYVn_rBuVw]\n 2: p_term [10], isa_ids [ZeAEVPa7RbqSf-oBdPytPQ, rUMgbhaXTAOSxyVwOR7AVg]\n 3: p_term [9], isa_ids [zC8SN4lKQ52u3LDq8texNw, 9cnWiDfBQeGtYz049jT9VA]\n 4: p_term [7], isa_ids [i5kiJCKFSR6qPlGAmD4Utw, AtREmfFKSL6VpAg3kns5Ag]\n [exittraffic-20190422/yt58Kn3HS9y6KvzRrNzQBA]: v[26], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [k52mbo6qS5ak1iVMHV_g7w, IEyTey0cS3Clo6J_8uLp8g]\n [clients-20190401/JcQV6C8nS8qIVm20O9kPtg]: v[23], mv[1], sv[1], av[1]\n 0: p_term [7], isa_ids [tLEYfNz6QcGDVS7QtX-mVw, 1-RDHu87QbKmJRiKChKdCw]\n [clients-20190802/Jkr1pyDoSsGua12h2R83Vw]: v[10], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [rMuVAI07SRS1qRx8TyRPCQ, gg5D6PJvRguEgz0NIJIChw]\n [exittraffic-20190525/9J5Z3TycSAqBngCQjFgx-Q]: v[10], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [DuHNb6L6TAeg32Oya_ziMQ, 6E5rpoL2S62r3SNnDm8WEg]\n [clients-20190123/Y5B_oWL7RZewhBn3ax2Okg]: v[31], mv[1], sv[1], av[1]\n 0: p_term [13], isa_ids [22ieBRrLQcqyti9r50rGFg, rTqf03D9T0Oai8CxTY5quQ]\n [exittraffic-20190811/qD6q25bbRhakTx2b0KOnbg]: v[10], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [ITSE_Ly3Q1-hKzWckoTeXQ, PZcdJ_hKSdy-Q0fuFmuKKw]\n [exittraffic-20190709/9U8GxIM3SqyvwblHLyK39A]: v[13], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [j5Oz-ZsER0KQaZZJt_bZng, qMGW0LXKTd-KQlXSPUZJ_Q]\n [exittraffic-20190520/JfxNMaAKT_mlBIZpD1npTg]: v[16], mv[1], sv[1], av[1]\n 0: p_term [5], isa_ids [UMv7pPAsT2OTE99oungn-Q, tyZSB6NMSy2oIjrgLnjZGg]\n [exittraffic-20190622/kZyPl7oJTme1rkMLzYDLNw]: v[14], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [JITYJW0GRHaTvwljhGT0ew, ltdIW7UbQw2nGN1pvnuVVA]\n [exittraffic-20190124/_axEZ3FrSquIe9-weIgQWg]: v[30], mv[1], sv[1], av[1]\n 0: p_term [10], isa_ids [Tz7qCofSQhaVEVgrHZxyiw, KA4llvKBT46PAmoTwuGz8A]\n [exittraffic-20190207/ucPYrNR7T8aVNpI-875mhQ]: v[22], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [6pJ1UCg4SqKV_JySlr_2lw, W-t16dQTSfem--zQWdK8Iw]\n [clients-20190307/WiBrP7FgS2aa3Kx867vBtg]: v[22], mv[1], sv[1], av[1]\n 0: p_term [7], isa_ids [lF8fjZjvQDCZr-MPa65Z-w, SGw4BUFSTj-O1h61HS5baw]\n [clients-20190506/uJ-MBd2KQymioS5sIMFAIQ]: v[19], mv[1], sv[1], av[1]\n 0: p_term [5], isa_ids [vCs0w0yvQOqYSjhHAOXrfw, J7xlS7OFQBW_1_CfFVzQbw]\n [exittraffic-20190506/g-0Q-UpYReeUV0ED-8o4HQ]: v[17], mv[1], sv[1], av[1]\n 0: p_term [5], isa_ids [_PGMKTynQbikceIlE3P_xw, EOPusQlqTvieNIllRqkS9g]\n [clients-20190117/Ijy0W1HTTlO-UEe4oDZK4Q]: v[33], mv[1], sv[1], av[1]\n 0: p_term [11], isa_ids [kwyzRbfZTvK0D-3rW49WuA, jhs15VMLThK0IBaQQ1FX9Q]\n [exittraffic-20190317/XoKk-WR1RCqe5MP-xXm2uA]: v[26], mv[1], sv[1], av[1]\n 0: p_term [9], isa_ids [urtx4MNJSD2FHOMGFKneYA, HSa9fwXrQ0m963Jqn9aHSw]\n [exittraffic-20190518/_IW-4B3LQJuJ-zecpNCbNQ]: v[17], mv[1], sv[1], av[1]\n 0: p_term [4], isa_ids [hKOlH5nySNebY6hqxJ7PRg, q3odPH71RiGV3hnam61lLw]\n [clients-20190213/zUOxgGkOTU6yu3qfYJbexg]: v[27], mv[1], sv[1], av[1]\n 0: p_term [10], isa_ids [7Qi1saK_RwuQw80-StKcQg, 6b511fzbQ3qMap3EMM6oaw]\n [exittraffic-20190513/qX7c0zv0TCiwr8vxIMgfyA]: v[21], mv[1], sv[1], av[1]\n 0: p_term [5], isa_ids [R3bjrVuDQ5yTnjFhPa75Tg, fQrWufOuS1u8SXe1PXJD1w]\n [clients-20190127/slbgCJPIRGaga-Bj2QBQEg]: v[32], mv[1], sv[1], av[1]\n 0: p_term [11], isa_ids [HaZGgBORT4WVSujpJ9nC5g, 2VLkln9nTK6xoO38bgR7Eg]\n [clients-20190807/lu3aJ7EdQIO3HPmz_lftGg]: v[10], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [yZS5MhziQceSKF2NbmDKHQ, Ezm3QEctTeGCzSe-hssmgg]\n [clients-20190815/Max-rDC0QNuR3B1fln3Zww]: v[11], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [29TTTA2kRHWseXx0QeukkA, e0TThgTjQ3yGiORnJ-SEaw]\n [clients-20190814/9WQUGEC5SBSyTxY_WXDTOw]: v[11], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [9Re1c9m5Rz-s1JyGOzorBw, gWcOSgKWTROO05Bnk6Y4oQ]\n [clients-20190403/5Ubx-y8JTMqWO8emSky4pg]: v[27], mv[1], sv[1], av[1]\n 0: p_term [10], isa_ids [aPB5t9ouRcen7V3JLrdC_w, ht1ib8b-SjyWJ2F7L2D-OQ]\n [exittraffic-20190728/VJ9_U-GYQO6H1K5VAC233A]: v[8], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [twv8vUwbTpS7PnTpbOvz6w, r_oy1C1sS3asupNZrFulUQ]\n [clients-20190305/4P1Q1b7xRHuKXRzjAp--OA]: v[22], mv[1], sv[1], av[1]\n 0: p_term [7], isa_ids [yRlROZ1wROqDgvv0yNDjJA, B6y4hOBvQUKFQVVPX_5wtw]\n [exittraffic-20190227/GLJIPN2jTzOinVG4IyWuXA]: v[29], mv[1], sv[1], av[1]\n 0: p_term [8], isa_ids [h7KXoVcATHKCPGuDUhsiTA, T1nrePd6RdyCvuhnvexjaQ]\n [clients-20190121/O1g_L7YaShWE_7DBiPYxKA]: v[35], mv[1], sv[1], av[1]\n 0: p_term [14], isa_ids [TQgTgDNuTIenGSmV3AT_VQ, ki_8E7-QTf6VrUcdrNY2XQ]\n [clients-20190620/7Hug8XENQKuWM28Zhq6IDA]: v[13], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [lP_yfj1vRpCtFgKEAtBcRQ, 9XMtAnkeTP2v3rGMhWGUgA]\n [clients-20190529/sq-kIzqGR_y7Vu-rfq10Bw]: v[13], mv[1], sv[1], av[1]\n 0: p_term [3], isa_ids [8hIR-xjCShuvpz3fca3DWw, hRkI93nTSjODTsBw0YRKdg]\n [clients-20190328/di8OuGmiT9WaQ22dhN-CNA]: v[21], mv[1], sv[1], av[1]\n 0: p_term [7], isa_ids [6c2H-CliQoK94_j-APjkGQ, LTWRJgOhRbG8cc46uhJtog]\n [clients-20190519/B1lFlH8fQcaTcXdbabNOUA]: v[21], mv[1], sv[1], av[1]\n 0: p_term [6], isa_ids [DzrB_5bUTLmlaeGvZsTaMA, IisLQHMcROKI4B-AVg6-5A]\n [exittraffic-20190819/tr230HHRQlKVc78SDF3U9A]: v[5], mv[2], sv[1], av[1]\n 0: p_term [1], isa_ids [zxmEP085RZ6cTOecHN0-iw, kAls7b19TIKg3-m9xYTi9A]\n [clients-20190517/IaWdazVOSO69HKVXvS1VZw]: v[21], mv[1], sv[1], av[1]\n 0: p_term [6], isa_ids [iFcMUhq7R3e14QxbsYUW5Q, jU2oA65JS16RFAxJvyaCQw]\n [nutchpages2/gokxilFZQcei2ZN6ESa2SQ]: v[13], mv[5], sv[1], av[1]\n 0: p_term [1], isa_ids [3YSaabiTS2-Tk2Lww9IKqg, MmeR-smkRji5-TjutTcS3w]\n 1: p_term [1], isa_ids [3B2LK1njSLGtFJ0t2FgGbw, BVg21WhbSOG93BI1tqGPpw]\n 2: p_term [1], isa_ids [f4SqICyeTCu5MIEfC0t3Xw, ZrZphvWoTSeIgBhcuEi8gg]\n 3: p_term [1], isa_ids [-th3cL3eQaGhdQK8Bf4HCQ, FfrX18sSTK2ck2eZmrJQFA]\n 4: p_term [1], isa_ids [A2wIpD6rRCGRAtwxxUG_hQ, l8sZLsutR6SwrxfyCXIIEQ]\n [clients-20190731/MW_mgHQ3TySLQ6vQozDShw]: v[8], mv[1], sv[1], av[1]\n 0: p_term [2], isa_ids [3uyNhhkaRq-VcO5dGlGp2g, " }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,003+0000", "level": "TRACE", "component": "o.e.c.c.CoordinationState", "cluster.name": "docker-cluster", "node.name": "bex", "message": "handlePublishRequest: accepting publish request for version [119] and term [300]" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,041+0000", "level": "DEBUG", "component": "o.e.c.c.ElectionSchedulerFactory", "cluster.name": "docker-cluster", "node.name": "bex", "message": "scheduleNextElection{gracePeriod=0s, thisAttempt=0, maxDelayMillis=100, delayMillis=69, ElectionScheduler{attempt=1, ElectionSchedulerFactory{initialTimeout=100ms, backoffTime=100ms, maxTimeout=10s}}} not starting election" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,534+0000", "level": "WARN", "component": "r.suppressed", "cluster.name": "docker-cluster", "node.name": "bex", "message": "path: /_bulk, params: {}" ,
- elasticsearch | "stacktrace": ["org.elasticsearch.cluster.block.ClusterBlockException: blocked by: [SERVICE_UNAVAILABLE/1/state not recovered / initialized, SERVICE_UNAVAILABLE/2/no master];",
- elasticsearch | "at org.elasticsearch.cluster.block.ClusterBlocks.globalBlockedException(ClusterBlocks.java:189) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.action.bulk.TransportBulkAction$BulkOperation.handleBlockExceptions(TransportBulkAction.java:481) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.action.bulk.TransportBulkAction$BulkOperation.doRun(TransportBulkAction.java:362) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.action.bulk.TransportBulkAction$BulkOperation$2.onTimeout(TransportBulkAction.java:515) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.cluster.ClusterStateObserver$ContextPreservingListener.onTimeout(ClusterStateObserver.java:325) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.cluster.ClusterStateObserver$ObserverClusterStateListener.onTimeout(ClusterStateObserver.java:252) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.cluster.service.ClusterApplierService$NotifyTimeout.run(ClusterApplierService.java:572) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]",
- elasticsearch | "at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]",
- elasticsearch | "at java.lang.Thread.run(Thread.java:835) [?:?]",
- elasticsearch | "Suppressed: org.elasticsearch.discovery.MasterNotDiscoveredException",
- elasticsearch | "\tat org.elasticsearch.action.support.master.TransportMasterNodeAction$AsyncSingleAction$3.onTimeout(TransportMasterNodeAction.java:251) ~[elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "\tat org.elasticsearch.cluster.ClusterStateObserver$ContextPreservingListener.onTimeout(ClusterStateObserver.java:325) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "\tat org.elasticsearch.cluster.ClusterStateObserver$ObserverClusterStateListener.onTimeout(ClusterStateObserver.java:252) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "\tat org.elasticsearch.cluster.service.ClusterApplierService$NotifyTimeout.run(ClusterApplierService.java:572) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "\tat org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:688) [elasticsearch-7.3.1.jar:7.3.1]",
- elasticsearch | "\tat java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]",
- elasticsearch | "\tat java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]",
- elasticsearch | "\tat java.lang.Thread.run(Thread.java:835) [?:?]"] }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,723+0000", "level": "DEBUG", "component": "o.e.a.a.i.c.TransportCreateIndexAction", "cluster.name": "docker-cluster", "node.name": "bex", "message": "timed out while retrying [indices:admin/create] after failure (timeout [1m])" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,990+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "checking {el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true} with [cluster.fault_detection.leader_check.timeout] = 10s" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,990+0000", "level": "TRACE", "component": "o.e.c.c.FollowersChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "responding to FollowerCheckRequest{term=300, sender={el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true}} on fast path" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,994+0000", "level": "TRACE", "component": "o.e.c.c.LeaderChecker", "cluster.name": "docker-cluster", "node.name": "bex", "message": "scheduling next check of {el3}{JZiS21uQRo23OpLpEmeCHw}{Zgt_NayVRBO6EN5MuRcoig}{10.230.0.3}{10.230.0.3:9300}{dim}{ml.machine_memory=33459941376, ml.max_open_jobs=20, xpack.installed=true} for [cluster.fault_detection.leader_check.interval] = 1s" }
- elasticsearch | {"type": "server", "timestamp": "2019-08-23T20:56:00,997+0000", "level": "DEBUG", "component": "o.e.a.a.i.c.TransportCreateIndexAction", "cluster.name": "docker-cluster", "node.name": "bex", "message": "no known master node, scheduling a retry" }
Advertisement
Add Comment
Please, Sign In to add comment