Advertisement
Guest User

Storm Logger

a guest
May 6th, 2016
174
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 79.75 KB | None | 0 0
  1. 13:02:02.098 [main] INFO o.a.s.d.s.Supervisor - Starting Supervisor with conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.cgroup.resources=[cpu, memory], storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=1, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.messaging.netty.min_wait_ms=100, storm.resource.isolation.plugin=org.apache.storm.container.cgroup.CgroupManager, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, drpc.queue.size=128, topology.backpressure.enable=true, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=[org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter], storm.cgroup.hierarchy.dir=/cgroup/storm_resources, topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=null, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=[localhost], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, storm.cgroup.cgexec.cmd=/bin/cgexec, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/b39c0ddf-0fbd-4f3c-a3f5-dd1d95419e68, topology.disable.loadaware=false, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=4096.0, storm.supervisor.cgroup.rootdir=storm, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, storm.cgroup.hierarchy.name=storm, task.credentials.poll.secs=30, blobstore.superuser=daniel, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[1024, 1025, 1026], topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.cgroup.memory.limit.tolerance.margin.mb=128.0, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster.ZKStateStorageFactory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, storm.resource.isolation.plugin.enable=false, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, nimbus.impersonation.authorizer=org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=[localhost], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, storm.supervisor.worker.manager.plugin=org.apache.storm.daemon.supervisor.workermanager.DefaultWorkerManager, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}
  2. 13:02:02.141 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  3. 13:02:02.142 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@708e76
  4. 13:02:02.142 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  5. 13:02:02.142 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  6. 13:02:02.142 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60773
  7. 13:02:02.143 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60773
  8. 13:02:02.165 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af0009, negotiated timeout = 20000
  9. 13:02:02.165 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af0009 with negotiated timeout 20000 for client /127.0.0.1:60773
  10. 13:02:02.166 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  11. 13:02:02.166 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  12. 13:02:02.167 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af0009
  13. 13:02:02.173 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af0009 closed
  14. 13:02:02.173 [main-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  15. 13:02:02.173 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60773 which had sessionid 0x15486cde5af0009
  16. 13:02:02.174 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  17. 13:02:02.174 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@1f089a59
  18. 13:02:02.175 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  19. 13:02:02.176 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  20. 13:02:02.176 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60774
  21. 13:02:02.176 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60774
  22. 13:02:02.182 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000a with negotiated timeout 20000 for client /127.0.0.1:60774
  23. 13:02:02.182 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000a, negotiated timeout = 20000
  24. 13:02:02.182 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  25. 13:02:02.188 [main] INFO o.a.s.l.Localizer - Reconstruct localized resource: /tmp/b39c0ddf-0fbd-4f3c-a3f5-dd1d95419e68/supervisor/usercache
  26. 13:02:02.188 [main] WARN o.a.s.l.Localizer - No left over resources found for any user during reconstructing of local resources at: /tmp/b39c0ddf-0fbd-4f3c-a3f5-dd1d95419e68/supervisor/usercache
  27. 13:02:02.239 [main] INFO o.a.s.d.s.Supervisor - Starting supervisor with id 85441820-1179-41fe-9b70-5d4e6de3e821 at host hp13-HP-EliteDesk-800-G1-SFF.
  28. 13:02:02.242 [main] INFO o.a.s.d.local-supervisor - Starting local Supervisor with conf {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", "storm.messaging.netty.client_worker_threads" 1, "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.cgroup.resources" ["cpu" "memory"], "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "storm.resource.isolation.plugin" "org.apache.storm.container.cgroup.CgroupManager", "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "drpc.queue.size" 128, "topology.backpressure.enable" true, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "storm.cgroup.hierarchy.dir" "/cgroup/storm_resources", "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" nil, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" nil, "storm.cgroup.cgexec.cmd" "/bin/cgexec", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/tmp/727574c3-2346-4211-9f65-fd19fb432995", "topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" nil, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 4096.0, "storm.supervisor.cgroup.rootdir" "storm", "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "storm.cgroup.hierarchy.name" "storm", "task.credentials.poll.secs" 30, "blobstore.superuser" "daniel", "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" (1027 1028 1029), "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.cgroup.memory.limit.tolerance.margin.mb" 128.0, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster.ZKStateStorageFactory", "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "storm.resource.isolation.plugin.enable" false, "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "nimbus.impersonation.authorizer" "org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "supervisor.childopts" "-Xmx256m", "topology.worker.max.heap.size.mb" 768.0, "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "storm.supervisor.worker.manager.plugin" "org.apache.storm.daemon.supervisor.workermanager.DefaultWorkerManager", "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib", "ui.port" 8080, "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600}
  29. 13:02:02.244 [main] INFO o.a.s.d.s.Supervisor - Starting Supervisor with conf {topology.builtin.metrics.bucket.size.secs=60, nimbus.childopts=-Xmx1024m, ui.filter.params=null, storm.cluster.mode=local, storm.messaging.netty.client_worker_threads=1, logviewer.max.per.worker.logs.size.mb=2048, supervisor.run.worker.as.user=false, topology.max.task.parallelism=null, topology.priority=29, zmq.threads=1, storm.group.mapping.service=org.apache.storm.security.auth.ShellBasedGroupsMapping, transactional.zookeeper.root=/transactional, topology.sleep.spout.wait.strategy.time.ms=1, scheduler.display.resource=false, topology.max.replication.wait.time.sec=60, drpc.invocations.port=3773, supervisor.localizer.cache.target.size.mb=10240, topology.multilang.serializer=org.apache.storm.multilang.JsonSerializer, storm.messaging.netty.server_worker_threads=1, nimbus.blobstore.class=org.apache.storm.blobstore.LocalFsBlobStore, resource.aware.scheduler.eviction.strategy=org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy, topology.max.error.report.per.interval=5, storm.thrift.transport=org.apache.storm.security.auth.SimpleTransportPlugin, zmq.hwm=0, storm.cgroup.resources=[cpu, memory], storm.group.mapping.service.params=null, worker.profiler.enabled=false, storm.principal.tolocal=org.apache.storm.security.auth.DefaultPrincipalToLocal, supervisor.worker.shutdown.sleep.secs=1, pacemaker.host=localhost, storm.zookeeper.retry.times=5, ui.actions.enabled=true, zmq.linger.millis=0, supervisor.enable=true, topology.stats.sample.rate=0.05, storm.messaging.netty.min_wait_ms=100, storm.resource.isolation.plugin=org.apache.storm.container.cgroup.CgroupManager, worker.log.level.reset.poll.secs=30, storm.zookeeper.port=2000, supervisor.heartbeat.frequency.secs=5, topology.enable.message.timeouts=true, supervisor.cpu.capacity=400.0, drpc.worker.threads=64, supervisor.blobstore.download.thread.count=5, drpc.queue.size=128, topology.backpressure.enable=true, supervisor.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, storm.blobstore.inputstream.buffer.size.bytes=65536, topology.shellbolt.max.pending=100, drpc.https.keystore.password=, nimbus.code.sync.freq.secs=120, logviewer.port=8000, topology.scheduler.strategy=org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy, topology.executor.send.buffer.size=1024, resource.aware.scheduler.priority.strategy=org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy, pacemaker.auth.method=NONE, storm.daemon.metrics.reporter.plugins=[org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter], storm.cgroup.hierarchy.dir=/cgroup/storm_resources, topology.worker.logwriter.childopts=-Xmx64m, topology.spout.wait.strategy=org.apache.storm.spout.SleepSpoutWaitStrategy, ui.host=0.0.0.0, storm.nimbus.retry.interval.millis=2000, nimbus.inbox.jar.expiration.secs=3600, dev.zookeeper.path=/tmp/dev-storm-zookeeper, topology.acker.executors=null, topology.fall.back.on.java.serialization=true, topology.eventlogger.executors=null, supervisor.localizer.cleanup.interval.ms=600000, storm.zookeeper.servers=[localhost], nimbus.thrift.threads=64, logviewer.cleanup.age.mins=10080, topology.worker.childopts=null, topology.classpath=null, storm.cgroup.cgexec.cmd=/bin/cgexec, supervisor.monitor.frequency.secs=3, nimbus.credential.renewers.freq.secs=600, topology.skip.missing.kryo.registrations=true, drpc.authorizer.acl.filename=drpc-auth-acl.yaml, pacemaker.kerberos.users=[], storm.group.mapping.service.cache.duration.secs=120, topology.testing.always.try.serialize=false, nimbus.monitor.freq.secs=10, storm.health.check.timeout.ms=5000, supervisor.supervisors=[], topology.tasks=null, topology.bolts.outgoing.overflow.buffer.enable=false, storm.messaging.netty.socket.backlog=500, topology.workers=1, pacemaker.base.threads=10, storm.local.dir=/tmp/727574c3-2346-4211-9f65-fd19fb432995, topology.disable.loadaware=false, worker.childopts=-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump, storm.auth.simple-white-list.users=[], topology.disruptor.batch.timeout.millis=1, topology.message.timeout.secs=30, topology.state.synchronization.timeout.secs=60, topology.tuple.serializer=org.apache.storm.serialization.types.ListDelegateSerializer, supervisor.supervisors.commands=[], nimbus.blobstore.expiration.secs=600, logviewer.childopts=-Xmx128m, topology.environment=null, topology.debug=false, topology.disruptor.batch.size=100, storm.messaging.netty.max_retries=300, ui.childopts=-Xmx768m, storm.network.topography.plugin=org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping, storm.zookeeper.session.timeout=20000, drpc.childopts=-Xmx768m, drpc.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.connection.timeout=15000, storm.zookeeper.auth.user=null, storm.meta.serialization.delegate=org.apache.storm.serialization.GzipThriftSerializationDelegate, topology.max.spout.pending=null, storm.codedistributor.class=org.apache.storm.codedistributor.LocalFileSystemCodeDistributor, nimbus.supervisor.timeout.secs=60, nimbus.task.timeout.secs=30, drpc.port=3772, pacemaker.max.threads=50, storm.zookeeper.retry.intervalceiling.millis=30000, nimbus.thrift.port=6627, storm.auth.simple-acl.admins=[], topology.component.cpu.pcore.percent=10.0, supervisor.memory.capacity.mb=4096.0, storm.supervisor.cgroup.rootdir=storm, storm.nimbus.retry.times=5, supervisor.worker.start.timeout.secs=120, storm.zookeeper.retry.interval=1000, logs.users=null, worker.profiler.command=flight.bash, transactional.zookeeper.port=null, drpc.max_buffer_size=1048576, pacemaker.thread.timeout=10, storm.cgroup.hierarchy.name=storm, task.credentials.poll.secs=30, blobstore.superuser=daniel, drpc.https.keystore.type=JKS, topology.worker.receiver.thread.count=1, topology.state.checkpoint.interval.ms=1000, supervisor.slots.ports=[1027, 1028, 1029], topology.transfer.buffer.size=1024, storm.health.check.dir=healthchecks, topology.worker.shared.thread.pool.size=4, drpc.authorizer.acl.strict=false, nimbus.file.copy.expiration.secs=600, worker.profiler.childopts=-XX:+UnlockCommercialFeatures -XX:+FlightRecorder, topology.executor.receive.buffer.size=1024, backpressure.disruptor.low.watermark=0.4, nimbus.task.launch.secs=120, storm.local.mode.zmq=false, storm.cgroup.memory.limit.tolerance.margin.mb=128.0, storm.messaging.netty.buffer_size=5242880, storm.cluster.state.store=org.apache.storm.cluster.ZKStateStorageFactory, worker.heartbeat.frequency.secs=1, storm.log4j2.conf.dir=log4j2, storm.resource.isolation.plugin.enable=false, ui.http.creds.plugin=org.apache.storm.security.auth.DefaultHttpCredentialsPlugin, storm.zookeeper.root=/storm, topology.tick.tuple.freq.secs=null, drpc.https.port=-1, storm.workers.artifacts.dir=workers-artifacts, supervisor.blobstore.download.max_retries=3, task.refresh.poll.secs=10, storm.exhibitor.port=8080, task.heartbeat.frequency.secs=3, pacemaker.port=6699, storm.messaging.netty.max_wait_ms=1000, nimbus.impersonation.authorizer=org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer, topology.component.resources.offheap.memory.mb=0.0, drpc.http.port=3774, topology.error.throttle.interval.secs=10, storm.messaging.transport=org.apache.storm.messaging.netty.Context, storm.messaging.netty.authentication=false, topology.component.resources.onheap.memory.mb=128.0, topology.kryo.factory=org.apache.storm.serialization.DefaultKryoFactory, worker.gc.childopts=, nimbus.topology.validator=org.apache.storm.nimbus.DefaultTopologyValidator, nimbus.seeds=[localhost], nimbus.queue.size=100000, nimbus.cleanup.inbox.freq.secs=600, storm.blobstore.replication.factor=3, worker.heap.memory.mb=768, logviewer.max.sum.worker.logs.size.mb=4096, pacemaker.childopts=-Xmx1024m, ui.users=null, transactional.zookeeper.servers=null, supervisor.worker.timeout.secs=30, storm.zookeeper.auth.password=null, client.blobstore.class=org.apache.storm.blobstore.NimbusBlobStore, supervisor.childopts=-Xmx256m, topology.worker.max.heap.size.mb=768.0, backpressure.disruptor.high.watermark=0.9, ui.filter=null, ui.header.buffer.bytes=4096, topology.min.replication.count=1, storm.supervisor.worker.manager.plugin=org.apache.storm.daemon.supervisor.workermanager.DefaultWorkerManager, topology.disruptor.wait.timeout.millis=1000, storm.nimbus.retry.intervalceiling.millis=60000, topology.trident.batch.emit.interval.millis=50, storm.auth.simple-acl.users=[], drpc.invocations.threads=64, java.library.path=/usr/local/lib:/opt/local/lib:/usr/lib, ui.port=8080, storm.exhibitor.poll.uripath=/exhibitor/v1/cluster/list, storm.messaging.netty.transfer.batch.size=262144, logviewer.appender.name=A1, nimbus.thrift.max_buffer_size=1048576, storm.auth.simple-acl.users.commands=[], drpc.request.timeout.secs=600}
  30. 13:02:02.245 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  31. 13:02:02.245 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@1d28050e
  32. 13:02:02.246 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  33. 13:02:02.246 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  34. 13:02:02.246 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60775
  35. 13:02:02.247 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60775
  36. 13:02:02.299 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000b with negotiated timeout 20000 for client /127.0.0.1:60775
  37. 13:02:02.299 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000b, negotiated timeout = 20000
  38. 13:02:02.300 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  39. 13:02:02.302 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  40. 13:02:02.303 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af000b
  41. 13:02:02.307 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af000b closed
  42. 13:02:02.307 [main-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  43. 13:02:02.307 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60775 which had sessionid 0x15486cde5af000b
  44. 13:02:02.307 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  45. 13:02:02.308 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@54917c5e
  46. 13:02:02.308 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  47. 13:02:02.309 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  48. 13:02:02.309 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60776
  49. 13:02:02.309 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60776
  50. 13:02:02.315 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000c with negotiated timeout 20000 for client /127.0.0.1:60776
  51. 13:02:02.315 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000c, negotiated timeout = 20000
  52. 13:02:02.316 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  53. 13:02:02.320 [main] INFO o.a.s.l.Localizer - Reconstruct localized resource: /tmp/727574c3-2346-4211-9f65-fd19fb432995/supervisor/usercache
  54. 13:02:02.320 [main] WARN o.a.s.l.Localizer - No left over resources found for any user during reconstructing of local resources at: /tmp/727574c3-2346-4211-9f65-fd19fb432995/supervisor/usercache
  55. 13:02:02.332 [main] INFO o.a.s.d.s.Supervisor - Starting supervisor with id a8b00f70-af32-4781-a9e8-68d1a5864ccb at host hp13-HP-EliteDesk-800-G1-SFF.
  56. 13:02:02.354 [main] INFO o.a.s.l.ThriftAccessLogger - Request ID: 1 access from: null principal: null operation: submitTopology
  57. 13:02:02.382 [main] INFO o.a.s.d.nimbus - Received topology submission for countApp with conf {"topology.max.task.parallelism" nil, "topology.submitter.principal" "", "topology.acker.executors" nil, "topology.eventlogger.executors" nil, "topology.workers" 1, "topology.message.timeout.secs" 120, "storm.zookeeper.superACL" nil, "topology.users" (), "topology.submitter.user" "daniel", "topology.kryo.register" nil, "topology.kryo.decorators" (), "storm.id" "countApp-1-1462550522", "topology.name" "countApp"}
  58. 13:02:02.399 [main] INFO o.a.s.d.nimbus - uploadedJar
  59. 13:02:02.423 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  60. 13:02:02.424 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@64ea622a
  61. 13:02:02.424 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  62. 13:02:02.425 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  63. 13:02:02.425 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60777
  64. 13:02:02.425 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60777
  65. 13:02:02.457 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000d with negotiated timeout 20000 for client /127.0.0.1:60777
  66. 13:02:02.457 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000d, negotiated timeout = 20000
  67. 13:02:02.457 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  68. 13:02:02.459 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Got user-level KeeperException when processing sessionid:0x15486cde5af000d type:create cxid:0x2 zxid:0x26 txntype:-1 reqpath:n/a Error Path:/storm/blobstoremaxkeysequencenumber Error:KeeperErrorCode = NoNode for /storm/blobstoremaxkeysequencenumber
  69. 13:02:02.490 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  70. 13:02:02.491 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af000d
  71. 13:02:02.498 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af000d closed
  72. 13:02:02.498 [main-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  73. 13:02:02.499 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60777 which had sessionid 0x15486cde5af000d
  74. 13:02:02.499 [main] INFO o.a.s.c.StormClusterStateImpl - set-path: /blobstore/countApp-1-1462550522-stormconf.ser/hp13-HP-EliteDesk-800-G1-SFF:6627-1
  75. 13:02:02.574 [main] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  76. 13:02:02.575 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@77cdc895
  77. 13:02:02.576 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  78. 13:02:02.576 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  79. 13:02:02.576 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60778
  80. 13:02:02.576 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60778
  81. 13:02:02.602 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000e with negotiated timeout 20000 for client /127.0.0.1:60778
  82. 13:02:02.602 [main-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000e, negotiated timeout = 20000
  83. 13:02:02.603 [main-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  84. 13:02:02.624 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  85. 13:02:02.626 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af000e
  86. 13:02:02.632 [main-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  87. 13:02:02.632 [main] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af000e closed
  88. 13:02:02.633 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60778 which had sessionid 0x15486cde5af000e
  89. 13:02:02.634 [main] INFO o.a.s.c.StormClusterStateImpl - set-path: /blobstore/countApp-1-1462550522-stormcode.ser/hp13-HP-EliteDesk-800-G1-SFF:6627-1
  90. 13:02:02.665 [main] INFO o.a.s.d.nimbus - desired replication count of 1 not achieved but we have hit the max wait time 60 so moving on with replication count for conf key = 1 for code key = 1for jar key = 1
  91. 13:02:02.717 [main] INFO o.a.s.d.nimbus - Activating countApp: countApp-1-1462550522
  92. 13:02:02.900 [timer] INFO o.a.s.s.EvenScheduler - Available slots: [a8b00f70-af32-4781-a9e8-68d1a5864ccb:1027, a8b00f70-af32-4781-a9e8-68d1a5864ccb:1028, a8b00f70-af32-4781-a9e8-68d1a5864ccb:1029, 85441820-1179-41fe-9b70-5d4e6de3e821:1024, 85441820-1179-41fe-9b70-5d4e6de3e821:1025, 85441820-1179-41fe-9b70-5d4e6de3e821:1026]
  93. 13:02:02.923 [timer] INFO o.a.s.d.nimbus - Setting new assignment for topology id countApp-1-1462550522: #org.apache.storm.daemon.common.Assignment{:master-code-dir "/tmp/f0efa436-870f-4f84-88a7-adbb8d89828d", :node->host {"85441820-1179-41fe-9b70-5d4e6de3e821" "hp13-HP-EliteDesk-800-G1-SFF"}, :executor->node+port {[3 3] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024], [6 6] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024], [5 5] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024], [4 4] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024], [2 2] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024], [1 1] ["85441820-1179-41fe-9b70-5d4e6de3e821" 1024]}, :executor->start-time-secs {[1 1] 1462550522, [2 2] 1462550522, [3 3] 1462550522, [4 4] 1462550522, [5 5] 1462550522, [6 6] 1462550522}, :worker->resources {["85441820-1179-41fe-9b70-5d4e6de3e821" 1024] [0.0 0.0 0.0]}}
  94. 13:02:03.230 [Thread-12] INFO o.a.s.d.s.SyncSupervisorEvent - Downloading code for storm id countApp-1-1462550522.
  95. 13:02:03.232 [Thread-12] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  96. 13:02:03.232 [Thread-12] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5747596e
  97. 13:02:03.233 [Thread-12] INFO o.a.s.b.FileBlobStoreImpl - Creating new blob store based in /tmp/f0efa436-870f-4f84-88a7-adbb8d89828d/blobs
  98. 13:02:03.233 [Thread-12-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  99. 13:02:03.233 [Thread-12-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  100. 13:02:03.233 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60779
  101. 13:02:03.233 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60779
  102. 13:02:03.257 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af000f with negotiated timeout 20000 for client /127.0.0.1:60779
  103. 13:02:03.257 [Thread-12-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af000f, negotiated timeout = 20000
  104. 13:02:03.257 [Thread-12-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  105. 13:02:03.267 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  106. 13:02:03.268 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af000f
  107. 13:02:03.282 [Thread-12-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  108. 13:02:03.282 [Thread-12] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af000f closed
  109. 13:02:03.282 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60779 which had sessionid 0x15486cde5af000f
  110. 13:02:03.323 [Thread-12] INFO o.a.s.d.s.SyncSupervisorEvent - Extracting resources from jar at /home/daniel/.m2/repository/org/apache/storm/flux-core/2.0.0-SNAPSHOT/flux-core-2.0.0-SNAPSHOT.jar to /tmp/b39c0ddf-0fbd-4f3c-a3f5-dd1d95419e68/supervisor/stormdist/countApp-1-1462550522/resources
  111. 13:02:03.348 [Thread-12] INFO o.a.s.d.s.SyncSupervisorEvent - Finished downloading code for storm id countApp-1-1462550522
  112. 13:02:03.352 [Thread-13] INFO o.a.s.d.s.SyncProcessEvent - Launching worker with assignment LocalAssignment(topology_id:countApp-1-1462550522, executors:[ExecutorInfo(task_start:6, task_end:6), ExecutorInfo(task_start:1, task_end:1), ExecutorInfo(task_start:2, task_end:2), ExecutorInfo(task_start:3, task_end:3), ExecutorInfo(task_start:4, task_end:4), ExecutorInfo(task_start:5, task_end:5)], resources:WorkerResources(mem_on_heap:0.0, mem_off_heap:0.0, cpu:0.0)) for this supervisor 85441820-1179-41fe-9b70-5d4e6de3e821 on port 1024 with id 64057fe5-5ab6-4136-b686-cf7324edb920
  113. 13:02:03.355 [Thread-13] INFO o.a.s.d.worker - Launching worker for countApp-1-1462550522 on 85441820-1179-41fe-9b70-5d4e6de3e821:1024 with id 64057fe5-5ab6-4136-b686-cf7324edb920 and conf {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", "storm.messaging.netty.client_worker_threads" 1, "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.cgroup.resources" ["cpu" "memory"], "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "storm.resource.isolation.plugin" "org.apache.storm.container.cgroup.CgroupManager", "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "drpc.queue.size" 128, "topology.backpressure.enable" true, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "storm.cgroup.hierarchy.dir" "/cgroup/storm_resources", "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" nil, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" nil, "storm.cgroup.cgexec.cmd" "/bin/cgexec", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/tmp/b39c0ddf-0fbd-4f3c-a3f5-dd1d95419e68", "topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" nil, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 4096.0, "storm.supervisor.cgroup.rootdir" "storm", "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "storm.cgroup.hierarchy.name" "storm", "task.credentials.poll.secs" 30, "blobstore.superuser" "daniel", "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" (1024 1025 1026), "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.cgroup.memory.limit.tolerance.margin.mb" 128.0, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster.ZKStateStorageFactory", "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "storm.resource.isolation.plugin.enable" false, "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "nimbus.impersonation.authorizer" "org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "supervisor.childopts" "-Xmx256m", "topology.worker.max.heap.size.mb" 768.0, "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "storm.supervisor.worker.manager.plugin" "org.apache.storm.daemon.supervisor.workermanager.DefaultWorkerManager", "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib", "ui.port" 8080, "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600}
  114. 13:02:03.362 [Thread-13] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  115. 13:02:03.362 [Thread-13] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5abce6eb
  116. 13:02:03.363 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  117. 13:02:03.363 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  118. 13:02:03.363 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60780
  119. 13:02:03.363 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60780
  120. 13:02:03.382 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af0010 with negotiated timeout 20000 for client /127.0.0.1:60780
  121. 13:02:03.382 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af0010, negotiated timeout = 20000
  122. 13:02:03.382 [Thread-13-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  123. 13:02:03.383 [Curator-Framework-0] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - backgroundOperationsLoop exiting
  124. 13:02:03.383 [ProcessThread(sid:0 cport:-1):] INFO o.a.s.s.o.a.z.s.PrepRequestProcessor - Processed session termination for sessionid: 0x15486cde5af0010
  125. 13:02:03.390 [Thread-13] INFO o.a.s.s.o.a.z.ZooKeeper - Session: 0x15486cde5af0010 closed
  126. 13:02:03.390 [Thread-13-EventThread] INFO o.a.s.s.o.a.z.ClientCnxn - EventThread shut down
  127. 13:02:03.390 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxn - Closed socket connection for client /127.0.0.1:60780 which had sessionid 0x15486cde5af0010
  128. 13:02:03.391 [Thread-13] INFO o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl - Starting
  129. 13:02:03.391 [Thread-13] INFO o.a.s.s.o.a.z.ZooKeeper - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4955cde5
  130. 13:02:03.391 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Opening socket connection to server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error)
  131. 13:02:03.392 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Socket connection established to hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, initiating session
  132. 13:02:03.392 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.NIOServerCnxnFactory - Accepted socket connection from /127.0.0.1:60781
  133. 13:02:03.392 [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2000] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Client attempting to establish new session at /127.0.0.1:60781
  134. 13:02:03.398 [SyncThread:0] INFO o.a.s.s.o.a.z.s.ZooKeeperServer - Established session 0x15486cde5af0011 with negotiated timeout 20000 for client /127.0.0.1:60781
  135. 13:02:03.399 [Thread-13-SendThread(hp13-HP-EliteDesk-800-G1-SFF:2000)] INFO o.a.s.s.o.a.z.ClientCnxn - Session establishment complete on server hp13-HP-EliteDesk-800-G1-SFF/127.0.0.1:2000, sessionid = 0x15486cde5af0011, negotiated timeout = 20000
  136. 13:02:03.399 [Thread-13-EventThread] INFO o.a.s.s.o.a.c.f.s.ConnectionStateManager - State change: CONNECTED
  137. 13:02:03.403 [Thread-13] INFO o.a.s.s.a.AuthUtils - Got AutoCreds []
  138. 13:02:03.406 [Thread-13] INFO o.a.s.d.worker - Reading Assignments.
  139. 13:02:03.460 [Thread-13] INFO o.a.s.d.worker - Registering IConnectionCallbacks for 85441820-1179-41fe-9b70-5d4e6de3e821:1024
  140. 13:02:03.494 [Thread-13] INFO o.a.s.d.executor - Loading executor __eventlogger:[2 2]
  141. 13:02:03.511 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks __eventlogger:[2 2]
  142. 13:02:03.679 [Thread-13] INFO o.a.s.d.executor - Timeouts disabled for executor __eventlogger:[2 2]
  143. 13:02:03.680 [Thread-13] INFO o.a.s.d.executor - Finished loading executor __eventlogger:[2 2]
  144. 13:02:03.695 [Thread-13] INFO o.a.s.d.executor - Loading executor defaultSpout:[3 3]
  145. 13:02:03.709 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks defaultSpout:[3 3]
  146. 13:02:03.714 [Thread-13] INFO o.a.s.d.executor - Finished loading executor defaultSpout:[3 3]
  147. 13:02:03.720 [Thread-13] INFO o.a.s.d.executor - Loading executor __acker:[1 1]
  148. 13:02:03.734 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks __acker:[1 1]
  149. 13:02:03.735 [Thread-13] INFO o.a.s.d.executor - Timeouts disabled for executor __acker:[1 1]
  150. 13:02:03.735 [Thread-13] INFO o.a.s.d.executor - Finished loading executor __acker:[1 1]
  151. 13:02:03.741 [Thread-13] INFO o.a.s.d.executor - Loading executor textBolt:[6 6]
  152. 13:02:03.754 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks textBolt:[6 6]
  153. 13:02:03.756 [Thread-13] INFO o.a.s.d.executor - Finished loading executor textBolt:[6 6]
  154. 13:02:03.761 [Thread-13] INFO o.a.s.d.executor - Loading executor __system:[-1 -1]
  155. 13:02:03.770 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks __system:[-1 -1]
  156. 13:02:03.771 [Thread-13] INFO o.a.s.d.executor - Finished loading executor __system:[-1 -1]
  157. 13:02:03.777 [Thread-13] INFO o.a.s.d.executor - Loading executor numberBolt:[5 5]
  158. 13:02:03.786 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks numberBolt:[5 5]
  159. 13:02:03.788 [Thread-13] INFO o.a.s.d.executor - Finished loading executor numberBolt:[5 5]
  160. 13:02:03.794 [Thread-13] INFO o.a.s.d.executor - Loading executor mergeBolt:[4 4]
  161. 13:02:03.800 [Thread-13] INFO o.a.s.d.executor - Loaded executor tasks mergeBolt:[4 4]
  162. 13:02:03.802 [Thread-13] INFO o.a.s.d.executor - Finished loading executor mergeBolt:[4 4]
  163. 13:02:03.812 [Thread-13] INFO o.a.s.d.worker - Started with log levels: {"" #object[org.apache.logging.log4j.Level 0x3536a242 "INFO"], "org.apache.zookeeper" #object[org.apache.logging.log4j.Level 0x15d6b659 "WARN"]}
  164. 13:02:03.817 [Thread-13] INFO o.a.s.d.worker - Worker has topology config {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", "storm.messaging.netty.client_worker_threads" 1, "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.cgroup.resources" ["cpu" "memory"], "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "storm.resource.isolation.plugin" "org.apache.storm.container.cgroup.CgroupManager", "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "drpc.queue.size" 128, "topology.backpressure.enable" true, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "storm.cgroup.hierarchy.dir" "/cgroup/storm_resources", "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "topology.submitter.principal" "", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" nil, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" nil, "storm.cgroup.cgexec.cmd" "/bin/cgexec", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/tmp/f0efa436-870f-4f84-88a7-adbb8d89828d", "topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 120, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" nil, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "storm.zookeeper.superACL" nil, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 4096.0, "storm.supervisor.cgroup.rootdir" "storm", "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "storm.cgroup.hierarchy.name" "storm", "task.credentials.poll.secs" 30, "blobstore.superuser" "daniel", "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" [6700 6701 6702 6703], "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "topology.users" [], "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.cgroup.memory.limit.tolerance.margin.mb" 128.0, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster.ZKStateStorageFactory", "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "storm.resource.isolation.plugin.enable" false, "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.submitter.user" "daniel", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "nimbus.impersonation.authorizer" "org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "topology.kryo.register" nil, "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "supervisor.childopts" "-Xmx256m", "topology.worker.max.heap.size.mb" 768.0, "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "storm.supervisor.worker.manager.plugin" "org.apache.storm.daemon.supervisor.workermanager.DefaultWorkerManager", "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib", "ui.port" 8080, "topology.kryo.decorators" [], "storm.id" "countApp-1-1462550522", "topology.name" "countApp", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600}
  165. 13:02:03.817 [Thread-13] INFO o.a.s.d.worker - Worker 64057fe5-5ab6-4136-b686-cf7324edb920 for storm countApp-1-1462550522 on 85441820-1179-41fe-9b70-5d4e6de3e821:1024 has finished loading
  166. 13:02:03.817 [Thread-13] INFO o.a.s.u.ConfigUtils - SET worker-user 64057fe5-5ab6-4136-b686-cf7324edb920
  167. 13:02:04.445 [refresh-active-timer] INFO o.a.s.d.worker - All connections are ready for worker 85441820-1179-41fe-9b70-5d4e6de3e821:1024 with id 64057fe5-5ab6-4136-b686-cf7324edb920
  168. 13:02:04.460 [Thread-36-textBolt-executor[6 6]] INFO o.a.s.d.executor - Preparing bolt textBolt:(6)
  169. 13:02:04.462 [Thread-36-textBolt-executor[6 6]] INFO o.a.s.d.executor - Prepared bolt textBolt:(6)
  170. 13:02:04.472 [Thread-38-__system-executor[-1 -1]] INFO o.a.s.d.executor - Preparing bolt __system:(-1)
  171. 13:02:04.475 [Thread-38-__system-executor[-1 -1]] INFO o.a.s.d.executor - Prepared bolt __system:(-1)
  172. 13:02:04.480 [Thread-30-__eventlogger-executor[2 2]] INFO o.a.s.d.executor - Preparing bolt __eventlogger:(2)
  173. 13:02:04.480 [Thread-30-__eventlogger-executor[2 2]] INFO o.a.s.m.EventLoggerBolt - EventLoggerBolt prepare called
  174. 13:02:04.489 [Thread-40-numberBolt-executor[5 5]] INFO o.a.s.d.executor - Preparing bolt numberBolt:(5)
  175. 13:02:04.490 [Thread-40-numberBolt-executor[5 5]] INFO o.a.s.d.executor - Prepared bolt numberBolt:(5)
  176. 13:02:04.502 [Thread-42-mergeBolt-executor[4 4]] INFO o.a.s.d.executor - Preparing bolt mergeBolt:(4)
  177. 13:02:04.503 [Thread-42-mergeBolt-executor[4 4]] INFO o.a.s.d.executor - Prepared bolt mergeBolt:(4)
  178. 13:02:04.514 [Thread-32-defaultSpout-executor[3 3]] INFO o.a.s.d.executor - Opening spout defaultSpout:(3)
  179. 13:02:04.515 [Thread-32-defaultSpout-executor[3 3]] INFO o.a.s.d.executor - Opened spout defaultSpout:(3)
  180. 13:02:04.517 [Thread-32-defaultSpout-executor[3 3]] INFO o.a.s.d.executor - Activating spout defaultSpout:(3)
  181. 13:02:04.517 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Activate DefaultSpout
  182. 13:02:04.521 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 4"
  183. 13:02:04.536 [Thread-34-__acker-executor[1 1]] INFO o.a.s.d.executor - Preparing bolt __acker:(1)
  184. 13:02:04.537 [Thread-34-__acker-executor[1 1]] INFO o.a.s.d.executor - Prepared bolt __acker:(1)
  185. 13:02:04.569 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072916-1462550524517
  186. 13:02:04.572 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 1"
  187. 13:02:04.587 [Thread-30-__eventlogger-executor[2 2]] INFO o.a.s.m.FileBasedEventLogger - logFilePath /home/daniel/Proyectos/FONDEF/Storm/countNumber/logs/workers-artifacts/countApp-1-1462550522/1024/events.log
  188. 13:02:04.589 [Thread-30-__eventlogger-executor[2 2]] INFO o.a.s.d.executor - Prepared bolt __eventlogger:(2)
  189. 13:02:04.621 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072919-1462550524571
  190. 13:02:04.623 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 4"
  191. 13:02:04.672 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072916-1462550524621
  192. 13:02:04.674 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 7"
  193. 13:02:04.723 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072913-1462550524672
  194. 13:02:04.725 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 7"
  195. 13:02:04.774 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072913-1462550524723
  196. 13:02:04.776 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 9"
  197. 13:02:04.824 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072911-1462550524774
  198. 13:02:04.826 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 6"
  199. 13:02:04.875 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072914-1462550524825
  200. 13:02:04.878 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  201. 13:02:04.926 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550524876
  202. 13:02:04.928 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 6"
  203. 13:02:04.977 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072914-1462550524927
  204. 13:02:04.979 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  205. 13:02:05.028 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550524977
  206. 13:02:05.030 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  207. 13:02:05.079 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550525028
  208. 13:02:05.081 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 8"
  209. 13:02:05.130 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072912-1462550525079
  210. 13:02:05.133 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 1"
  211. 13:02:05.181 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072919-1462550525130
  212. 13:02:05.183 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  213. 13:02:05.232 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550525181
  214. 13:02:05.234 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 4"
  215. 13:02:05.283 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072916-1462550525232
  216. 13:02:05.286 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 8"
  217. 13:02:05.334 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072912-1462550525283
  218. 13:02:05.336 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 2"
  219. 13:02:05.385 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072918-1462550525335
  220. 13:02:05.387 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 1"
  221. 13:02:05.436 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072919-1462550525385
  222. 13:02:05.437 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 8"
  223. 13:02:05.486 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072912-1462550525436
  224. 13:02:05.489 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 2"
  225. 13:02:05.537 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072918-1462550525487
  226. 13:02:05.539 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 7"
  227. 13:02:05.588 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072913-1462550525537
  228. 13:02:05.589 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 9"
  229. 13:02:05.638 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072911-1462550525588
  230. 13:02:05.640 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 5"
  231. 13:02:05.689 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072915-1462550525639
  232. 13:02:05.692 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 1"
  233. 13:02:05.740 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072919-1462550525689
  234. 13:02:05.742 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  235. 13:02:05.790 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550525740
  236. 13:02:05.793 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 2"
  237. 13:02:05.841 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072918-1462550525791
  238. 13:02:05.843 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 3"
  239. 13:02:05.892 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072917-1462550525841
  240. 13:02:05.894 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 1"
  241. 13:02:05.943 [Thread-32-defaultSpout-executor[3 3]] INFO c.c.s.DefaultSpout - Delete tuple -292072919-1462550525892
  242. 13:02:05.945 [Thread-36-textBolt-executor[6 6]] INFO c.c.b.TextBolt - La tupla es: "Este es el tweet número 10"
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement