Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- 09:23:23.050 [ScalaTest-run] DEBUG com.spotify.docker.client.DockerCertificates - /Users/viktortnk/.docker/ca.pem, /Users/viktortnk/.docker/key.pem or /Users/viktortnk/.docker/cert.pem does not exist, not using SSL
- 09:23:23.222 [ScalaTest-run] DEBUG com.spotify.docker.client.DockerConfigReader - Using configfile: /Users/viktortnk/.dockercfg
- 09:23:23.586 [ScalaTest-run] DEBUG com.whisk.docker.testkit.DockerContainerManager - Starting containers
- 09:23:24.232 [ScalaTest-run] DEBUG com.whisk.docker.testkit.DockerContainerManager - Starting container: spotify/kafka
- 09:23:24.248 [ForkJoinPool-1-worker-4] INFO com.spotify.docker.client.DefaultDockerClient - Creating container with ContainerConfig: ContainerConfig{hostname=null, domainname=null, user=null, attachStdin=null, attachStdout=null, attachStderr=null, portSpecs=null, exposedPorts=[9092, 2181], tty=null, openStdin=null, stdinOnce=null, env=[ADVERTISED_PORT=9092, ADVERTISED_HOST=localhost], cmd=null, image=spotify/kafka, volumes={}, workingDir=null, entrypoint=null, networkDisabled=null, onBuild=null, labels=null, macAddress=null, hostConfig=HostConfig{binds=null, blkioWeight=null, blkioWeightDevice=null, blkioDeviceReadBps=null, blkioDeviceWriteBps=null, blkioDeviceReadIOps=null, blkioDeviceWriteIOps=null, containerIdFile=null, lxcConf=null, privileged=null, portBindings={9092=[PortBinding{hostIp=0.0.0.0, hostPort=9092}], 2181=[PortBinding{hostIp=0.0.0.0, hostPort=2181}]}, links=null, publishAllPorts=null, dns=null, dnsOptions=null, dnsSearch=null, extraHosts=null, volumesFrom=null, capAdd=null, capDrop=null, networkMode=null, securityOpt=null, devices=null, memory=null, memorySwap=null, memorySwappiness=null, memoryReservation=null, nanoCpus=null, cpuPeriod=null, cpuShares=null, cpusetCpus=null, cpusetMems=null, cpuQuota=null, cgroupParent=null, restartPolicy=null, logConfig=null, ipcMode=null, ulimits=null, pidMode=null, shmSize=null, oomKillDisable=null, oomScoreAdj=null, autoRemove=null, pidsLimit=null, tmpfs=null, readonlyRootfs=null, storageOpt=null}, stopSignal=null, healthcheck=null, networkingConfig=null}
- 09:23:24.421 [ForkJoinPool-1-worker-4] INFO com.whisk.docker.testkit.DockerContainerManager - starting container with id: 5cc27b972746c1389bcb3af9715b9701a77fac8cec03e110f5d9c5691e958875
- 09:23:24.422 [ForkJoinPool-1-worker-1] INFO com.spotify.docker.client.DefaultDockerClient - Starting container with Id: 5cc27b972746c1389bcb3af9715b9701a77fac8cec03e110f5d9c5691e958875
- 09:23:24.772 [ForkJoinPool-1-worker-1] INFO com.whisk.docker.testkit.DockerContainerManager - container is starting. id=5cc27b972746c1389bcb3af9715b9701a77fac8cec03e110f5d9c5691e958875
- 09:23:24.805 [ForkJoinPool-1-worker-1] DEBUG com.whisk.docker.testkit.DockerContainerManager - container entered running state. id=5cc27b972746c1389bcb3af9715b9701a77fac8cec03e110f5d9c5691e958875
- 09:23:24.805 [ForkJoinPool-1-worker-1] DEBUG com.whisk.docker.testkit.DockerContainerManager - preparing to execute ready check for container
- 09:23:27.028 [ForkJoinPool-1-worker-1] DEBUG com.whisk.docker.testkit.DockerContainerManager - container is ready. id=5cc27b972746c1389bcb3af9715b9701a77fac8cec03e110f5d9c5691e958875
- 09:23:27.028 [ForkJoinPool-1-worker-1] INFO com.whisk.docker.testkit.DockerContainerManager - container spotify/kafka is ready after 2.795s
- 9092
- 09:23:27.109 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
- 09:23:27.109 [ScalaTest-run-running-KafkaServiceTest] INFO org.apache.kafka.clients.admin.AdminClientConfig - AdminClientConfig values:
- bootstrap.servers = [localhost:9092]
- client.id = test
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 1000
- request.timeout.ms = 120000
- retries = 5
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- 09:23:27.162 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-closed:
- 09:23:27.163 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-created:
- 09:23:27.163 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent-received:
- 09:23:27.164 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent:
- 09:23:27.165 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-received:
- 09:23:27.165 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name select-time:
- 09:23:27.165 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name io-time:
- 09:23:27.170 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.Metadata - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = [])
- 09:23:27.172 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - Created Kafka admin client test
- 09:23:27.173 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: queueing Call(callName=listTopics, deadlineMs=1505550327173) with a timeout 120000 ms from now.
- 09:23:27.274 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initialize connection to node -1 for sending metadata request
- 09:23:27.274 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating connection to node -1 at localhost:9092.
- 09:23:27.291 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.bytes-sent
- 09:23:27.291 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.bytes-received
- 09:23:27.292 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.latency
- 09:23:27.295 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.network.Selector - Created socket with SO_RCVBUF = 342972, SO_SNDBUF = 146988, SO_TIMEOUT = 0 to node -1
- 09:23:27.298 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Completed connection to node -1. Fetching API versions.
- 09:23:27.299 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node -1.
- 09:23:27.345 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.network.Selector - Connection with localhost/127.0.0.1 disconnected
- java.io.EOFException: null
- at org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:87)
- at org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:75)
- at org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:203)
- at org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:167)
- at org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:379)
- at org.apache.kafka.common.network.Selector.poll(Selector.java:326)
- at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:432)
- at org.apache.kafka.clients.admin.KafkaAdminClient$AdminClientRunnable.run(KafkaAdminClient.java:927)
- at java.lang.Thread.run(Thread.java:748)
- 09:23:27.346 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Node -1 disconnected.
- 09:23:27.347 [kafka-admin-client-thread | test] ERROR org.apache.kafka.clients.admin.KafkaAdminClient - Internal server error on -1: server returned information about unknown correlation ID 0. requestHeader = {api_key=18,api_version=1,correlation_id=0,client_id=test}
- 09:23:27.347 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Manually disconnected from -1. Removed requests: .
- 09:23:27.347 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Give up sending metadata request since no node is available
- 09:23:28.348 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initialize connection to node -1 for sending metadata request
- 09:23:28.348 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating connection to node -1 at localhost:9092.
- 09:23:28.349 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.network.Selector - Created socket with SO_RCVBUF = 342972, SO_SNDBUF = 146988, SO_TIMEOUT = 0 to node -1
- 09:23:28.349 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Completed connection to node -1. Fetching API versions.
- 09:23:28.349 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node -1.
- 09:23:28.418 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node -1.
- 09:23:28.418 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v0 to send API_VERSIONS {} to node -1.
- 09:23:28.422 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 2 [usable: 2], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 1 [usable: 1], FindCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 [usable: 0], DeleteTopics(20): 0 [usable: 0], DeleteRecords(21): UNSUPPORTED, InitProducerId(22): UNSUPPORTED, OffsetForLeaderEpoch(23): UNSUPPORTED, AddPartitionsToTxn(24): UNSUPPORTED, AddOffsetsToTxn(25): UNSUPPORTED, EndTxn(26): UNSUPPORTED, WriteTxnMarkers(27): UNSUPPORTED, TxnOffsetCommit(28): UNSUPPORTED, DescribeAcls(29): UNSUPPORTED, CreateAcls(30): UNSUPPORTED, DeleteAcls(31): UNSUPPORTED, DescribeConfigs(32): UNSUPPORTED, AlterConfigs(33): UNSUPPORTED)
- 09:23:28.423 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Sending metadata request (type=MetadataRequest, topics=) to node -1
- 09:23:28.423 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v2 to send METADATA {topics=[]} to node -1.
- 09:23:28.429 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.Metadata - Updated cluster metadata version 2 to Cluster(id = eFI4GzAGTH2MLKOck__I6A, nodes = [localhost:9092 (id: 0 rack: null)], partitions = [])
- 09:23:28.429 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating connection to node 0 at localhost:9092.
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.bytes-sent
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.bytes-received
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.latency
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.network.Selector - Created socket with SO_RCVBUF = 342972, SO_SNDBUF = 146988, SO_TIMEOUT = 0 to node 0
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Completed connection to node 0. Fetching API versions.
- 09:23:28.430 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node 0.
- 09:23:28.432 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node 0.
- 09:23:28.432 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v0 to send API_VERSIONS {} to node 0.
- 09:23:28.433 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Recorded API versions for node 0: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 2 [usable: 2], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 1 [usable: 1], FindCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 [usable: 0], DeleteTopics(20): 0 [usable: 0], DeleteRecords(21): UNSUPPORTED, InitProducerId(22): UNSUPPORTED, OffsetForLeaderEpoch(23): UNSUPPORTED, AddPartitionsToTxn(24): UNSUPPORTED, AddOffsetsToTxn(25): UNSUPPORTED, EndTxn(26): UNSUPPORTED, WriteTxnMarkers(27): UNSUPPORTED, TxnOffsetCommit(28): UNSUPPORTED, DescribeAcls(29): UNSUPPORTED, CreateAcls(30): UNSUPPORTED, DeleteAcls(31): UNSUPPORTED, DescribeConfigs(32): UNSUPPORTED, AlterConfigs(33): UNSUPPORTED)
- 09:23:28.433 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v2 to send METADATA {topics=null} to node 0.
- 09:23:28.436 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: queueing Call(callName=createTopics, deadlineMs=1505550328436) with a timeout 120000 ms from now.
- 09:23:28.436 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v0 to send CREATE_TOPICS {create_topic_requests=[{topic=my-topic,num_partitions=1,replication_factor=1,replica_assignment=[],config_entries=[]}],timeout=120000} to node 0.
- 09:23:28.792 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: queueing Call(callName=listTopics, deadlineMs=1505550328792) with a timeout 120000 ms from now.
- 09:23:28.793 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v2 to send METADATA {topics=null} to node 0.
- 09:23:28.805 [ScalaTest-run-running-KafkaServiceTest] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [localhost:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 1
- max.block.ms = 60000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 0
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- 09:23:28.811 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bufferpool-wait-time
- 09:23:28.813 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name buffer-exhausted-records
- 09:23:28.813 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.Metadata - Updated cluster metadata version 1 to Cluster(id = null, nodes = [localhost:9092 (id: -1 rack: null)], partitions = [])
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name produce-throttle-time
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-closed:
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-created:
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent-received:
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent:
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-received:
- 09:23:28.815 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name select-time:
- 09:23:28.816 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name io-time:
- 09:23:28.816 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name batch-size
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name compression-rate
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name queue-time
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name request-time
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name records-per-request
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name record-retries
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name errors
- 09:23:28.817 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name record-size-max
- 09:23:28.818 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name batch-split-rate
- 09:23:28.818 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.producer.internals.Sender - Starting Kafka producer I/O thread.
- 09:23:28.819 [ScalaTest-run-running-KafkaServiceTest] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version : 0.11.0.0
- 09:23:28.820 [ScalaTest-run-running-KafkaServiceTest] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId : cb8625948210849f
- 09:23:28.820 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.producer.KafkaProducer - Kafka producer started
- 09:23:28.822 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initialize connection to node -1 for sending metadata request
- 09:23:28.822 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating connection to node -1 at localhost:9092.
- 09:23:28.822 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.bytes-sent
- 09:23:28.823 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.bytes-received
- 09:23:28.823 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node--1.latency
- 09:23:28.823 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.network.Selector - Created socket with SO_RCVBUF = 326640, SO_SNDBUF = 146988, SO_TIMEOUT = 0 to node -1
- 09:23:28.823 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Completed connection to node -1. Fetching API versions.
- 09:23:28.823 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node -1.
- 09:23:28.825 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node -1.
- 09:23:28.825 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v0 to send API_VERSIONS {} to node -1.
- 09:23:28.826 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Recorded API versions for node -1: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 2 [usable: 2], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 1 [usable: 1], FindCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 [usable: 0], DeleteTopics(20): 0 [usable: 0], DeleteRecords(21): UNSUPPORTED, InitProducerId(22): UNSUPPORTED, OffsetForLeaderEpoch(23): UNSUPPORTED, AddPartitionsToTxn(24): UNSUPPORTED, AddOffsetsToTxn(25): UNSUPPORTED, EndTxn(26): UNSUPPORTED, WriteTxnMarkers(27): UNSUPPORTED, TxnOffsetCommit(28): UNSUPPORTED, DescribeAcls(29): UNSUPPORTED, CreateAcls(30): UNSUPPORTED, DeleteAcls(31): UNSUPPORTED, DescribeConfigs(32): UNSUPPORTED, AlterConfigs(33): UNSUPPORTED)
- 09:23:28.826 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Sending metadata request (type=MetadataRequest, topics=my-topic) to node -1
- 09:23:28.826 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v2 to send METADATA {topics=[my-topic]} to node -1.
- 09:23:28.827 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.Metadata - Updated cluster metadata version 2 to Cluster(id = eFI4GzAGTH2MLKOck__I6A, nodes = [localhost:9092 (id: 0 rack: null)], partitions = [Partition(topic = my-topic, partition = 0, leader = 0, replicas = [0], isr = [0])])
- 09:23:28.837 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating connection to node 0 at localhost:9092.
- 09:23:28.837 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.bytes-sent
- 09:23:28.838 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.bytes-received
- 09:23:28.838 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name node-0.latency
- 09:23:28.838 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.network.Selector - Created socket with SO_RCVBUF = 326640, SO_SNDBUF = 146988, SO_TIMEOUT = 0 to node 0
- 09:23:28.838 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Completed connection to node 0. Fetching API versions.
- 09:23:28.838 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node 0.
- 09:23:28.839 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Initiating API versions fetch from node 0.
- 09:23:28.840 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v0 to send API_VERSIONS {} to node 0.
- 09:23:28.841 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Recorded API versions for node 0: (Produce(0): 0 to 2 [usable: 2], Fetch(1): 0 to 3 [usable: 3], Offsets(2): 0 to 1 [usable: 1], Metadata(3): 0 to 2 [usable: 2], LeaderAndIsr(4): 0 [usable: 0], StopReplica(5): 0 [usable: 0], UpdateMetadata(6): 0 to 2 [usable: 2], ControlledShutdown(7): 1 [usable: 1], OffsetCommit(8): 0 to 2 [usable: 2], OffsetFetch(9): 0 to 1 [usable: 1], FindCoordinator(10): 0 [usable: 0], JoinGroup(11): 0 to 1 [usable: 1], Heartbeat(12): 0 [usable: 0], LeaveGroup(13): 0 [usable: 0], SyncGroup(14): 0 [usable: 0], DescribeGroups(15): 0 [usable: 0], ListGroups(16): 0 [usable: 0], SaslHandshake(17): 0 [usable: 0], ApiVersions(18): 0 [usable: 0], CreateTopics(19): 0 [usable: 0], DeleteTopics(20): 0 [usable: 0], DeleteRecords(21): UNSUPPORTED, InitProducerId(22): UNSUPPORTED, OffsetForLeaderEpoch(23): UNSUPPORTED, AddPartitionsToTxn(24): UNSUPPORTED, AddOffsetsToTxn(25): UNSUPPORTED, EndTxn(26): UNSUPPORTED, WriteTxnMarkers(27): UNSUPPORTED, TxnOffsetCommit(28): UNSUPPORTED, DescribeAcls(29): UNSUPPORTED, CreateAcls(30): UNSUPPORTED, DeleteAcls(31): UNSUPPORTED, DescribeConfigs(32): UNSUPPORTED, AlterConfigs(33): UNSUPPORTED)
- 09:23:28.842 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name topic.my-topic.records-per-batch
- 09:23:28.842 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name topic.my-topic.bytes
- 09:23:28.842 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name topic.my-topic.compression-rate
- 09:23:28.842 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name topic.my-topic.record-retries
- 09:23:28.842 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name topic.my-topic.record-errors
- 09:23:28.844 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - Using older server API v2 to send PRODUCE {acks=-1,timeout=30000,partitionSizes=[my-topic-0=44]} to node 0.
- 09:23:28.871 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: initiating close operation.
- 09:23:28.871 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: waiting for the I/O thread to exit. Hard shutdown in 31536000000 ms.
- 09:23:28.872 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-closed:
- 09:23:28.872 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-created:
- 09:23:28.872 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent-received:
- 09:23:28.872 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent:
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-received:
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name select-time:
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name io-time:
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.bytes-sent
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.bytes-received
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.latency
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-sent
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-received
- 09:23:28.873 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.latency
- 09:23:28.874 [kafka-admin-client-thread | test] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: exiting AdminClientRunnable thread.
- 09:23:28.874 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.admin.KafkaAdminClient - test: closed.
- 09:23:28.874 [ScalaTest-run-running-KafkaServiceTest] INFO org.apache.kafka.clients.producer.KafkaProducer - Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms.
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.producer.internals.Sender - Beginning shutdown of Kafka producer I/O thread, sending remaining records.
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-closed:
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name connections-created:
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent-received:
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-sent:
- 09:23:28.874 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name bytes-received:
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name select-time:
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name io-time:
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.bytes-sent
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.bytes-received
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node--1.latency
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-sent
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.bytes-received
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.metrics.Metrics - Removed sensor with name node-0.latency
- 09:23:28.875 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.producer.internals.Sender - Shutdown of Kafka producer I/O thread has completed.
- 09:23:28.876 [ScalaTest-run-running-KafkaServiceTest] DEBUG org.apache.kafka.clients.producer.KafkaProducer - The Kafka producer has closed.
- Process finished with exit code 0
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement