Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [2016-07-31 03:18:36,005] INFO StandaloneConfig values:
- value.converter = class io.confluent.connect.avro.AvroConverter
- offset.storage.file.filename = /tmp/connect.offsets
- access.control.allow.methods =
- key.converter = class io.confluent.connect.avro.AvroConverter
- offset.flush.timeout.ms = 5000
- rest.port = 8083
- rest.advertised.port = null
- access.control.allow.origin =
- rest.advertised.host.name = null
- bootstrap.servers = [localhost:9092]
- task.shutdown.graceful.timeout.ms = 5000
- rest.host.name = null
- internal.value.converter = class org.apache.kafka.connect.json.JsonConverter
- cluster = connect
- internal.key.converter = class org.apache.kafka.connect.json.JsonConverter
- offset.flush.interval.ms = 60000
- (org.apache.kafka.connect.runtime.standalone.StandaloneConfig:178)
- [2016-07-31 03:18:36,203] INFO Logging initialized @1642ms (org.eclipse.jetty.util.log:186)
- [2016-07-31 03:18:36,976] INFO Kafka Connect starting (org.apache.kafka.connect.runtime.Connect:52)
- [2016-07-31 03:18:36,977] INFO Herder starting (org.apache.kafka.connect.runtime.standalone.StandaloneHerder:71)
- [2016-07-31 03:18:36,977] INFO Worker starting (org.apache.kafka.connect.runtime.Worker:102)
- [2016-07-31 03:18:36,991] INFO ProducerConfig values:
- interceptor.classes = null
- request.timeout.ms = 2147483647
- ssl.truststore.password = null
- retry.backoff.ms = 100
- buffer.memory = 33554432
- batch.size = 16384
- ssl.keymanager.algorithm = SunX509
- receive.buffer.bytes = 32768
- ssl.key.password = null
- ssl.cipher.suites = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.service.name = null
- ssl.provider = null
- max.in.flight.requests.per.connection = 1
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- bootstrap.servers = [localhost:9092]
- client.id =
- max.request.size = 1048576
- acks = all
- linger.ms = 0
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- metadata.fetch.timeout.ms = 60000
- ssl.endpoint.identification.algorithm = null
- ssl.keystore.location = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- ssl.truststore.location = null
- ssl.keystore.password = null
- block.on.buffer.full = false
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- metrics.sample.window.ms = 30000
- security.protocol = PLAINTEXT
- metadata.max.age.ms = 300000
- ssl.protocol = TLS
- sasl.kerberos.min.time.before.relogin = 60000
- timeout.ms = 30000
- connections.max.idle.ms = 540000
- ssl.trustmanager.algorithm = PKIX
- metric.reporters = []
- ssl.truststore.type = JKS
- compression.type = none
- retries = 2147483647
- max.block.ms = 9223372036854775807
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- send.buffer.bytes = 131072
- reconnect.backoff.ms = 50
- metrics.num.samples = 2
- ssl.keystore.type = JKS
- (org.apache.kafka.clients.producer.ProducerConfig:178)
- [2016-07-31 03:18:37,034] INFO ProducerConfig values:
- interceptor.classes = null
- request.timeout.ms = 2147483647
- ssl.truststore.password = null
- retry.backoff.ms = 100
- buffer.memory = 33554432
- batch.size = 16384
- ssl.keymanager.algorithm = SunX509
- receive.buffer.bytes = 32768
- ssl.key.password = null
- ssl.cipher.suites = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.service.name = null
- ssl.provider = null
- max.in.flight.requests.per.connection = 1
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- bootstrap.servers = [localhost:9092]
- client.id = producer-1
- max.request.size = 1048576
- acks = all
- linger.ms = 0
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- metadata.fetch.timeout.ms = 60000
- ssl.endpoint.identification.algorithm = null
- ssl.keystore.location = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- ssl.truststore.location = null
- ssl.keystore.password = null
- block.on.buffer.full = false
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- metrics.sample.window.ms = 30000
- security.protocol = PLAINTEXT
- metadata.max.age.ms = 300000
- ssl.protocol = TLS
- sasl.kerberos.min.time.before.relogin = 60000
- timeout.ms = 30000
- connections.max.idle.ms = 540000
- ssl.trustmanager.algorithm = PKIX
- metric.reporters = []
- ssl.truststore.type = JKS
- compression.type = none
- retries = 2147483647
- max.block.ms = 9223372036854775807
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- send.buffer.bytes = 131072
- reconnect.backoff.ms = 50
- metrics.num.samples = 2
- ssl.keystore.type = JKS
- (org.apache.kafka.clients.producer.ProducerConfig:178)
- [2016-07-31 03:18:37,040] INFO Kafka version : 0.10.0.0-cp1 (org.apache.kafka.common.utils.AppInfoParser:83)
- [2016-07-31 03:18:37,040] INFO Kafka commitId : 7aeb2e89dbc741f6 (org.apache.kafka.common.utils.AppInfoParser:84)
- [2016-07-31 03:18:37,041] INFO Starting FileOffsetBackingStore with file /tmp/connect.offsets (org.apache.kafka.connect.storage.FileOffsetBackingStore:60)
- [2016-07-31 03:18:37,047] INFO Worker started (org.apache.kafka.connect.runtime.Worker:124)
- [2016-07-31 03:18:37,048] INFO Herder started (org.apache.kafka.connect.runtime.standalone.StandaloneHerder:73)
- [2016-07-31 03:18:37,048] INFO Starting REST server (org.apache.kafka.connect.runtime.rest.RestServer:98)
- [2016-07-31 03:18:37,433] INFO jetty-9.2.15.v20160210 (org.eclipse.jetty.server.Server:327)
- [2016-07-31 03:18:40,025] INFO Started o.e.j.s.ServletContextHandler@625800d1{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler:744)
- [2016-07-31 03:18:40,068] INFO Started ServerConnector@6de34792{HTTP/1.1}{0.0.0.0:8083} (org.eclipse.jetty.server.ServerConnector:266)
- [2016-07-31 03:18:40,070] INFO Started @5510ms (org.eclipse.jetty.server.Server:379)
- [2016-07-31 03:18:40,070] INFO REST server listening at http://192.168.0.4:8083/, advertising URL http://192.168.0.4:8083/ (org.apache.kafka.connect.runtime.rest.RestServer:150)
- [2016-07-31 03:18:40,070] INFO Kafka Connect started (org.apache.kafka.connect.runtime.Connect:58)
- [2016-07-31 03:18:40,083] INFO ConnectorConfig values:
- name = test-mysql-jdbc
- tasks.max = 1
- connector.class = io.confluent.connect.jdbc.JdbcSourceConnector
- (org.apache.kafka.connect.runtime.ConnectorConfig:178)
- [2016-07-31 03:18:40,087] INFO Creating connector test-mysql-jdbc of type io.confluent.connect.jdbc.JdbcSourceConnector (org.apache.kafka.connect.runtime.Worker:168)
- [2016-07-31 03:18:40,107] INFO Instantiated connector test-mysql-jdbc with version 3.0.0 of type io.confluent.connect.jdbc.JdbcSourceConnector (org.apache.kafka.connect.runtime.Worker:176)
- [2016-07-31 03:18:40,110] INFO JdbcSourceConnectorConfig values:
- query =
- validate.non.null = true
- connection.url = jdbc:mysql://localhost:3306/demo?user=root&password=mypassword
- topic.prefix = test_jdbc_
- table.blacklist = []
- mode = timestamp+incrementing
- table.poll.interval.ms = 60000
- timestamp.delay.interval.ms = 0
- incrementing.column.name = id
- timestamp.column.name = modified
- poll.interval.ms = 5000
- batch.max.rows = 100
- table.whitelist = []
- (io.confluent.connect.jdbc.JdbcSourceConnectorConfig:178)
- [2016-07-31 03:18:40,947] INFO Finished creating connector test-mysql-jdbc (org.apache.kafka.connect.runtime.Worker:181)
- [2016-07-31 03:18:40,950] INFO SourceConnectorConfig values:
- name = test-mysql-jdbc
- tasks.max = 1
- connector.class = io.confluent.connect.jdbc.JdbcSourceConnector
- (org.apache.kafka.connect.runtime.SourceConnectorConfig:178)
- [2016-07-31 03:18:40,959] INFO TaskConfig values:
- task.class = class io.confluent.connect.jdbc.JdbcSourceTask
- (org.apache.kafka.connect.runtime.TaskConfig:178)
- [2016-07-31 03:18:40,959] INFO Creating task test-mysql-jdbc-0 (org.apache.kafka.connect.runtime.Worker:315)
- [2016-07-31 03:18:40,959] INFO Instantiated task test-mysql-jdbc-0 with version 3.0.0 of type io.confluent.connect.jdbc.JdbcSourceTask (org.apache.kafka.connect.runtime.Worker:326)
- [2016-07-31 03:18:40,986] INFO Created connector test-mysql-jdbc (org.apache.kafka.connect.cli.ConnectStandalone:91)
- [2016-07-31 03:18:40,986] INFO JdbcSourceTaskConfig values:
- query =
- validate.non.null = true
- connection.url = jdbc:mysql://localhost:3306/demo?user=root&password=mypassword
- topic.prefix = test_jdbc_
- table.blacklist = []
- mode = timestamp+incrementing
- tables = [users]
- table.poll.interval.ms = 60000
- timestamp.delay.interval.ms = 0
- incrementing.column.name = id
- timestamp.column.name = modified
- poll.interval.ms = 5000
- batch.max.rows = 100
- table.whitelist = []
- (io.confluent.connect.jdbc.JdbcSourceTaskConfig:178)
- [2016-07-31 03:18:40,996] INFO ConnectorConfig values:
- name = hdfs-sink
- tasks.max = 1
- connector.class = io.confluent.connect.hdfs.HdfsSinkConnector
- (org.apache.kafka.connect.runtime.ConnectorConfig:178)
- [2016-07-31 03:18:40,997] INFO Creating connector hdfs-sink of type io.confluent.connect.hdfs.HdfsSinkConnector (org.apache.kafka.connect.runtime.Worker:168)
- [2016-07-31 03:18:41,003] INFO Instantiated connector hdfs-sink with version 3.0.0 of type io.confluent.connect.hdfs.HdfsSinkConnector (org.apache.kafka.connect.runtime.Worker:176)
- [2016-07-31 03:18:41,011] INFO HdfsSinkConnectorConfig values:
- kerberos.ticket.renew.period.ms = 3600000
- rotate.interval.ms = -1
- hadoop.home =
- partition.duration.ms = -1
- hdfs.namenode.principal =
- schema.cache.size = 1000
- format.class = io.confluent.connect.hdfs.avro.AvroFormat
- locale =
- hive.integration = true
- hive.metastore.uris = thrift://localhost:9083
- storage.class = io.confluent.connect.hdfs.storage.HdfsStorage
- retry.backoff.ms = 5000
- timezone =
- hive.database = default
- partition.field.name = department
- hadoop.conf.dir =
- connect.hdfs.principal =
- path.format =
- filename.offset.zero.pad.width = 10
- hive.conf.dir =
- flush.size = 2
- topics.dir = topics
- schema.compatibility = BACKWARD
- shutdown.timeout.ms = 3000
- hdfs.url = hdfs://localhost:9000
- connect.hdfs.keytab =
- hdfs.authentication.kerberos = false
- partitioner.class = io.confluent.connect.hdfs.partitioner.FieldPartitioner
- hive.home =
- logs.dir = logs
- (io.confluent.connect.hdfs.HdfsSinkConnectorConfig:178)
- [2016-07-31 03:18:41,012] INFO Finished creating connector hdfs-sink (org.apache.kafka.connect.runtime.Worker:181)
- [2016-07-31 03:18:41,012] INFO SourceConnectorConfig values:
- name = hdfs-sink
- tasks.max = 1
- connector.class = io.confluent.connect.hdfs.HdfsSinkConnector
- (org.apache.kafka.connect.runtime.SourceConnectorConfig:178)
- [2016-07-31 03:18:41,013] INFO TaskConfig values:
- task.class = class io.confluent.connect.hdfs.HdfsSinkTask
- (org.apache.kafka.connect.runtime.TaskConfig:178)
- [2016-07-31 03:18:41,016] INFO Creating task hdfs-sink-0 (org.apache.kafka.connect.runtime.Worker:315)
- [2016-07-31 03:18:41,016] INFO Instantiated task hdfs-sink-0 with version 3.0.0 of type io.confluent.connect.hdfs.HdfsSinkTask (org.apache.kafka.connect.runtime.Worker:326)
- [2016-07-31 03:18:41,050] INFO ConsumerConfig values:
- interceptor.classes = null
- request.timeout.ms = 40000
- check.crcs = true
- ssl.truststore.password = null
- retry.backoff.ms = 100
- ssl.keymanager.algorithm = SunX509
- receive.buffer.bytes = 65536
- ssl.key.password = null
- ssl.cipher.suites = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.service.name = null
- ssl.provider = null
- session.timeout.ms = 30000
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- max.poll.records = 2147483647
- bootstrap.servers = [localhost:9092]
- client.id =
- fetch.max.wait.ms = 500
- fetch.min.bytes = 1
- key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- auto.offset.reset = earliest
- value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- max.partition.fetch.bytes = 1048576
- partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor]
- ssl.endpoint.identification.algorithm = null
- ssl.keystore.location = null
- ssl.truststore.location = null
- exclude.internal.topics = true
- ssl.keystore.password = null
- metrics.sample.window.ms = 30000
- security.protocol = PLAINTEXT
- metadata.max.age.ms = 300000
- auto.commit.interval.ms = 5000
- ssl.protocol = TLS
- sasl.kerberos.min.time.before.relogin = 60000
- connections.max.idle.ms = 540000
- ssl.trustmanager.algorithm = PKIX
- group.id = connect-hdfs-sink
- enable.auto.commit = false
- metric.reporters = []
- ssl.truststore.type = JKS
- send.buffer.bytes = 131072
- reconnect.backoff.ms = 50
- metrics.num.samples = 2
- ssl.keystore.type = JKS
- heartbeat.interval.ms = 3000
- (org.apache.kafka.clients.consumer.ConsumerConfig:178)
- [2016-07-31 03:18:41,072] INFO ConsumerConfig values:
- interceptor.classes = null
- request.timeout.ms = 40000
- check.crcs = true
- ssl.truststore.password = null
- retry.backoff.ms = 100
- ssl.keymanager.algorithm = SunX509
- receive.buffer.bytes = 65536
- ssl.key.password = null
- ssl.cipher.suites = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.service.name = null
- ssl.provider = null
- session.timeout.ms = 30000
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- max.poll.records = 2147483647
- bootstrap.servers = [localhost:9092]
- client.id = consumer-1
- fetch.max.wait.ms = 500
- fetch.min.bytes = 1
- key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- auto.offset.reset = earliest
- value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- max.partition.fetch.bytes = 1048576
- partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor]
- ssl.endpoint.identification.algorithm = null
- ssl.keystore.location = null
- ssl.truststore.location = null
- exclude.internal.topics = true
- ssl.keystore.password = null
- metrics.sample.window.ms = 30000
- security.protocol = PLAINTEXT
- metadata.max.age.ms = 300000
- auto.commit.interval.ms = 5000
- ssl.protocol = TLS
- sasl.kerberos.min.time.before.relogin = 60000
- connections.max.idle.ms = 540000
- ssl.trustmanager.algorithm = PKIX
- group.id = connect-hdfs-sink
- enable.auto.commit = false
- metric.reporters = []
- ssl.truststore.type = JKS
- send.buffer.bytes = 131072
- reconnect.backoff.ms = 50
- metrics.num.samples = 2
- ssl.keystore.type = JKS
- heartbeat.interval.ms = 3000
- (org.apache.kafka.clients.consumer.ConsumerConfig:178)
- [2016-07-31 03:18:41,104] INFO Kafka version : 0.10.0.0-cp1 (org.apache.kafka.common.utils.AppInfoParser:83)
- [2016-07-31 03:18:41,155] INFO Kafka commitId : 7aeb2e89dbc741f6 (org.apache.kafka.common.utils.AppInfoParser:84)
- [2016-07-31 03:18:41,160] INFO Created connector hdfs-sink (org.apache.kafka.connect.cli.ConnectStandalone:91)
- [2016-07-31 03:18:41,167] INFO HdfsSinkConnectorConfig values:
- kerberos.ticket.renew.period.ms = 3600000
- rotate.interval.ms = -1
- hadoop.home =
- partition.duration.ms = -1
- hdfs.namenode.principal =
- schema.cache.size = 1000
- format.class = io.confluent.connect.hdfs.avro.AvroFormat
- locale =
- hive.integration = true
- hive.metastore.uris = thrift://localhost:9083
- storage.class = io.confluent.connect.hdfs.storage.HdfsStorage
- retry.backoff.ms = 5000
- timezone =
- hive.database = default
- partition.field.name = department
- hadoop.conf.dir =
- connect.hdfs.principal =
- path.format =
- filename.offset.zero.pad.width = 10
- hive.conf.dir =
- flush.size = 2
- topics.dir = topics
- schema.compatibility = BACKWARD
- shutdown.timeout.ms = 3000
- hdfs.url = hdfs://localhost:9000
- connect.hdfs.keytab =
- hdfs.authentication.kerberos = false
- partitioner.class = io.confluent.connect.hdfs.partitioner.FieldPartitioner
- hive.home =
- logs.dir = logs
- (io.confluent.connect.hdfs.HdfsSinkConnectorConfig:178)
- [2016-07-31 03:18:41,175] INFO Source task WorkerSourceTask{id=test-mysql-jdbc-0} finished initialization and start (org.apache.kafka.connect.runtime.WorkerSourceTask:138)
- [2016-07-31 03:18:41,194] INFO Hadoop configuration directory (io.confluent.connect.hdfs.DataWriter:94)
- [2016-07-31 03:18:41,549] ERROR Task test-mysql-jdbc-0 threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask:142)
- org.apache.kafka.connect.errors.DataException: Failed to serialize Avro data:
- at io.confluent.connect.avro.AvroConverter.fromConnectData(AvroConverter.java:92)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.sendRecords(WorkerSourceTask.java:183)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:160)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:140)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:175)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
- at java.util.concurrent.FutureTask.run(FutureTask.java:262)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
- Caused by: java.net.ConnectException: Connection refused
- at java.net.PlainSocketImpl.socketConnect(Native Method)
- at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
- at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
- at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
- at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
- at java.net.Socket.connect(Socket.java:579)
- at java.net.Socket.connect(Socket.java:528)
- at sun.net.NetworkClient.doConnect(NetworkClient.java:180)
- at sun.net.www.http.HttpClient.openServer(HttpClient.java:432)
- at sun.net.www.http.HttpClient.openServer(HttpClient.java:527)
- at sun.net.www.http.HttpClient.<init>(HttpClient.java:211)
- at sun.net.www.http.HttpClient.New(HttpClient.java:308)
- at sun.net.www.http.HttpClient.New(HttpClient.java:326)
- at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:997)
- at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:933)
- at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:851)
- at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1092)
- at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:141)
- at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:181)
- at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:232)
- at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:224)
- at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:219)
- at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:57)
- at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:89)
- at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:72)
- at io.confluent.connect.avro.AvroConverter$Serializer.serialize(AvroConverter.java:120)
- at io.confluent.connect.avro.AvroConverter.fromConnectData(AvroConverter.java:90)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.sendRecords(WorkerSourceTask.java:183)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:160)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:140)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:175)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
- at java.util.concurrent.FutureTask.run(FutureTask.java:262)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
- at java.lang.Thread.run(Thread.java:745)
- [2016-07-31 03:18:41,550] ERROR Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask:143)
- [2016-07-31 03:18:42,322] WARN Unable to load native-hadoop library for your platform... using builtin-java classes where applicable (org.apache.hadoop.util.NativeCodeLoader:62)
- [2016-07-31 03:18:45,726] INFO Trying to connect to metastore with URI thrift://localhost:9083 (hive.metastore:376)
- [2016-07-31 03:18:47,440] INFO Connected to metastore. (hive.metastore:472)
- [2016-07-31 03:18:47,890] INFO Sink task WorkerSinkTask{id=hdfs-sink-0} finished initialization and start (org.apache.kafka.connect.runtime.WorkerSinkTask:208)
- [2016-07-31 03:18:48,157] WARN Error while fetching metadata with correlation id 1 : {test_jdbc_users=LEADER_NOT_AVAILABLE} (org.apache.kafka.clients.NetworkClient:600)
- [2016-07-31 03:18:48,180] INFO Discovered coordinator vagrant-ubuntu-trusty-64:9092 (id: 2147483647 rack: null) for group connect-hdfs-sink. (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:505)
- [2016-07-31 03:18:48,181] INFO Revoking previously assigned partitions [] for group connect-hdfs-sink (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:280)
- [2016-07-31 03:18:48,181] INFO (Re-)joining group connect-hdfs-sink (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:326)
- [2016-07-31 03:18:48,283] INFO Successfully joined group connect-hdfs-sink with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator:434)
- [2016-07-31 03:18:48,283] INFO Setting newly assigned partitions [] for group connect-hdfs-sink (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator:219)
- [2016-07-31 03:18:53,238] INFO Reflections took 15972 ms to scan 261 urls, producing 12056 keys and 79134 values (org.reflections.Reflections:229)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement