Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- . /etc/confluent/docker/mesos-setup.sh
- + . /etc/confluent/docker/mesos-setup.sh
- #!/usr/bin/env bash
- set +o nounset
- ++ set +o nounset
- if [ -z $SKIP_MESOS_AUTO_SETUP ]; then
- if [ -n $MESOS_SANDBOX ] && [ -e $MESOS_SANDBOX/.ssl/scheduler.crt ] && [ -e $MESOS_SANDBOX/.ssl/scheduler.key ]; then
- echo "Entering Mesos auto setup for Java SSL truststore. You should not see this if you are not on mesos ..."
- openssl pkcs12 -export -in $MESOS_SANDBOX/.ssl/scheduler.crt -inkey $MESOS_SANDBOX/.ssl/scheduler.key \
- -out /tmp/keypair.p12 -name keypair \
- -CAfile $MESOS_SANDBOX/.ssl/ca-bundle.crt -caname root -passout pass:export
- keytool -importkeystore \
- -deststorepass changeit -destkeypass changeit -destkeystore /tmp/kafka-keystore.jks \
- -srckeystore /tmp/keypair.p12 -srcstoretype PKCS12 -srcstorepass export \
- -alias keypair
- keytool -import \
- -trustcacerts \
- -alias root \
- -file $MESOS_SANDBOX/.ssl/ca-bundle.crt \
- -storepass changeit \
- -keystore /tmp/kafka-truststore.jks -noprompt
- fi
- fi
- ++ '[' -z ']'
- ++ '[' -n ']'
- ++ '[' -e /.ssl/scheduler.crt ']'
- set -o nounset
- ++ set -o nounset
- . /etc/confluent/docker/apply-mesos-overrides
- + . /etc/confluent/docker/apply-mesos-overrides
- #!/usr/bin/env bash
- #
- # Copyright 2016 Confluent Inc.
- #
- # Licensed under the Apache License, Version 2.0 (the "License");
- # you may not use this file except in compliance with the License.
- # You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- # Mesos DC/OS docker deployments will have HOST and PORT0
- # set for the proxying of the service.
- #
- # Use those values provide things we know we'll need.
- [ -n "${HOST:-}" ] && [ -z "${CONNECT_REST_ADVERTISED_HOST_NAME:-}" ] && \
- export CONNECT_REST_ADVERTISED_HOST_NAME=$HOST || true
- ++ '[' -n '' ']'
- ++ true
- [ -n "${PORT0:-}" ] && [ -z "${CONNECT_REST_ADVERTISED_PORT:-}" ] && \
- export CONNECT_REST_ADVERTISED_PORT=$PORT0 || true
- ++ '[' -n '' ']'
- ++ true
- # And default to 8083, which MUST match the containerPort specification
- # in the Mesos package for this service.
- [ -z "${CONNECT_REST_PORT:-}" ] && \
- export CONNECT_REST_PORT=8083 || true
- ++ '[' -z 8083 ']'
- ++ true
- echo "===> ENV Variables ..."
- + echo '===> ENV Variables ...'
- env | sort
- ===> ENV Variables ...
- + env
- + sort
- ALLOW_UNSIGNED=false
- APITEST_PORT=tcp://10.59.244.229:80
- APITEST_PORT_50001_TCP=tcp://10.59.244.229:50001
- APITEST_PORT_50001_TCP_ADDR=10.59.244.229
- APITEST_PORT_50001_TCP_PORT=50001
- APITEST_PORT_50001_TCP_PROTO=tcp
- APITEST_PORT_80_TCP=tcp://10.59.244.229:80
- APITEST_PORT_80_TCP_ADDR=10.59.244.229
- APITEST_PORT_80_TCP_PORT=80
- APITEST_PORT_80_TCP_PROTO=tcp
- APITEST_SERVICE_HOST=10.59.244.229
- APITEST_SERVICE_PORT=80
- APITEST_SERVICE_PORT_HTTP=80
- APITEST_SERVICE_PORT_WEBHOOK_PORT=50001
- API_GATEWAY_PORT=tcp://10.59.246.60:8052
- API_GATEWAY_PORT_8052_TCP=tcp://10.59.246.60:8052
- API_GATEWAY_PORT_8052_TCP_ADDR=10.59.246.60
- API_GATEWAY_PORT_8052_TCP_PORT=8052
- API_GATEWAY_PORT_8052_TCP_PROTO=tcp
- API_GATEWAY_PORT_8952_TCP=tcp://10.59.246.60:8952
- API_GATEWAY_PORT_8952_TCP_ADDR=10.59.246.60
- API_GATEWAY_PORT_8952_TCP_PORT=8952
- API_GATEWAY_PORT_8952_TCP_PROTO=tcp
- API_GATEWAY_SERVICE_HOST=10.59.246.60
- API_GATEWAY_SERVICE_PORT=8052
- API_GATEWAY_SERVICE_PORT_ADMIN=8952
- API_GATEWAY_SERVICE_PORT_API=8052
- AUTHENTICATION_SERVER_PORT=tcp://10.59.241.3:8016
- AUTHENTICATION_SERVER_PORT_8016_TCP=tcp://10.59.241.3:8016
- AUTHENTICATION_SERVER_PORT_8016_TCP_ADDR=10.59.241.3
- AUTHENTICATION_SERVER_PORT_8016_TCP_PORT=8016
- AUTHENTICATION_SERVER_PORT_8016_TCP_PROTO=tcp
- AUTHENTICATION_SERVER_PORT_8991_TCP=tcp://10.59.241.3:8991
- AUTHENTICATION_SERVER_PORT_8991_TCP_ADDR=10.59.241.3
- AUTHENTICATION_SERVER_PORT_8991_TCP_PORT=8991
- AUTHENTICATION_SERVER_PORT_8991_TCP_PROTO=tcp
- AUTHENTICATION_SERVER_SERVICE_HOST=10.59.241.3
- AUTHENTICATION_SERVER_SERVICE_PORT=8016
- AUTHENTICATION_SERVER_SERVICE_PORT_ADMIN=8991
- AUTHENTICATION_SERVER_SERVICE_PORT_API=8016
- AUTH_SERVER_PORT=tcp://10.59.240.135:8013
- AUTH_SERVER_PORT_8013_TCP=tcp://10.59.240.135:8013
- AUTH_SERVER_PORT_8013_TCP_ADDR=10.59.240.135
- AUTH_SERVER_PORT_8013_TCP_PORT=8013
- AUTH_SERVER_PORT_8013_TCP_PROTO=tcp
- AUTH_SERVER_PORT_8996_TCP=tcp://10.59.240.135:8996
- AUTH_SERVER_PORT_8996_TCP_ADDR=10.59.240.135
- AUTH_SERVER_PORT_8996_TCP_PORT=8996
- AUTH_SERVER_PORT_8996_TCP_PROTO=tcp
- AUTH_SERVER_SERVICE_HOST=10.59.240.135
- AUTH_SERVER_SERVICE_PORT=8013
- AUTH_SERVER_SERVICE_PORT_ADMIN=8996
- AUTH_SERVER_SERVICE_PORT_API=8013
- CLIENT_ANALYTICS_SERVER_PORT=tcp://10.59.246.194:8017
- CLIENT_ANALYTICS_SERVER_PORT_8017_TCP=tcp://10.59.246.194:8017
- CLIENT_ANALYTICS_SERVER_PORT_8017_TCP_ADDR=10.59.246.194
- CLIENT_ANALYTICS_SERVER_PORT_8017_TCP_PORT=8017
- CLIENT_ANALYTICS_SERVER_PORT_8017_TCP_PROTO=tcp
- CLIENT_ANALYTICS_SERVER_PORT_8997_TCP=tcp://10.59.246.194:8997
- CLIENT_ANALYTICS_SERVER_PORT_8997_TCP_ADDR=10.59.246.194
- CLIENT_ANALYTICS_SERVER_PORT_8997_TCP_PORT=8997
- CLIENT_ANALYTICS_SERVER_PORT_8997_TCP_PROTO=tcp
- CLIENT_ANALYTICS_SERVER_SERVICE_HOST=10.59.246.194
- CLIENT_ANALYTICS_SERVER_SERVICE_PORT=8017
- CLIENT_ANALYTICS_SERVER_SERVICE_PORT_ADMIN=8997
- CLIENT_ANALYTICS_SERVER_SERVICE_PORT_API=8017
- COMPONENT=kafka-connect
- CONFLUENT_DEB_VERSION=1
- CONFLUENT_MAJOR_VERSION=4
- CONFLUENT_MINOR_VERSION=0
- CONFLUENT_MVN_LABEL=
- CONFLUENT_PATCH_VERSION=2
- CONFLUENT_PLATFORM_LABEL=
- CONFLUENT_VERSION=4.0.2
- CONNECT_BOOTSTRAP_SERVERS=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092
- CONNECT_CONFIG_STORAGE_REPLICATION_FACTOR=3
- CONNECT_CONFIG_STORAGE_TOPIC=my_connect_diff-115379-68670-0_source_configs
- CONNECT_CONSUMER_REQUEST_TIMEOUT_MS=20000
- CONNECT_CONSUMER_RETRY_BACKOFF_MS=500
- CONNECT_CONSUMER_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username="user" password="3232";
- CONNECT_CONSUMER_SASL_MECHANISM=PLAIN
- CONNECT_CONSUMER_SECURITY_PROTOCOL=SASL_SSL
- CONNECT_CONSUMER_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM=https
- CONNECT_DATABASE_HISTORY_CONSUMER_REQUEST_TIMEOUT_MS=20000
- CONNECT_DATABASE_HISTORY_CONSUMER_RETRY_BACKOFF_MS=500
- CONNECT_DATABASE_HISTORY_CONSUMER_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username="user" password="pswd";
- CONNECT_DATABASE_HISTORY_CONSUMER_SASL_MECHANISM=PLAIN
- CONNECT_DATABASE_HISTORY_CONSUMER_SECURITY_PROTOCOL=SASL_SSL
- CONNECT_DATABASE_HISTORY_CONSUMER_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM=https
- CONNECT_DATABASE_HISTORY_PRODUCER_REQUEST_TIMEOUT_MS=20000
- CONNECT_DATABASE_HISTORY_PRODUCER_RETRY_BACKOFF_MS=500
- CONNECT_DATABASE_HISTORY_PRODUCER_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username="user" password="pswd";
- CONNECT_DATABASE_HISTORY_PRODUCER_SASL_MECHANISM=PLAIN
- CONNECT_DATABASE_HISTORY_PRODUCER_SECURITY_PROTOCOL=SASL_SSL
- CONNECT_DATABASE_HISTORY_PRODUCER_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM=https
- CONNECT_GROUP_ID=ci-connect-diff-115379-68670-0-source-group
- CONNECT_INTERNAL_KEY_CONVERTER=org.apache.kafka.connect.json.JsonConverter
- CONNECT_INTERNAL_VALUE_CONVERTER=org.apache.kafka.connect.json.JsonConverter
- CONNECT_KEY_CONVERTER=org.apache.kafka.connect.json.JsonConverter
- CONNECT_KEY_CONVERTER_SCHEMA_REGISTRY_URL=http://schema-registry-service:8081
- CONNECT_LOG4J_LOGGERS=org.reflections=ERROR,io.debezium.connector.mysql.SnapshotReader=WARN,org.apache.kafka.connect.runtime.WorkerSourceTask=WARN,org.apache.kafka.connect.runtime.WorkerSinkTask=WARN
- CONNECT_LOG4J_ROOT_LOGLEVEL=INFO
- CONNECT_OFFSET_FLUSH_INTERVAL_MS=30000
- CONNECT_OFFSET_FLUSH_TIMEOUT_MS=60000
- CONNECT_OFFSET_STORAGE_REPLICATION_FACTOR=3
- CONNECT_OFFSET_STORAGE_TOPIC=my_connect_diff-115379-68670-0_source_offsets
- CONNECT_PLUGIN_PATH=/opt/clover/kafka-connect-plugin/
- CONNECT_PRODUCER_REQUEST_TIMEOUT_MS=20000
- CONNECT_PRODUCER_RETRY_BACKOFF_MS=500
- CONNECT_PRODUCER_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username="user" password="pswd";
- CONNECT_PRODUCER_SASL_MECHANISM=PLAIN
- CONNECT_PRODUCER_SECURITY_PROTOCOL=SASL_SSL
- CONNECT_PRODUCER_SSL_ENDPOINT_IDENTIFICATION_ALGORITHM=https
- CONNECT_REQUEST_TIMEOUT_MS=20000
- CONNECT_REST_ADVERTISED_HOST_NAME=kafka-connect-source-service-0
- CONNECT_REST_PORT=8083
- CONNECT_RETRY_BACKOFF_MS=500
- CONNECT_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username="pswd" password="user";
- CONNECT_SASL_MECHANISM=PLAIN
- CONNECT_SECURITY_PROTOCOL=SASL_SSL
- CONNECT_SSL_ENDPOINT_INDENTIFICATION_ALGORITHM=https
- CONNECT_STATUS_STORAGE_REPLICATION_FACTOR=3
- CONNECT_STATUS_STORAGE_TOPIC=my_connect_diff-115379-68670-0_source_status
- CONNECT_TASK_SHUTDOWN_GRACEFUL_TIMEOUT_MS=60000
- CONNECT_VALUE_CONVERTER=io.confluent.connect.avro.AvroConverter
- CONNECT_VALUE_CONVERTER_SCHEMA_REGISTRY_URL=http://schema-registry-service:8081
- COS_SERVER_BACKEND_PORT=tcp://10.59.242.232:8018
- COS_SERVER_BACKEND_PORT_8018_TCP=tcp://10.59.242.232:8018
- COS_SERVER_BACKEND_PORT_8018_TCP_ADDR=10.59.242.232
- COS_SERVER_BACKEND_PORT_8018_TCP_PORT=8018
- COS_SERVER_BACKEND_PORT_8018_TCP_PROTO=tcp
- COS_SERVER_BACKEND_PORT_8998_TCP=tcp://10.59.242.232:8998
- COS_SERVER_BACKEND_PORT_8998_TCP_ADDR=10.59.242.232
- COS_SERVER_BACKEND_PORT_8998_TCP_PORT=8998
- COS_SERVER_BACKEND_PORT_8998_TCP_PROTO=tcp
- COS_SERVER_BACKEND_SERVICE_HOST=10.59.242.232
- COS_SERVER_BACKEND_SERVICE_PORT=8018
- COS_SERVER_BACKEND_SERVICE_PORT_ADMIN=8998
- COS_SERVER_BACKEND_SERVICE_PORT_API=8018
- COS_SERVER_PORT=tcp://10.59.242.3:8010
- COS_SERVER_PORT_8010_TCP=tcp://10.59.242.3:8010
- COS_SERVER_PORT_8010_TCP_ADDR=10.59.242.3
- COS_SERVER_PORT_8010_TCP_PORT=8010
- COS_SERVER_PORT_8010_TCP_PROTO=tcp
- COS_SERVER_PORT_8999_TCP=tcp://10.59.242.3:8999
- COS_SERVER_PORT_8999_TCP_ADDR=10.59.242.3
- echo "===> User"
- COS_SERVER_PORT_8999_TCP_PORT=8999
- COS_SERVER_PORT_8999_TCP_PROTO=tcp
- COS_SERVER_SERVICE_HOST=10.59.242.3
- COS_SERVER_SERVICE_PORT=8010
- COS_SERVER_SERVICE_PORT_ADMIN=8999
- COS_SERVER_SERVICE_PORT_API=8010
- CRASH_SERVER_PORT=tcp://10.59.252.95:8054
- CRASH_SERVER_PORT_8054_TCP=tcp://10.59.252.95:8054
- CRASH_SERVER_PORT_8054_TCP_ADDR=10.59.252.95
- CRASH_SERVER_PORT_8054_TCP_PORT=8054
- CRASH_SERVER_PORT_8054_TCP_PROTO=tcp
- CRASH_SERVER_PORT_8954_TCP=tcp://10.59.252.95:8954
- CRASH_SERVER_PORT_8954_TCP_ADDR=10.59.252.95
- CRASH_SERVER_PORT_8954_TCP_PORT=8954
- CRASH_SERVER_PORT_8954_TCP_PROTO=tcp
- CRASH_SERVER_SERVICE_HOST=10.59.252.95
- CRASH_SERVER_SERVICE_PORT=8054
- CRASH_SERVER_SERVICE_PORT_ADMIN=8954
- CRASH_SERVER_SERVICE_PORT_API=8054
- + echo '===> User'
- id
- + id
- CUB_CLASSPATH=/etc/confluent/docker/docker-utils.jar
- DATABASE_HISTORY_KAFKA_BOOTSTRAP_SERVERS=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092
- DATA_INFRA_BOOTSTRAP_PORT=tcp://10.59.251.121:80
- DATA_INFRA_BOOTSTRAP_PORT_80_TCP=tcp://10.59.251.121:80
- DATA_INFRA_BOOTSTRAP_PORT_80_TCP_ADDR=10.59.251.121
- DATA_INFRA_BOOTSTRAP_PORT_80_TCP_PORT=80
- DATA_INFRA_BOOTSTRAP_PORT_80_TCP_PROTO=tcp
- DATA_INFRA_BOOTSTRAP_SERVICE_HOST=10.59.251.121
- DATA_INFRA_BOOTSTRAP_SERVICE_PORT=80
- DATA_INFRA_BOOTSTRAP_SERVICE_PORT_HTTP=80
- DATA_MIGRATION_PORT=tcp://10.59.246.82:8053
- DATA_MIGRATION_PORT_8053_TCP=tcp://10.59.246.82:8053
- DATA_MIGRATION_PORT_8053_TCP_ADDR=10.59.246.82
- DATA_MIGRATION_PORT_8053_TCP_PORT=8053
- DATA_MIGRATION_PORT_8053_TCP_PROTO=tcp
- DATA_MIGRATION_PORT_8953_TCP=tcp://10.59.246.82:8953
- DATA_MIGRATION_PORT_8953_TCP_ADDR=10.59.246.82
- DATA_MIGRATION_PORT_8953_TCP_PORT=8953
- DATA_MIGRATION_PORT_8953_TCP_PROTO=tcp
- DATA_MIGRATION_SERVICE_HOST=10.59.246.82
- DATA_MIGRATION_SERVICE_PORT=8053
- DATA_MIGRATION_SERVICE_PORT_ADMIN=8953
- DATA_MIGRATION_SERVICE_PORT_API=8053
- DEBEZIUM_VERSION=0.8.3.Final
- HOME=/root
- HOSTNAME=kafka-connect-source-service-0-7755889d5c-pjxvf
- KAFKA_ADVERTISED_LISTENERS=
- KAFKA_CONNECT_SINK_SERVICE_0_PORT=tcp://10.59.242.197:8083
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_18083_TCP=tcp://10.59.242.197:18083
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_18083_TCP_ADDR=10.59.242.197
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_18083_TCP_PORT=18083
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_18083_TCP_PROTO=tcp
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_8083_TCP=tcp://10.59.242.197:8083
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_8083_TCP_ADDR=10.59.242.197
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_8083_TCP_PORT=8083
- KAFKA_CONNECT_SINK_SERVICE_0_PORT_8083_TCP_PROTO=tcp
- KAFKA_CONNECT_SINK_SERVICE_0_SERVICE_HOST=10.59.242.197
- KAFKA_CONNECT_SINK_SERVICE_0_SERVICE_PORT=8083
- KAFKA_CONNECT_SINK_SERVICE_0_SERVICE_PORT_JMX=18083
- KAFKA_CONNECT_SINK_SERVICE_0_SERVICE_PORT_KAFKA_CONNECT=8083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT=tcp://10.59.255.52:8083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_18083_TCP=tcp://10.59.255.52:18083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_18083_TCP_ADDR=10.59.255.52
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_18083_TCP_PORT=18083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_18083_TCP_PROTO=tcp
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_8083_TCP=tcp://10.59.255.52:8083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_8083_TCP_ADDR=10.59.255.52
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_8083_TCP_PORT=8083
- KAFKA_CONNECT_SOURCE_SERVICE_0_PORT_8083_TCP_PROTO=tcp
- KAFKA_CONNECT_SOURCE_SERVICE_0_SERVICE_HOST=10.59.255.52
- KAFKA_CONNECT_SOURCE_SERVICE_0_SERVICE_PORT=8083
- KAFKA_CONNECT_SOURCE_SERVICE_0_SERVICE_PORT_JMX=18083
- KAFKA_CONNECT_SOURCE_SERVICE_0_SERVICE_PORT_KAFKA_CONNECT=8083
- KAFKA_HEAP_OPTS=-Xms2g -Xmx3g
- KAFKA_JMX_HOSTNAME=0.0.0.0
- KAFKA_JMX_PORT=18083
- KAFKA_VERSION=1.0.2
- KAFKA_ZOOKEEPER_CONNECT=
- KUBERNETES_PORT=tcp://10.59.240.1:443
- KUBERNETES_PORT_443_TCP=tcp://10.59.240.1:443
- KUBERNETES_PORT_443_TCP_ADDR=10.59.240.1
- KUBERNETES_PORT_443_TCP_PORT=443
- KUBERNETES_PORT_443_TCP_PROTO=tcp
- KUBERNETES_SERVICE_HOST=10.59.240.1
- KUBERNETES_SERVICE_PORT=443
- KUBERNETES_SERVICE_PORT_HTTPS=443
- LANG=C.UTF-8
- MEMCACHED_PORT=tcp://10.59.250.150:11211
- MEMCACHED_PORT_11211_TCP=tcp://10.59.250.150:11211
- MEMCACHED_PORT_11211_TCP_ADDR=10.59.250.150
- MEMCACHED_PORT_11211_TCP_PORT=11211
- MEMCACHED_PORT_11211_TCP_PROTO=tcp
- MEMCACHED_SERVICE_HOST=10.59.250.150
- MEMCACHED_SERVICE_PORT=11211
- MEMCACHED_SERVICE_PORT_MEMCACHED=11211
- PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
- PUSH_SERVER_PORT=tcp://10.59.243.147:8011
- PUSH_SERVER_PORT_8011_TCP=tcp://10.59.243.147:8011
- PUSH_SERVER_PORT_8011_TCP_ADDR=10.59.243.147
- PUSH_SERVER_PORT_8011_TCP_PORT=8011
- PUSH_SERVER_PORT_8011_TCP_PROTO=tcp
- PUSH_SERVER_PORT_8998_TCP=tcp://10.59.243.147:8998
- PUSH_SERVER_PORT_8998_TCP_ADDR=10.59.243.147
- PUSH_SERVER_PORT_8998_TCP_PORT=8998
- PUSH_SERVER_PORT_8998_TCP_PROTO=tcp
- PUSH_SERVER_SERVICE_HOST=10.59.243.147
- PUSH_SERVER_SERVICE_PORT=8011
- PUSH_SERVER_SERVICE_PORT_ADMIN=8998
- PUSH_SERVER_SERVICE_PORT_API=8011
- PWD=/
- PYTHON_PIP_VERSION=8.1.2
- PYTHON_VERSION=2.7.9-1
- RAPCON_0_PORT=tcp://10.59.240.70:8844
- RAPCON_0_PORT_8844_TCP=tcp://10.59.240.70:8844
- RAPCON_0_PORT_8844_TCP_ADDR=10.59.240.70
- RAPCON_0_PORT_8844_TCP_PORT=8844
- RAPCON_0_PORT_8844_TCP_PROTO=tcp
- RAPCON_0_SERVICE_HOST=10.59.240.70
- RAPCON_0_SERVICE_PORT=8844
- RAPCON_0_SERVICE_PORT_RAPCON=8844
- RAPCON_1_PORT=tcp://10.59.254.116:8845
- RAPCON_1_PORT_8845_TCP=tcp://10.59.254.116:8845
- RAPCON_1_PORT_8845_TCP_ADDR=10.59.254.116
- RAPCON_1_PORT_8845_TCP_PORT=8845
- RAPCON_1_PORT_8845_TCP_PROTO=tcp
- RAPCON_1_SERVICE_HOST=10.59.254.116
- RAPCON_1_SERVICE_PORT=8845
- RAPCON_1_SERVICE_PORT_RAPCON=8845
- REDIS_PORT=tcp://10.59.249.188:6379
- REDIS_PORT_6379_TCP=tcp://10.59.249.188:6379
- REDIS_PORT_6379_TCP_ADDR=10.59.249.188
- REDIS_PORT_6379_TCP_PORT=6379
- REDIS_PORT_6379_TCP_PROTO=tcp
- REDIS_SERVICE_HOST=10.59.249.188
- REDIS_SERVICE_PORT=6379
- REDIS_SERVICE_PORT_REDIS=6379
- REPORTING_SERVER_PORT=tcp://10.59.248.169:8051
- REPORTING_SERVER_PORT_8051_TCP=tcp://10.59.248.169:8051
- REPORTING_SERVER_PORT_8051_TCP_ADDR=10.59.248.169
- REPORTING_SERVER_PORT_8051_TCP_PORT=8051
- REPORTING_SERVER_PORT_8051_TCP_PROTO=tcp
- REPORTING_SERVER_PORT_8951_TCP=tcp://10.59.248.169:8951
- REPORTING_SERVER_PORT_8951_TCP_ADDR=10.59.248.169
- REPORTING_SERVER_PORT_8951_TCP_PORT=8951
- REPORTING_SERVER_PORT_8951_TCP_PROTO=tcp
- REPORTING_SERVER_SERVICE_HOST=10.59.248.169
- REPORTING_SERVER_SERVICE_PORT=8051
- REPORTING_SERVER_SERVICE_PORT_ADMIN=8951
- REPORTING_SERVER_SERVICE_PORT_API=8051
- SCALA_VERSION=2.11
- SCHEMA_REGISTRY_SERVICE_PORT=tcp://10.59.243.48:8081
- SCHEMA_REGISTRY_SERVICE_PORT_18081_TCP=tcp://10.59.243.48:18081
- SCHEMA_REGISTRY_SERVICE_PORT_18081_TCP_ADDR=10.59.243.48
- SCHEMA_REGISTRY_SERVICE_PORT_18081_TCP_PORT=18081
- SCHEMA_REGISTRY_SERVICE_PORT_18081_TCP_PROTO=tcp
- SCHEMA_REGISTRY_SERVICE_PORT_8081_TCP=tcp://10.59.243.48:8081
- SCHEMA_REGISTRY_SERVICE_PORT_8081_TCP_ADDR=10.59.243.48
- SCHEMA_REGISTRY_SERVICE_PORT_8081_TCP_PORT=8081
- SCHEMA_REGISTRY_SERVICE_PORT_8081_TCP_PROTO=tcp
- SCHEMA_REGISTRY_SERVICE_SERVICE_HOST=10.59.243.48
- SCHEMA_REGISTRY_SERVICE_SERVICE_PORT=8081
- SCHEMA_REGISTRY_SERVICE_SERVICE_PORT_JMX=18081
- SCHEMA_REGISTRY_SERVICE_SERVICE_PORT_SCHEMA_REGISTRY=8081
- SHLVL=1
- TOKENIZATION_SERVER_PORT=tcp://10.59.240.114:8059
- TOKENIZATION_SERVER_PORT_8059_TCP=tcp://10.59.240.114:8059
- TOKENIZATION_SERVER_PORT_8059_TCP_ADDR=10.59.240.114
- TOKENIZATION_SERVER_PORT_8059_TCP_PORT=8059
- TOKENIZATION_SERVER_PORT_8059_TCP_PROTO=tcp
- TOKENIZATION_SERVER_PORT_8959_TCP=tcp://10.59.240.114:8959
- TOKENIZATION_SERVER_PORT_8959_TCP_ADDR=10.59.240.114
- TOKENIZATION_SERVER_PORT_8959_TCP_PORT=8959
- TOKENIZATION_SERVER_PORT_8959_TCP_PROTO=tcp
- TOKENIZATION_SERVER_SERVICE_HOST=10.59.240.114
- TOKENIZATION_SERVER_SERVICE_PORT=8059
- TOKENIZATION_SERVER_SERVICE_PORT_ADMIN=8959
- TOKENIZATION_SERVER_SERVICE_PORT_API=8059
- ZULU_OPENJDK_VERSION=8=8.17.0.3
- _=/usr/bin/env
- ===> User
- uid=0(root) gid=0(root) groups=0(root)
- echo "===> Configuring ..."
- + echo '===> Configuring ...'
- /etc/confluent/docker/configure
- + /etc/confluent/docker/configure
- ===> Configuring ...
- dub ensure CONNECT_BOOTSTRAP_SERVERS
- + dub ensure CONNECT_BOOTSTRAP_SERVERS
- dub ensure CONNECT_GROUP_ID
- + dub ensure CONNECT_GROUP_ID
- dub ensure CONNECT_CONFIG_STORAGE_TOPIC
- + dub ensure CONNECT_CONFIG_STORAGE_TOPIC
- dub ensure CONNECT_OFFSET_STORAGE_TOPIC
- + dub ensure CONNECT_OFFSET_STORAGE_TOPIC
- dub ensure CONNECT_STATUS_STORAGE_TOPIC
- + dub ensure CONNECT_STATUS_STORAGE_TOPIC
- dub ensure CONNECT_KEY_CONVERTER
- + dub ensure CONNECT_KEY_CONVERTER
- dub ensure CONNECT_VALUE_CONVERTER
- + dub ensure CONNECT_VALUE_CONVERTER
- dub ensure CONNECT_INTERNAL_KEY_CONVERTER
- + dub ensure CONNECT_INTERNAL_KEY_CONVERTER
- dub ensure CONNECT_INTERNAL_VALUE_CONVERTER
- + dub ensure CONNECT_INTERNAL_VALUE_CONVERTER
- # This is required to avoid config bugs. You should set this to a value that is
- # resolvable by all containers.
- dub ensure CONNECT_REST_ADVERTISED_HOST_NAME
- + dub ensure CONNECT_REST_ADVERTISED_HOST_NAME
- # Default to 8083, which matches the mesos-overrides. This is here in case we extend the containers to remove the mesos overrides.
- if [ -z "$CONNECT_REST_PORT" ]; then
- export CONNECT_REST_PORT=8083
- fi
- + '[' -z 8083 ']'
- # Fix for https://issues.apache.org/jira/browse/KAFKA-3988
- if [[ $CONNECT_INTERNAL_KEY_CONVERTER == "org.apache.kafka.connect.json.JsonConverter" ]] || [[ $CONNECT_INTERNAL_VALUE_CONVERTER == "org.apache.kafka.connect.json.JsonConverter" ]]
- then
- export CONNECT_INTERNAL_KEY_CONVERTER_SCHEMAS_ENABLE=false
- export CONNECT_INTERNAL_VALUE_CONVERTER_SCHEMAS_ENABLE=false
- fi
- + [[ org.apache.kafka.connect.json.JsonConverter == \o\r\g\.\a\p\a\c\h\e\.\k\a\f\k\a\.\c\o\n\n\e\c\t\.\j\s\o\n\.\J\s\o\n\C\o\n\v\e\r\t\e\r ]]
- + export CONNECT_INTERNAL_KEY_CONVERTER_SCHEMAS_ENABLE=false
- + CONNECT_INTERNAL_KEY_CONVERTER_SCHEMAS_ENABLE=false
- + export CONNECT_INTERNAL_VALUE_CONVERTER_SCHEMAS_ENABLE=false
- + CONNECT_INTERNAL_VALUE_CONVERTER_SCHEMAS_ENABLE=false
- if [[ $CONNECT_KEY_CONVERTER == "io.confluent.connect.avro.AvroConverter" ]]
- then
- dub ensure CONNECT_KEY_CONVERTER_SCHEMA_REGISTRY_URL
- fi
- + [[ org.apache.kafka.connect.json.JsonConverter == \i\o\.\c\o\n\f\l\u\e\n\t\.\c\o\n\n\e\c\t\.\a\v\r\o\.\A\v\r\o\C\o\n\v\e\r\t\e\r ]]
- if [[ $CONNECT_VALUE_CONVERTER == "io.confluent.connect.avro.AvroConverter" ]]
- then
- dub ensure CONNECT_VALUE_CONVERTER_SCHEMA_REGISTRY_URL
- fi
- + [[ io.confluent.connect.avro.AvroConverter == \i\o\.\c\o\n\f\l\u\e\n\t\.\c\o\n\n\e\c\t\.\a\v\r\o\.\A\v\r\o\C\o\n\v\e\r\t\e\r ]]
- + dub ensure CONNECT_VALUE_CONVERTER_SCHEMA_REGISTRY_URL
- dub path /etc/"${COMPONENT}"/ writable
- + dub path /etc/kafka-connect/ writable
- dub template "/etc/confluent/docker/${COMPONENT}.properties.template" "/etc/${COMPONENT}/${COMPONENT}.properties"
- + dub template /etc/confluent/docker/kafka-connect.properties.template /etc/kafka-connect/kafka-connect.properties
- # The connect-distributed script expects the log4j config at /etc/kafka/connect-log4j.properties.
- dub template "/etc/confluent/docker/log4j.properties.template" "/etc/kafka/connect-log4j.properties"
- + dub template /etc/confluent/docker/log4j.properties.template /etc/kafka/connect-log4j.properties
- ===> Running preflight checks ...
- echo "===> Running preflight checks ... "
- + echo '===> Running preflight checks ... '
- /etc/confluent/docker/ensure
- + /etc/confluent/docker/ensure
- ===> Check if Kafka is healthy ...
- echo "===> Check if Kafka is healthy ..."
- + echo '===> Check if Kafka is healthy ...'
- if [[ -n "${CONNECT_SECURITY_PROTOCOL-}" ]] && [[ $CONNECT_SECURITY_PROTOCOL != "PLAINTEXT" ]]
- then
- cub kafka-ready \
- "${CONNECT_CUB_KAFKA_MIN_BROKERS:-1}" \
- "${CONNECT_CUB_KAFKA_TIMEOUT:-40}" \
- -b "$CONNECT_BOOTSTRAP_SERVERS" \
- --config /etc/"${COMPONENT}"/kafka-connect.properties
- else
- cub kafka-ready \
- "${CONNECT_CUB_KAFKA_MIN_BROKERS:-1}" \
- "${CONNECT_CUB_KAFKA_TIMEOUT:-40}" \
- -b "$CONNECT_BOOTSTRAP_SERVERS"
- fi
- + [[ -n SASL_SSL ]]
- + [[ SASL_SSL != \P\L\A\I\N\T\E\X\T ]]
- + cub kafka-ready 1 40 -b pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 --config /etc/kafka-connect/kafka-connect.properties
- [main] INFO org.apache.kafka.clients.admin.AdminClientConfig - AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id =
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 5
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- [main] INFO org.apache.kafka.common.security.authenticator.AbstractLogin - Successfully logged in.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'log4j.loggers' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.request.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'status.storage.replication.factor' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.security.protocol' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'offset.storage.topic' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'key.converter' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'config.storage.topic' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'rest.advertised.host.name' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'config.storage.replication.factor' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'rest.port' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'internal.value.converter' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'group.id' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'plugin.path' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'value.converter' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.security.protocol' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'status.storage.topic' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.sasl.mechanism' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'offset.flush.interval.ms' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'internal.key.converter' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'offset.storage.replication.factor' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'log4j.root.loglevel' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config.
- [main] WARN org.apache.kafka.clients.admin.AdminClientConfig - The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config.
- [main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version : 1.0.2-cp1
- [main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId : b7643f6655acac82
- echo "===> Launching ... "
- + echo '===> Launching ... '
- exec /etc/confluent/docker/launch
- + exec /etc/confluent/docker/launch
- ===> Launching ...
- ===> Launching kafka-connect ...
- [2019-02-19 15:38:49,770] INFO Kafka Connect distributed worker initializing ... (org.apache.kafka.connect.cli.ConnectDistributed)
- [2019-02-19 15:38:49,776] INFO WorkerInfo values:
- jvm.args = -Xms2g, -Xmx3g, -XX:+UseG1GC, -XX:MaxGCPauseMillis=20, -XX:InitiatingHeapOccupancyPercent=35, -XX:+ExplicitGCInvokesConcurrent, -Djava.awt.headless=true, -Dcom.sun.management.jmxremote=true, -Dcom.sun.management.jmxremote.authenticate=false, -Dcom.sun.management.jmxremote.ssl=false, -Djava.rmi.server.hostname=0.0.0.0, -Dcom.sun.management.jmxremote.local.only=false, -Dcom.sun.management.jmxremote.rmi.port=18083, -Dcom.sun.management.jmxremote.port=18083, -Dcom.sun.management.jmxremote.port=18083, -Dkafka.logs.dir=/var/log/kafka, -Dlog4j.configuration=file:/etc/kafka/connect-log4j.properties
- jvm.spec = Azul Systems, Inc., OpenJDK 64-Bit Server VM, 1.8.0_102, 25.102-b14
- jvm.classpath = /etc/kafka-connect/jars/bcpg-jdk15on-1.51.jar:/etc/kafka-connect/jars/log4j-1.2.16-c2.jar:/etc/kafka-connect/jars/bcpkix-jdk15on-1.51.jar:/etc/kafka-connect/jars/bcprov-jdk15on-1.51.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1-test.jar:/usr/share/java/kafka/paranamer-2.7.jar:/usr/share/java/kafka/javax.inject-2.5.0-b32.jar:/usr/share/java/kafka/javax.inject-1.jar:/usr/share/java/kafka/hk2-utils-2.5.0-b32.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1.jar:/usr/share/java/kafka/guava-20.0.jar:/usr/share/java/kafka/jersey-server-2.25.1.jar:/usr/share/java/kafka/snappy-java-1.1.4.jar:/usr/share/java/kafka/commons-logging-1.2.jar:/usr/share/java/kafka/jersey-container-servlet-2.25.1.jar:/usr/share/java/kafka/kafka-streams-examples-1.0.2-cp1.jar:/usr/share/java/kafka/avro-1.8.2.jar:/usr/share/java/kafka/aopalliance-repackaged-2.5.0-b32.jar:/usr/share/java/kafka/javassist-3.20.0-GA.jar:/usr/share/java/kafka/jackson-annotations-2.9.6.jar:/usr/share/java/kafka/jackson-core-2.9.6.jar:/usr/share/java/kafka/metrics-core-2.2.0.jar:/usr/share/java/kafka/jetty-continuation-9.2.22.v20170606.jar:/usr/share/java/kafka/jersey-client-2.25.1.jar:/usr/share/java/kafka/slf4j-api-1.7.25.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1-sources.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1-test-sources.jar:/usr/share/java/kafka/jackson-databind-2.9.6.jar:/usr/share/java/kafka/scala-library-2.11.12.jar:/usr/share/java/kafka/jackson-core-asl-1.9.13.jar:/usr/share/java/kafka/javax.annotation-api-1.2.jar:/usr/share/java/kafka/connect-runtime-1.0.2-cp1.jar:/usr/share/java/kafka/jetty-io-9.2.22.v20170606.jar:/usr/share/java/kafka/jersey-container-servlet-core-2.25.1.jar:/usr/share/java/kafka/jetty-util-9.2.22.v20170606.jar:/usr/share/java/kafka/jersey-common-2.25.1.jar:/usr/share/java/kafka/log4j-1.2.17.jar:/usr/share/java/kafka/zookeeper-3.4.10.jar:/usr/share/java/kafka/commons-lang3-3.5.jar:/usr/share/java/kafka/connect-transforms-1.0.2-cp1.jar:/usr/share/java/kafka/commons-digester-1.8.1.jar:/usr/share/java/kafka/hk2-api-2.5.0-b32.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1-javadoc.jar:/usr/share/java/kafka/jetty-http-9.2.22.v20170606.jar:/usr/share/java/kafka/support-metrics-common-4.0.2.jar:/usr/share/java/kafka/plexus-utils-3.0.24.jar:/usr/share/java/kafka/jersey-guava-2.25.1.jar:/usr/share/java/kafka/httpclient-4.5.2.jar:/usr/share/java/kafka/validation-api-1.1.0.Final.jar:/usr/share/java/kafka/hk2-locator-2.5.0-b32.jar:/usr/share/java/kafka/kafka_2.11-1.0.2-cp1-scaladoc.jar:/usr/share/java/kafka/commons-codec-1.9.jar:/usr/share/java/kafka/kafka-streams-1.0.2-cp1.jar:/usr/share/java/kafka/commons-compress-1.8.1.jar:/usr/share/java/kafka/xz-1.5.jar:/usr/share/java/kafka/slf4j-log4j12-1.7.25.jar:/usr/share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/share/java/kafka/commons-lang3-3.1.jar:/usr/share/java/kafka/connect-json-1.0.2-cp1.jar:/usr/share/java/kafka/commons-collections-3.2.1.jar:/usr/share/java/kafka/commons-validator-1.4.1.jar:/usr/share/java/kafka/zkclient-0.10.jar:/usr/share/java/kafka/kafka-tools-1.0.2-cp1.jar:/usr/share/java/kafka/jackson-jaxrs-json-provider-2.9.6.jar:/usr/share/java/kafka/maven-artifact-3.5.0.jar:/usr/share/java/kafka/jopt-simple-5.0.4.jar:/usr/share/java/kafka/jetty-server-9.2.22.v20170606.jar:/usr/share/java/kafka/javax.ws.rs-api-2.0.1.jar:/usr/share/java/kafka/httpcore-4.4.4.jar:/usr/share/java/kafka/kafka-log4j-appender-1.0.2-cp1.jar:/usr/share/java/kafka/lz4-java-1.4.jar:/usr/share/java/kafka/jackson-jaxrs-base-2.9.6.jar:/usr/share/java/kafka/jetty-servlets-9.2.22.v20170606.jar:/usr/share/java/kafka/commons-beanutils-1.8.3.jar:/usr/share/java/kafka/connect-api-1.0.2-cp1.jar:/usr/share/java/kafka/kafka.jar:/usr/share/java/kafka/jetty-security-9.2.22.v20170606.jar:/usr/share/java/kafka/javassist-3.21.0-GA.jar:/usr/share/java/kafka/kafka-clients-1.0.2-cp1.jar:/usr/share/java/kafka/httpmime-4.5.2.jar:/usr/share/java/kafka/jackson-mapper-asl-1.9.13.jar:/usr/share/java/kafka/jetty-servlet-9.2.22.v20170606.jar:/usr/share/java/kafka/connect-file-1.0.2-cp1.jar:/usr/share/java/kafka/argparse4j-0.7.0.jar:/usr/share/java/kafka/jackson-module-jaxb-annotations-2.9.6.jar:/usr/share/java/kafka/jersey-media-jaxb-2.25.1.jar:/usr/share/java/kafka/support-metrics-client-4.0.2.jar:/usr/share/java/kafka/rocksdbjni-5.7.3.jar:/usr/share/java/kafka/reflections-0.9.11.jar:/usr/share/java/kafka/osgi-resource-locator-1.0.1.jar:/usr/share/java/kafka/confluent-metrics-4.0.2.jar:/usr/share/java/confluent-common/slf4j-api-1.7.25.jar:/usr/share/java/confluent-common/common-utils-4.0.2.jar:/usr/share/java/confluent-common/build-tools-4.0.2.jar:/usr/share/java/confluent-common/log4j-1.2.17.jar:/usr/share/java/confluent-common/zookeeper-3.4.10.jar:/usr/share/java/confluent-common/common-config-4.0.2.jar:/usr/share/java/confluent-common/common-metrics-4.0.2.jar:/usr/share/java/confluent-common/zkclient-0.10.jar:/usr/share/java/confluent-common/jline-0.9.94.jar:/usr/share/java/confluent-common/netty-3.10.5.Final.jar:/usr/share/java/kafka-serde-tools/paranamer-2.7.jar:/usr/share/java/kafka-serde-tools/snappy-java-1.1.1.3.jar:/usr/share/java/kafka-serde-tools/jackson-annotations-2.9.6.jar:/usr/share/java/kafka-serde-tools/jackson-core-2.9.6.jar:/usr/share/java/kafka-serde-tools/slf4j-api-1.7.25.jar:/usr/share/java/kafka-serde-tools/jackson-databind-2.9.6.jar:/usr/share/java/kafka-serde-tools/jackson-core-asl-1.9.13.jar:/usr/share/java/kafka-serde-tools/kafka-avro-serializer-4.0.2.jar:/usr/share/java/kafka-serde-tools/commons-compress-1.8.1.jar:/usr/share/java/kafka-serde-tools/xz-1.5.jar:/usr/share/java/kafka-serde-tools/kafka-json-serializer-4.0.2.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-4.0.2.jar:/usr/share/java/kafka-serde-tools/kafka-connect-avro-converter-4.0.2.jar:/usr/share/java/kafka-serde-tools/jackson-mapper-asl-1.9.13.jar:/usr/share/java/kafka-serde-tools/avro-1.8.1.jar:/usr/share/java/monitoring-interceptors/monitoring-interceptors-4.0.2.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1-test.jar:/usr/bin/../share/java/kafka/paranamer-2.7.jar:/usr/bin/../share/java/kafka/javax.inject-2.5.0-b32.jar:/usr/bin/../share/java/kafka/javax.inject-1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.5.0-b32.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/guava-20.0.jar:/usr/bin/../share/java/kafka/jersey-server-2.25.1.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.4.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.25.1.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/avro-1.8.2.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.5.0-b32.jar:/usr/bin/../share/java/kafka/javassist-3.20.0-GA.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.9.6.jar:/usr/bin/../share/java/kafka/jackson-core-2.9.6.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/jersey-client-2.25.1.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.25.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1-sources.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1-test-sources.jar:/usr/bin/../share/java/kafka/jackson-databind-2.9.6.jar:/usr/bin/../share/java/kafka/scala-library-2.11.12.jar:/usr/bin/../share/java/kafka/jackson-core-asl-1.9.13.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.2.jar:/usr/bin/../share/java/kafka/connect-runtime-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/jetty-io-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.25.1.jar:/usr/bin/../share/java/kafka/jetty-util-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/jersey-common-2.25.1.jar:/usr/bin/../share/java/kafka/log4j-1.2.17.jar:/usr/bin/../share/java/kafka/zookeeper-3.4.10.jar:/usr/bin/../share/java/kafka/commons-lang3-3.5.jar:/usr/bin/../share/java/kafka/connect-transforms-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/commons-digester-1.8.1.jar:/usr/bin/../share/java/kafka/hk2-api-2.5.0-b32.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1-javadoc.jar:/usr/bin/../share/java/kafka/jetty-http-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/support-metrics-common-4.0.2.jar:/usr/bin/../share/java/kafka/plexus-utils-3.0.24.jar:/usr/bin/../share/java/kafka/jersey-guava-2.25.1.jar:/usr/bin/../share/java/kafka/httpclient-4.5.2.jar:/usr/bin/../share/java/kafka/validation-api-1.1.0.Final.jar:/usr/bin/../share/java/kafka/hk2-locator-2.5.0-b32.jar:/usr/bin/../share/java/kafka/kafka_2.11-1.0.2-cp1-scaladoc.jar:/usr/bin/../share/java/kafka/commons-codec-1.9.jar:/usr/bin/../share/java/kafka/kafka-streams-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/commons-compress-1.8.1.jar:/usr/bin/../share/java/kafka/xz-1.5.jar:/usr/bin/../share/java/kafka/slf4j-log4j12-1.7.25.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/commons-lang3-3.1.jar:/usr/bin/../share/java/kafka/connect-json-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.1.jar:/usr/bin/../share/java/kafka/commons-validator-1.4.1.jar:/usr/bin/../share/java/kafka/zkclient-0.10.jar:/usr/bin/../share/java/kafka/kafka-tools-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.9.6.jar:/usr/bin/../share/java/kafka/maven-artifact-3.5.0.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/jetty-server-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.0.1.jar:/usr/bin/../share/java/kafka/httpcore-4.4.4.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/lz4-java-1.4.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.9.6.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.8.3.jar:/usr/bin/../share/java/kafka/connect-api-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/jetty-security-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/javassist-3.21.0-GA.jar:/usr/bin/../share/java/kafka/kafka-clients-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/httpmime-4.5.2.jar:/usr/bin/../share/java/kafka/jackson-mapper-asl-1.9.13.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.2.22.v20170606.jar:/usr/bin/../share/java/kafka/connect-file-1.0.2-cp1.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.9.6.jar:/usr/bin/../share/java/kafka/jersey-media-jaxb-2.25.1.jar:/usr/bin/../share/java/kafka/support-metrics-client-4.0.2.jar:/usr/bin/../share/java/kafka/rocksdbjni-5.7.3.jar:/usr/bin/../share/java/kafka/reflections-0.9.11.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.1.jar:/usr/bin/../share/java/kafka/confluent-metrics-4.0.2.jar:/usr/bin/../share/java/confluent-support-metrics/*:/usr/share/java/confluent-support-metrics/*
- os.spec = Linux, amd64, 4.14.65+
- os.vcpus = 16
- (org.apache.kafka.connect.runtime.WorkerInfo)
- [2019-02-19 15:38:49,777] INFO Scanning for plugin classes. This might take a moment ... (org.apache.kafka.connect.cli.ConnectDistributed)
- [2019-02-19 15:38:49,815] INFO Loading plugin from: /opt/clover/kafka-connect-plugin/debezium-mysql-connector (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:50,699] INFO Registered loader: PluginClassLoader{pluginLocation=file:/opt/clover/kafka-connect-plugin/debezium-mysql-connector/} (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:50,700] INFO Added plugin 'io.debezium.connector.mysql.MySqlConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:50,700] INFO Added plugin 'io.debezium.transforms.ByLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:50,700] INFO Added plugin 'io.debezium.transforms.UnwrapFromEnvelope' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:50,713] INFO Loading plugin from: /opt/clover/kafka-connect-plugin/clover-connectors (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Registered loader: PluginClassLoader{pluginLocation=file:/opt/clover/kafka-connect-plugin/clover-connectors/} (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'com.clover.connectors.snowflake.sink.SnowflakeSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'com.clover.connectors.backfill.sink.BackfillSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'com.clover.connectors.backfill.source.BackfillSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'com.clover.converter.SecureAvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'io.confluent.connect.avro.AvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:38:56,913] INFO Added plugin 'com.clover.transformations.DebeziumShardedLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Registered loader: sun.misc.Launcher$AppClassLoader@764c12b6 (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.MockSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.MockSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.file.FileStreamSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.MockConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.file.FileStreamSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,669] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,670] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,671] INFO Added aliases 'BackfillSinkConnector' and 'BackfillSink' to plugin 'com.clover.connectors.backfill.sink.BackfillSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'BackfillSourceConnector' and 'BackfillSource' to plugin 'com.clover.connectors.backfill.source.BackfillSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'SnowflakeSinkConnector' and 'SnowflakeSink' to plugin 'com.clover.connectors.snowflake.sink.SnowflakeSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'MySqlConnector' and 'MySql' to plugin 'io.debezium.connector.mysql.MySqlConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'FileStreamSinkConnector' and 'FileStreamSink' to plugin 'org.apache.kafka.connect.file.FileStreamSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'FileStreamSourceConnector' and 'FileStreamSource' to plugin 'org.apache.kafka.connect.file.FileStreamSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'MockConnector' and 'Mock' to plugin 'org.apache.kafka.connect.tools.MockConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,672] INFO Added aliases 'MockSinkConnector' and 'MockSink' to plugin 'org.apache.kafka.connect.tools.MockSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'MockSourceConnector' and 'MockSource' to plugin 'org.apache.kafka.connect.tools.MockSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'SchemaSourceConnector' and 'SchemaSource' to plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'VerifiableSinkConnector' and 'VerifiableSink' to plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'VerifiableSourceConnector' and 'VerifiableSource' to plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'SecureAvroConverter' and 'SecureAvro' to plugin 'com.clover.converter.SecureAvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'AvroConverter' and 'Avro' to plugin 'io.confluent.connect.avro.AvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'ByteArrayConverter' and 'ByteArray' to plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'JsonConverter' and 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,673] INFO Added aliases 'StringConverter' and 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,674] INFO Added alias 'DebeziumShardedLogicalTableRouter' to plugin 'com.clover.transformations.DebeziumShardedLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,674] INFO Added alias 'ByLogicalTableRouter' to plugin 'io.debezium.transforms.ByLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,674] INFO Added alias 'UnwrapFromEnvelope' to plugin 'io.debezium.transforms.UnwrapFromEnvelope' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,674] INFO Added alias 'RegexRouter' to plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,675] INFO Added alias 'TimestampRouter' to plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,675] INFO Added alias 'ValueToKey' to plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader)
- [2019-02-19 15:39:03,692] INFO DistributedConfig values:
- access.control.allow.methods =
- access.control.allow.origin =
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id =
- config.storage.replication.factor = 3
- config.storage.topic = my_connect_diff-115379-68670-0_source_configs
- connections.max.idle.ms = 540000
- group.id = ci-connect-diff-115379-68670-0-source-group
- heartbeat.interval.ms = 3000
- internal.key.converter = class org.apache.kafka.connect.json.JsonConverter
- internal.value.converter = class org.apache.kafka.connect.json.JsonConverter
- key.converter = class org.apache.kafka.connect.json.JsonConverter
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- offset.flush.interval.ms = 30000
- offset.flush.timeout.ms = 60000
- offset.storage.partitions = 25
- offset.storage.replication.factor = 3
- offset.storage.topic = my_connect_diff-115379-68670-0_source_offsets
- plugin.path = [/opt/clover/kafka-connect-plugin/]
- rebalance.timeout.ms = 60000
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- rest.advertised.host.name = kafka-connect-source-service-0
- rest.advertised.port = null
- rest.host.name = null
- rest.port = 8083
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- session.timeout.ms = 10000
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- status.storage.partitions = 5
- status.storage.replication.factor = 3
- status.storage.topic = my_connect_diff-115379-68670-0_source_status
- task.shutdown.graceful.timeout.ms = 60000
- value.converter = class io.confluent.connect.avro.AvroConverter
- worker.sync.timeout.ms = 3000
- worker.unsync.backoff.ms = 300000
- (org.apache.kafka.connect.runtime.distributed.DistributedConfig)
- [2019-02-19 15:39:03,849] INFO Logging initialized @14423ms (org.eclipse.jetty.util.log)
- [2019-02-19 15:39:04,028] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,028] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,103] INFO Successfully logged in. (org.apache.kafka.common.security.authenticator.AbstractLogin)
- [2019-02-19 15:39:04,189] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,189] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,192] INFO Kafka Connect distributed worker initialization took 14421ms (org.apache.kafka.connect.cli.ConnectDistributed)
- [2019-02-19 15:39:04,192] INFO Kafka Connect starting (org.apache.kafka.connect.runtime.Connect)
- [2019-02-19 15:39:04,193] INFO Starting REST server (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:04,193] INFO Herder starting (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:04,194] INFO Worker starting (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:04,194] INFO Starting KafkaOffsetBackingStore (org.apache.kafka.connect.storage.KafkaOffsetBackingStore)
- [2019-02-19 15:39:04,194] INFO Starting KafkaBasedLog with topic my_connect_diff-115379-68670-0_source_offsets (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:04,199] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id =
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 5
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,225] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,225] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,225] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,226] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,227] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,228] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:04,229] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,229] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:04,285] INFO jetty-9.2.22.v20170606 (org.eclipse.jetty.server.Server)
- Feb 19, 2019 3:39:04 PM org.glassfish.jersey.internal.Errors logErrors
- WARNING: The following warnings have been detected: WARNING: The (sub)resource method createConnector in org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource contains empty path annotation.
- WARNING: The (sub)resource method listConnectors in org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource contains empty path annotation.
- WARNING: The (sub)resource method listConnectorPlugins in org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResource contains empty path annotation.
- WARNING: The (sub)resource method serverInfo in org.apache.kafka.connect.runtime.rest.resources.RootResource contains empty path annotation.
- [2019-02-19 15:39:04,706] INFO Started o.e.j.s.ServletContextHandler@7348899b{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler)
- [2019-02-19 15:39:04,718] INFO Started ServerConnector@aa4468b{HTTP/1.1}{0.0.0.0:8083} (org.eclipse.jetty.server.ServerConnector)
- [2019-02-19 15:39:04,719] INFO Started @15293ms (org.eclipse.jetty.server.Server)
- [2019-02-19 15:39:04,720] INFO REST server listening at http://10.56.152.227:8083/, advertising URL http://kafka-connect-source-service-0:8083/ (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:04,720] INFO Kafka Connect started (org.apache.kafka.connect.runtime.Connect)
- [2019-02-19 15:39:05,571] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 60000
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,596] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,597] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:05,598] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:05,598] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:05,604] INFO ConsumerConfig values:
- auto.commit.interval.ms = 5000
- auto.offset.reset = earliest
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- check.crcs = true
- client.id =
- connections.max.idle.ms = 540000
- enable.auto.commit = false
- exclude.internal.topics = true
- fetch.max.bytes = 52428800
- fetch.max.wait.ms = 500
- fetch.min.bytes = 1
- group.id = ci-connect-diff-115379-68670-0-source-group
- heartbeat.interval.ms = 3000
- interceptor.classes = null
- internal.leave.group.on.close = true
- isolation.level = read_uncommitted
- key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- max.partition.fetch.bytes = 1048576
- max.poll.interval.ms = 300000
- max.poll.records = 500
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- session.timeout.ms = 10000
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,629] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:05,630] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:05,630] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:07,450] INFO [Consumer clientId=consumer-1, groupId=ci-connect-diff-115379-68670-0-source-group] Discovered group coordinator b6-pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (id: 2147483641 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:07,826] INFO Finished reading KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_offsets (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:07,826] INFO Started KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_offsets (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:07,826] INFO Finished reading offsets topic and starting KafkaOffsetBackingStore (org.apache.kafka.connect.storage.KafkaOffsetBackingStore)
- [2019-02-19 15:39:07,827] INFO Worker started (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:07,827] INFO Starting KafkaBasedLog with topic my_connect_diff-115379-68670-0_source_status (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:07,828] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id =
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 5
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,842] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,842] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,843] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,843] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,844] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:07,845] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:07,845] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:09,039] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 60000
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 0
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,044] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:09,045] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:09,045] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:09,046] INFO ConsumerConfig values:
- auto.commit.interval.ms = 5000
- auto.offset.reset = earliest
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- check.crcs = true
- client.id =
- connections.max.idle.ms = 540000
- enable.auto.commit = false
- exclude.internal.topics = true
- fetch.max.bytes = 52428800
- fetch.max.wait.ms = 500
- fetch.min.bytes = 1
- group.id = ci-connect-diff-115379-68670-0-source-group
- heartbeat.interval.ms = 3000
- interceptor.classes = null
- internal.leave.group.on.close = true
- isolation.level = read_uncommitted
- key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
- max.partition.fetch.bytes = 1048576
- max.poll.interval.ms = 300000
- max.poll.records = 500
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- session.timeout.ms = 10000
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,049] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,050] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,051] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:09,052] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:09,052] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:10,608] INFO [Consumer clientId=consumer-2, groupId=ci-connect-diff-115379-68670-0-source-group] Discovered group coordinator b6-pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (id: 2147483641 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:11,060] INFO Finished reading KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_status (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:11,060] INFO Started KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_status (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:11,062] INFO Starting KafkaConfigBackingStore (org.apache.kafka.connect.storage.KafkaConfigBackingStore)
- [2019-02-19 15:39:11,062] INFO Starting KafkaBasedLog with topic my_connect_diff-115379-68670-0_source_configs (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:11,062] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id =
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 5
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,069] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:39:11,070] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:11,070] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:12,262] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 60000
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,268] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:12,269] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:12,269] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:12,269] INFO ConsumerConfig values:
- auto.commit.interval.ms = 5000
- auto.offset.reset = earliest
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- check.crcs = true
- client.id =
- connections.max.idle.ms = 540000
- enable.auto.commit = false
- exclude.internal.topics = true
- fetch.max.bytes = 52428800
- fetch.max.wait.ms = 500
- fetch.min.bytes = 1
- group.id = ci-connect-diff-115379-68670-0-source-group
- heartbeat.interval.ms = 3000
- interceptor.classes = null
- internal.leave.group.on.close = true
- isolation.level = read_uncommitted
- key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
- max.partition.fetch.bytes = 1048576
- max.poll.interval.ms = 300000
- max.poll.records = 500
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- session.timeout.ms = 10000
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer
- (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'log4j.loggers' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'producer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'consumer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'rest.advertised.host.name' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'rest.port' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'internal.value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'internal.key.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'producer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.consumer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.sasl.jaas.config' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'ssl.endpoint.indentification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'database.history.producer.security.protocol' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,273] WARN The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'consumer.request.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'consumer.retry.backoff.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'producer.sasl.mechanism' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'internal.key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'internal.value.converter.schemas.enable' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'log4j.root.loglevel' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] WARN The configuration 'producer.ssl.endpoint.identification.algorithm' was supplied but isn't a known config. (org.apache.kafka.clients.consumer.ConsumerConfig)
- [2019-02-19 15:39:12,274] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:12,274] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:13,680] INFO [Consumer clientId=consumer-3, groupId=ci-connect-diff-115379-68670-0-source-group] Discovered group coordinator b6-pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (id: 2147483641 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:14,130] INFO Finished reading KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_configs (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:14,130] INFO Started KafkaBasedLog for topic my_connect_diff-115379-68670-0_source_configs (org.apache.kafka.connect.util.KafkaBasedLog)
- [2019-02-19 15:39:14,130] INFO Started KafkaConfigBackingStore (org.apache.kafka.connect.storage.KafkaConfigBackingStore)
- [2019-02-19 15:39:14,130] INFO Herder started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:14,493] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Discovered group coordinator b6-pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (id: 2147483641 rack: null) (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:14,497] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:17,871] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 1 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:39:17,872] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-adee3e28-1122-4eff-a180-868279783d19', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=9, connectorIds=[mysql-master-orders-0, mysql-master-orders-1, mysql-master-orders-2], taskIds=[mysql-master-orders-0-0, mysql-master-orders-1-0, mysql-master-orders-2-0]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,873] WARN Catching up to assignment's config offset. (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,873] INFO Current config state offset -1 is behind group assignment 9, reading to end of config log (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,914] INFO Finished reading to end of log and updated config snapshot, new config log offset: 9 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,914] INFO Starting connectors and tasks using config offset 9 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,916] INFO Starting connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,917] INFO Starting connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,916] INFO Starting connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,917] INFO Starting task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,918] INFO Starting task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,918] INFO Starting task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:17,918] INFO Creating task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,918] INFO Creating task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,918] INFO Creating task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,919] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:39:17,924] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,924] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,924] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,924] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,925] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,926] INFO Creating connector mysql-master-orders-2 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,944] INFO Creating connector mysql-master-orders-0 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,945] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:17,945] INFO Creating connector mysql-master-orders-1 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,950] INFO Instantiated connector mysql-master-orders-2 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,950] INFO Instantiated connector mysql-master-orders-0 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,950] INFO Instantiated connector mysql-master-orders-1 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,951] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:39:17,951] INFO Instantiated task mysql-master-orders-2-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,951] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:39:17,951] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:39:17,951] INFO Instantiated task mysql-master-orders-1-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,951] INFO Instantiated task mysql-master-orders-0-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:17,959] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:39:17,959] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:39:17,959] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:39:18,145] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:39:18,145] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:39:18,145] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:39:18,148] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:39:18,148] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:39:18,148] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:39:18,157] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:18,157] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:18,157] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:39:18,165] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,166] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,166] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,167] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,167] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,167] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:39:18,349] INFO Finished creating connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:18,349] INFO Finished creating connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:18,349] INFO Finished creating connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:39:18,352] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:39:18,352] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:39:18,352] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:39:18,353] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:18,353] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:18,353] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:39:18,354] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:39:18,384] INFO 10.56.152.1 - - [19/Feb/2019:15:39:11 +0000] "GET /connectors HTTP/1.1" 200 73 7057 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:18,411] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,411] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,411] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-1.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO group.id = mysql_master_orders_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,414] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.whitelist = orders_shard_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.server.id = 2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-2.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO name = mysql-master-orders-1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO group.id = mysql_master_orders_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-0.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO group.id = mysql_master_orders_0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.whitelist = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.server.id = 1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,415] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO name = mysql-master-orders-0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.whitelist = orders_shard_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.server.id = 3 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,416] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO name = mysql-master-orders-2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:18,417] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:39:21,269] INFO 10.56.152.1 - - [19/Feb/2019:15:39:21 +0000] "GET /connectors HTTP/1.1" 200 73 5 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:27,191] INFO 10.56.73.73 - - [19/Feb/2019:15:39:27 +0000] "HEAD /connectors HTTP/1.1" 200 73 5 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:31,268] INFO 10.56.152.1 - - [19/Feb/2019:15:39:31 +0000] "GET /connectors HTTP/1.1" 200 73 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:41,268] INFO 10.56.152.1 - - [19/Feb/2019:15:39:41 +0000] "GET /connectors HTTP/1.1" 200 73 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:39:51,268] INFO 10.56.152.1 - - [19/Feb/2019:15:39:51 +0000] "GET /connectors HTTP/1.1" 200 73 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:40:01,268] INFO 10.56.152.1 - - [19/Feb/2019:15:40:01 +0000] "GET /connectors HTTP/1.1" 200 73 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:40:11,269] INFO 10.56.152.1 - - [19/Feb/2019:15:40:11 +0000] "GET /connectors HTTP/1.1" 200 73 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:40:17,909] INFO Removed connector mysql-master-orders-0 due to null configuration. This is usually intentional and does not indicate an issue. (org.apache.kafka.connect.storage.KafkaConfigBackingStore)
- [2019-02-19 15:40:17,909] INFO Connector mysql-master-orders-0 config removed (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:40:18,450] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:40:18,450] INFO 10.56.73.73 - - [19/Feb/2019:15:40:17 +0000] "DELETE /connectors/mysql-master-orders-0 HTTP/1.1" 204 - 1272 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:40:18,451] INFO Stopping connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,451] INFO Stopping connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,451] INFO Stopping connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,452] INFO Stopping task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,452] INFO Stopping task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,452] INFO Stopping task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,453] INFO Stopped connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,453] INFO Stopped connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:40:18,455] INFO Stopped connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:18,453] ERROR Graceful stop of task mysql-master-orders-0-0 failed. (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:18,453] ERROR Graceful stop of task mysql-master-orders-2-0 failed. (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:18,454] ERROR Graceful stop of task mysql-master-orders-1-0 failed. (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:18,454] INFO Finished stopping tasks in preparation for rebalance (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:18,454] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:41:21,539] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 3 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:41:21,540] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-95e580ed-3657-44e8-9161-383b8c605e1b', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=11, connectorIds=[mysql-master-orders-1, mysql-master-orders-2], taskIds=[mysql-master-orders-1-0, mysql-master-orders-2-0]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Starting connectors and tasks using config offset 11 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Starting connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Starting connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Starting task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Creating task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,541] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:41:21,541] INFO Starting task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,541] INFO Creating task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,541] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:41:21,541] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:41:21,541] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,541] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:41:21,541] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,542] INFO Creating connector mysql-master-orders-2 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,541] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,541] INFO Creating connector mysql-master-orders-1 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,542] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,543] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:41:21,543] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:41:21,544] INFO Instantiated task mysql-master-orders-2-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,544] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:41:21,545] INFO Instantiated connector mysql-master-orders-1 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,545] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:41:21,545] INFO Instantiated connector mysql-master-orders-2 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,545] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:41:21,545] INFO Instantiated task mysql-master-orders-1-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,545] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:41:21,545] INFO Finished creating connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,545] INFO Finished creating connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,545] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:41:21,548] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:41:21,548] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:41:21,548] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,548] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:41:21,548] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:41:21,549] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:41:21,549] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:41:21,585] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,586] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,586] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,587] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,592] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,594] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-2.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO group.id = mysql_master_orders_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-1.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO group.id = mysql_master_orders_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,594] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.whitelist = orders_shard_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.whitelist = orders_shard_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.server.id = 2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.server.id = 3 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO name = mysql-master-orders-2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO name = mysql-master-orders-1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,595] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:41:21,676] INFO Removed connector mysql-master-orders-1 due to null configuration. This is usually intentional and does not indicate an issue. (org.apache.kafka.connect.storage.KafkaConfigBackingStore)
- [2019-02-19 15:41:21,676] INFO Connector mysql-master-orders-1 config removed (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,677] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:41:21,677] INFO Stopping connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,677] INFO Stopping connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,677] INFO 10.56.73.73 - - [19/Feb/2019:15:40:18 +0000] "DELETE /connectors/mysql-master-orders-1 HTTP/1.1" 204 - 63222 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,677] INFO Stopping task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,677] INFO Stopping task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,679] INFO Stopped connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,679] INFO Stopped connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:41:21,720] INFO 10.56.152.1 - - [19/Feb/2019:15:40:31 +0000] "GET /connectors HTTP/1.1" 409 142 50456 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,721] INFO 10.56.152.1 - - [19/Feb/2019:15:41:01 +0000] "GET /connectors HTTP/1.1" 409 142 20457 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,722] INFO 10.56.152.1 - - [19/Feb/2019:15:40:41 +0000] "GET /connectors HTTP/1.1" 409 142 40458 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,723] INFO 10.56.152.1 - - [19/Feb/2019:15:40:51 +0000] "GET /connectors HTTP/1.1" 409 142 30458 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,723] INFO 10.56.152.1 - - [19/Feb/2019:15:41:21 +0000] "GET /connectors HTTP/1.1" 409 142 459 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,723] INFO 10.56.152.1 - - [19/Feb/2019:15:41:11 +0000] "GET /connectors HTTP/1.1" 409 142 10459 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,723] INFO 10.56.152.1 - - [19/Feb/2019:15:40:21 +0000] "GET /connectors HTTP/1.1" 409 142 60459 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:41:21,810] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-2-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-2-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:41:21,810] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-1-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-1-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:41:21,811] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-2-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:41:21,811] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-1-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:41:21,818] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-1-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:41:21,818] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-2-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:41:21,821] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,822] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,822] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:21,823] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:22,364] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-2-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,364] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-1-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,366] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,366] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:22,367] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:22,367] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:41:22,367] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:22,367] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:41:29,533] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Error reading MySQL variables: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:266)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:79)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:55)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:58)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:989)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:341)
- at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2251)
- at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2284)
- at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2083)
- at com.mysql.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:806)
- at com.mysql.jdbc.JDBC4Connection.<init>(JDBC4Connection.java:47)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:410)
- at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:328)
- at io.debezium.jdbc.JdbcConnection.lambda$patternBasedFactory$1(JdbcConnection.java:163)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:654)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:649)
- at io.debezium.jdbc.JdbcConnection.connect(JdbcConnection.java:276)
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:249)
- ... 12 more
- Caused by: java.net.ConnectException: Connection timed out
- at java.net.PlainSocketImpl.socketConnect(Native Method)
- at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
- at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
- at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
- at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
- at java.net.Socket.connect(Socket.java:589)
- at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:300)
- ... 29 more
- [2019-02-19 15:41:29,534] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:41:29,533] ERROR WorkerSourceTask{id=mysql-master-orders-0-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Error reading MySQL variables: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:266)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:79)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:55)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:58)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:989)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:341)
- at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2251)
- at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2284)
- at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2083)
- at com.mysql.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:806)
- at com.mysql.jdbc.JDBC4Connection.<init>(JDBC4Connection.java:47)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:410)
- at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:328)
- at io.debezium.jdbc.JdbcConnection.lambda$patternBasedFactory$1(JdbcConnection.java:163)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:654)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:649)
- at io.debezium.jdbc.JdbcConnection.connect(JdbcConnection.java:276)
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:249)
- ... 12 more
- Caused by: java.net.ConnectException: Connection timed out
- at java.net.PlainSocketImpl.socketConnect(Native Method)
- at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
- at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
- at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
- at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
- at java.net.Socket.connect(Socket.java:589)
- at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:300)
- ... 29 more
- [2019-02-19 15:41:29,533] ERROR WorkerSourceTask{id=mysql-master-orders-1-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Error reading MySQL variables: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:266)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:79)
- at io.debezium.connector.mysql.MySqlTaskContext.<init>(MySqlTaskContext.java:55)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:58)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
- The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:989)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:341)
- at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2251)
- at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2284)
- at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2083)
- at com.mysql.jdbc.ConnectionImpl.<init>(ConnectionImpl.java:806)
- at com.mysql.jdbc.JDBC4Connection.<init>(JDBC4Connection.java:47)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
- at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
- at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
- at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
- at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
- at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:410)
- at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:328)
- at io.debezium.jdbc.JdbcConnection.lambda$patternBasedFactory$1(JdbcConnection.java:163)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:654)
- at io.debezium.jdbc.JdbcConnection.connection(JdbcConnection.java:649)
- at io.debezium.jdbc.JdbcConnection.connect(JdbcConnection.java:276)
- at io.debezium.connector.mysql.MySqlJdbcContext.readMySqlSystemVariables(MySqlJdbcContext.java:249)
- ... 12 more
- Caused by: java.net.ConnectException: Connection timed out
- at java.net.PlainSocketImpl.socketConnect(Native Method)
- at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
- at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
- at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
- at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
- at java.net.Socket.connect(Socket.java:589)
- at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211)
- at com.mysql.jdbc.MysqlIO.<init>(MysqlIO.java:300)
- ... 29 more
- [2019-02-19 15:41:29,536] ERROR WorkerSourceTask{id=mysql-master-orders-1-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:41:29,535] ERROR WorkerSourceTask{id=mysql-master-orders-0-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:41:29,534] WARN Could not stop task (org.apache.kafka.connect.runtime.WorkerSourceTask)
- java.lang.NullPointerException
- at io.debezium.connector.mysql.MySqlConnectorTask.stop(MySqlConnectorTask.java:243)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.tryStop(WorkerSourceTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.close(WorkerSourceTask.java:134)
- at org.apache.kafka.connect.runtime.WorkerTask.doClose(WorkerTask.java:149)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:176)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:41:29,536] WARN Could not stop task (org.apache.kafka.connect.runtime.WorkerSourceTask)
- java.lang.NullPointerException
- at io.debezium.connector.mysql.MySqlConnectorTask.stop(MySqlConnectorTask.java:243)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.tryStop(WorkerSourceTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.close(WorkerSourceTask.java:134)
- at org.apache.kafka.connect.runtime.WorkerTask.doClose(WorkerTask.java:149)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:176)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:41:29,536] WARN Could not stop task (org.apache.kafka.connect.runtime.WorkerSourceTask)
- java.lang.NullPointerException
- at io.debezium.connector.mysql.MySqlConnectorTask.stop(MySqlConnectorTask.java:243)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.tryStop(WorkerSourceTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.close(WorkerSourceTask.java:134)
- at org.apache.kafka.connect.runtime.WorkerTask.doClose(WorkerTask.java:149)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:176)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:41:29,537] INFO [Producer clientId=producer-6] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:41:29,537] INFO [Producer clientId=producer-5] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:41:29,537] INFO [Producer clientId=producer-4] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:42:21,679] ERROR Graceful stop of task mysql-master-orders-1-0 failed. (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:21,679] ERROR Graceful stop of task mysql-master-orders-2-0 failed. (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:21,679] INFO Finished stopping tasks in preparation for rebalance (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:21,679] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:42:24,849] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 5 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:42:24,849] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-cadf63cf-2202-45da-99bb-1c91da3c580a', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=13, connectorIds=[mysql-master-orders-2], taskIds=[mysql-master-orders-2-0]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:24,850] INFO Starting connectors and tasks using config offset 13 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:24,850] INFO Starting connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:24,850] INFO Starting task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:24,850] INFO Creating task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:24,851] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:42:24,851] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:42:24,852] INFO Creating connector mysql-master-orders-2 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:24,851] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:42:24,852] INFO Instantiated connector mysql-master-orders-2 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:24,853] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:42:24,853] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:42:24,853] INFO Instantiated task mysql-master-orders-2-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:24,854] INFO Finished creating connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:42:24,855] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:42:24,855] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:42:24,855] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:42:24,855] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-2
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:42:24,855] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:42:24,855] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:42:24,863] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:24,864] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:24,864] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:42:24,866] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-2.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO group.id = mysql_master_orders_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.whitelist = orders_shard_2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.server.id = 3 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO name = mysql-master-orders-2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,867] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:42:24,868] INFO 10.56.152.1 - - [19/Feb/2019:15:42:21 +0000] "GET /connectors HTTP/1.1" 200 25 3604 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,868] INFO 10.56.152.1 - - [19/Feb/2019:15:41:41 +0000] "GET /connectors HTTP/1.1" 200 25 43604 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,868] INFO 10.56.152.1 - - [19/Feb/2019:15:42:01 +0000] "GET /connectors HTTP/1.1" 200 25 23550 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,868] INFO 10.56.152.1 - - [19/Feb/2019:15:41:31 +0000] "GET /connectors HTTP/1.1" 200 25 53604 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,869] INFO 10.56.152.1 - - [19/Feb/2019:15:42:11 +0000] "GET /connectors HTTP/1.1" 200 25 13605 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,871] INFO 10.56.152.1 - - [19/Feb/2019:15:41:51 +0000] "GET /connectors HTTP/1.1" 200 25 33607 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:24,881] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-2-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-2-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:42:24,882] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-2-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:42:24,882] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-2-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:42:24,884] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:24,884] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:24,885] WARN Error registering AppInfo mbean (org.apache.kafka.common.utils.AppInfoParser)
- javax.management.InstanceAlreadyExistsException: kafka.producer:type=app-info,id=mysql-master-orders-2-dbhistory
- at com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:437)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1898)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:966)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:900)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:324)
- at com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)
- at org.apache.kafka.common.utils.AppInfoParser.registerAppInfo(AppInfoParser.java:62)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:438)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:298)
- at io.debezium.relational.history.KafkaDatabaseHistory.start(KafkaDatabaseHistory.java:171)
- at io.debezium.connector.mysql.MySqlSchema.start(MySqlSchema.java:155)
- at io.debezium.connector.mysql.MySqlTaskContext.start(MySqlTaskContext.java:257)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:63)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:42:25,391] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-2-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:42:25,392] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:25,392] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:42:25,392] WARN Error registering AppInfo mbean (org.apache.kafka.common.utils.AppInfoParser)
- javax.management.InstanceAlreadyExistsException: kafka.admin.client:type=app-info,id=mysql-master-orders-2-dbhistory
- at com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:437)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1898)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:966)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:900)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:324)
- at com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)
- at org.apache.kafka.common.utils.AppInfoParser.registerAppInfo(AppInfoParser.java:62)
- at org.apache.kafka.clients.admin.KafkaAdminClient.<init>(KafkaAdminClient.java:378)
- at org.apache.kafka.clients.admin.KafkaAdminClient.createInternal(KafkaAdminClient.java:329)
- at org.apache.kafka.clients.admin.AdminClient.create(AdminClient.java:51)
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:336)
- at io.debezium.connector.mysql.MySqlSchema.intializeHistoryStorage(MySqlSchema.java:260)
- at io.debezium.connector.mysql.MySqlTaskContext.initializeHistoryStorage(MySqlTaskContext.java:194)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:127)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:42:31,268] INFO 10.56.152.1 - - [19/Feb/2019:15:42:31 +0000] "GET /connectors HTTP/1.1" 200 25 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:41,267] INFO 10.56.152.1 - - [19/Feb/2019:15:42:41 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:42:51,295] INFO 10.56.152.1 - - [19/Feb/2019:15:42:51 +0000] "GET /connectors HTTP/1.1" 200 25 31 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:01,309] INFO 10.56.152.1 - - [19/Feb/2019:15:43:01 +0000] "GET /connectors HTTP/1.1" 200 25 45 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:11,267] INFO 10.56.152.1 - - [19/Feb/2019:15:43:11 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:21,276] INFO 10.56.152.1 - - [19/Feb/2019:15:43:21 +0000] "GET /connectors HTTP/1.1" 200 25 12 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:22,429] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:43:22,430] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Creation of database history topic failed, please create the topic manually
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:349)
- at io.debezium.connector.mysql.MySqlSchema.intializeHistoryStorage(MySqlSchema.java:260)
- at io.debezium.connector.mysql.MySqlTaskContext.initializeHistoryStorage(MySqlTaskContext.java:194)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:127)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: java.util.concurrent.TimeoutException
- at org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:108)
- at org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:225)
- at io.debezium.relational.history.KafkaDatabaseHistory.getKafkaBrokerConfig(KafkaDatabaseHistory.java:354)
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:338)
- ... 12 more
- [2019-02-19 15:43:22,430] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:43:22,430] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:43:22,430] INFO [Producer clientId=producer-7] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:43:22,452] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:43:22,453] ERROR WorkerSourceTask{id=mysql-master-orders-1-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Creation of database history topic failed, please create the topic manually
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:349)
- at io.debezium.connector.mysql.MySqlSchema.intializeHistoryStorage(MySqlSchema.java:260)
- at io.debezium.connector.mysql.MySqlTaskContext.initializeHistoryStorage(MySqlTaskContext.java:194)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:127)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: java.util.concurrent.TimeoutException
- at org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:108)
- at org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:225)
- at io.debezium.relational.history.KafkaDatabaseHistory.getKafkaBrokerConfig(KafkaDatabaseHistory.java:354)
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:338)
- ... 12 more
- [2019-02-19 15:43:22,453] ERROR WorkerSourceTask{id=mysql-master-orders-1-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:43:22,453] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:43:22,453] INFO [Producer clientId=producer-8] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:43:31,272] INFO 10.56.152.1 - - [19/Feb/2019:15:43:31 +0000] "GET /connectors HTTP/1.1" 200 25 8 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:34,045] INFO Removed connector mysql-master-orders-2 due to null configuration. This is usually intentional and does not indicate an issue. (org.apache.kafka.connect.storage.KafkaConfigBackingStore)
- [2019-02-19 15:43:34,045] INFO Connector mysql-master-orders-2 config removed (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:43:34,045] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:43:34,045] INFO Stopping connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:43:34,045] INFO Stopping task mysql-master-orders-2-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:43:34,046] INFO 10.56.73.73 - - [19/Feb/2019:15:43:33 +0000] "DELETE /connectors/mysql-master-orders-2 HTTP/1.1" 204 - 115 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:43:34,046] INFO Stopped connector mysql-master-orders-2 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:25,396] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:44:25,396] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Creation of database history topic failed, please create the topic manually
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:349)
- at io.debezium.connector.mysql.MySqlSchema.intializeHistoryStorage(MySqlSchema.java:260)
- at io.debezium.connector.mysql.MySqlTaskContext.initializeHistoryStorage(MySqlTaskContext.java:194)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:127)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: java.util.concurrent.TimeoutException
- at org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:108)
- at org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:225)
- at io.debezium.relational.history.KafkaDatabaseHistory.getKafkaBrokerConfig(KafkaDatabaseHistory.java:354)
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:338)
- ... 12 more
- [2019-02-19 15:44:25,397] ERROR WorkerSourceTask{id=mysql-master-orders-2-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:44:25,397] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:44:25,397] INFO [Producer clientId=producer-9] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:44:25,442] INFO Finished stopping tasks in preparation for rebalance (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:25,442] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:44:25,527] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 6 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:44:25,527] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-cadf63cf-2202-45da-99bb-1c91da3c580a', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=15, connectorIds=[], taskIds=[]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:25,528] INFO Starting connectors and tasks using config offset 15 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:25,528] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:25,535] WARN The connection password is empty (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:44:25,542] INFO Successfully tested connection for jdbc:mysql://mysql-master-mms:3306/?useInformationSchema=true&nullCatalogMeansCurrent=false&useSSL=false&useUnicode=true&characterEncoding=UTF-8&characterSetResults=UTF-8&zeroDateTimeBehavior=convertToNull with user 'debezium' (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:44:25,590] INFO Connector mysql-master-orders-0 config updated (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,130] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,130] INFO Finished stopping tasks in preparation for rebalance (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,130] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:44:26,131] INFO 10.56.152.1 - - [19/Feb/2019:15:44:11 +0000] "GET /connectors HTTP/1.1" 409 142 14868 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,131] INFO 10.56.152.1 - - [19/Feb/2019:15:44:21 +0000] "GET /connectors HTTP/1.1" 409 142 4867 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,131] INFO 10.56.152.1 - - [19/Feb/2019:15:44:01 +0000] "GET /connectors HTTP/1.1" 409 142 24867 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,132] INFO 10.56.152.1 - - [19/Feb/2019:15:43:51 +0000] "GET /connectors HTTP/1.1" 409 142 34868 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,132] INFO 10.56.152.1 - - [19/Feb/2019:15:43:41 +0000] "GET /connectors HTTP/1.1" 409 142 44868 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,141] INFO 10.56.73.73 - - [19/Feb/2019:15:43:34 +0000] "PUT /connectors/mysql-master-orders-0/config HTTP/1.1" 201 1377 52091 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:26,214] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 7 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:44:26,214] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-cadf63cf-2202-45da-99bb-1c91da3c580a', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=16, connectorIds=[mysql-master-orders-0], taskIds=[mysql-master-orders-0-0]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,214] INFO Starting connectors and tasks using config offset 16 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,215] INFO Starting connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,215] INFO Starting task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,215] INFO Creating task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:26,215] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:44:26,215] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:44:26,215] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:44:26,215] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:44:26,215] INFO Instantiated task mysql-master-orders-0-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:26,216] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:44:26,216] INFO Creating connector mysql-master-orders-0 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:26,216] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:44:26,217] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:44:26,217] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:44:26,217] INFO Instantiated connector mysql-master-orders-0 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:26,217] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:44:26,217] INFO Finished creating connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:44:26,218] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:44:26,218] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:44:26,222] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:26,222] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:26,223] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:44:26,224] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-0.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO group.id = mysql_master_orders_0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.whitelist = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.server.id = 1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO name = mysql-master-orders-0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,224] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,225] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,225] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:44:26,236] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-0-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-0-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:44:26,236] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-0-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:44:26,237] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-0-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:44:26,238] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:26,238] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:26,415] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-0-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:44:26,416] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:26,416] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:44:31,267] INFO 10.56.152.1 - - [19/Feb/2019:15:44:31 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:41,267] INFO 10.56.152.1 - - [19/Feb/2019:15:44:41 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:44:51,267] INFO 10.56.152.1 - - [19/Feb/2019:15:44:51 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:01,267] INFO 10.56.152.1 - - [19/Feb/2019:15:45:01 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:11,267] INFO 10.56.152.1 - - [19/Feb/2019:15:45:11 +0000] "GET /connectors HTTP/1.1" 200 25 4 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:21,267] INFO 10.56.152.1 - - [19/Feb/2019:15:45:21 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:31,267] INFO 10.56.152.1 - - [19/Feb/2019:15:45:31 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:41,271] INFO 10.56.152.1 - - [19/Feb/2019:15:45:41 +0000] "GET /connectors HTTP/1.1" 200 25 7 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:45:51,267] INFO 10.56.152.1 - - [19/Feb/2019:15:45:51 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:01,267] INFO 10.56.152.1 - - [19/Feb/2019:15:46:01 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:11,266] INFO 10.56.152.1 - - [19/Feb/2019:15:46:11 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:21,267] INFO 10.56.152.1 - - [19/Feb/2019:15:46:21 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:26,468] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:46:26,469] ERROR WorkerSourceTask{id=mysql-master-orders-0-0} Task threw an uncaught and unrecoverable exception (org.apache.kafka.connect.runtime.WorkerTask)
- org.apache.kafka.connect.errors.ConnectException: Creation of database history topic failed, please create the topic manually
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:349)
- at io.debezium.connector.mysql.MySqlSchema.intializeHistoryStorage(MySqlSchema.java:260)
- at io.debezium.connector.mysql.MySqlTaskContext.initializeHistoryStorage(MySqlTaskContext.java:194)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:127)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- Caused by: java.util.concurrent.TimeoutException
- at org.apache.kafka.common.internals.KafkaFutureImpl$SingleWaiter.await(KafkaFutureImpl.java:108)
- at org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:225)
- at io.debezium.relational.history.KafkaDatabaseHistory.getKafkaBrokerConfig(KafkaDatabaseHistory.java:354)
- at io.debezium.relational.history.KafkaDatabaseHistory.initializeStorage(KafkaDatabaseHistory.java:338)
- ... 12 more
- [2019-02-19 15:46:26,469] ERROR WorkerSourceTask{id=mysql-master-orders-0-0} Task is being killed and will not recover until manually restarted (org.apache.kafka.connect.runtime.WorkerTask)
- [2019-02-19 15:46:26,469] INFO Stopping MySQL connector task (io.debezium.connector.mysql.MySqlConnectorTask)
- [2019-02-19 15:46:26,469] INFO [Producer clientId=producer-10] Closing the Kafka producer with timeoutMillis = 30000 ms. (org.apache.kafka.clients.producer.KafkaProducer)
- [2019-02-19 15:46:31,267] INFO 10.56.152.1 - - [19/Feb/2019:15:46:31 +0000] "GET /connectors HTTP/1.1" 200 25 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:38,257] WARN The connection password is empty (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:46:38,265] INFO Successfully tested connection for jdbc:mysql://mysql-master-mms:3306/?useInformationSchema=true&nullCatalogMeansCurrent=false&useSSL=false&useUnicode=true&characterEncoding=UTF-8&characterSetResults=UTF-8&zeroDateTimeBehavior=convertToNull with user 'debezium' (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:46:38,348] INFO Connector mysql-master-orders-1 config updated (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,349] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,349] INFO Stopping connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,349] INFO Stopping task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,350] INFO Stopped connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,351] INFO 10.56.73.73 - - [19/Feb/2019:15:46:38 +0000] "PUT /connectors/mysql-master-orders-1/config HTTP/1.1" 201 1385 99 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:38,392] INFO Finished stopping tasks in preparation for rebalance (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,392] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] (Re-)joining group (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:46:38,478] INFO [Worker clientId=connect-1, groupId=ci-connect-diff-115379-68670-0-source-group] Successfully joined group with generation 8 (org.apache.kafka.clients.consumer.internals.AbstractCoordinator)
- [2019-02-19 15:46:38,478] INFO Joined group and got assignment: Assignment{error=0, leader='connect-1-cadf63cf-2202-45da-99bb-1c91da3c580a', leaderUrl='http://kafka-connect-source-service-0:8083/', offset=17, connectorIds=[mysql-master-orders-0, mysql-master-orders-1], taskIds=[mysql-master-orders-0-0, mysql-master-orders-1-0]} (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO Starting connectors and tasks using config offset 17 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO Starting connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO Starting connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:46:38,478] INFO Starting task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO Starting task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,478] INFO Creating task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,478] INFO Creating task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,478] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:46:38,478] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,479] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:46:38,479] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,479] INFO Creating connector mysql-master-orders-1 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,479] INFO ConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig)
- [2019-02-19 15:46:38,480] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,479] INFO Creating connector mysql-master-orders-0 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,480] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:46:38,480] INFO Instantiated task mysql-master-orders-1-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,480] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,480] INFO TaskConfig values:
- task.class = class io.debezium.connector.mysql.MySqlConnectorTask
- (org.apache.kafka.connect.runtime.TaskConfig)
- [2019-02-19 15:46:38,480] INFO Instantiated connector mysql-master-orders-1 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,480] INFO Instantiated task mysql-master-orders-0-0 with version 0.8.3.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,480] INFO Instantiated connector mysql-master-orders-0 with version 0.8.3.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,481] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:46:38,481] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:46:38,481] INFO Finished creating connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,481] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:46:38,481] INFO AvroConverterConfig values:
- schema.registry.url = [http://schema-registry-service:8081]
- auto.register.schemas = true
- max.schemas.per.subject = 1000
- (io.confluent.connect.avro.AvroConverterConfig)
- [2019-02-19 15:46:38,481] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:46:38,481] INFO Finished creating connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:38,481] INFO AvroDataConfig values:
- schemas.cache.config = 1000
- enhanced.avro.schema.support = false
- connect.meta.data = true
- (io.confluent.connect.avro.AvroDataConfig)
- [2019-02-19 15:46:38,482] INFO SourceConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- value.converter = null
- (org.apache.kafka.connect.runtime.SourceConnectorConfig)
- [2019-02-19 15:46:38,482] INFO DebeziumShardedLogicalTableRouterConfig values:
- normalized.database.name = [orders]
- normalized.server.name = diff-115379-68670-0_mysql_master
- (com.clover.transformations.DebeziumShardedLogicalTableRouterConfig)
- [2019-02-19 15:46:38,482] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-1
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,482] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:46:38,482] INFO EnrichedConnectorConfig values:
- connector.class = io.debezium.connector.mysql.MySqlConnector
- key.converter = null
- name = mysql-master-orders-0
- tasks.max = 1
- transforms = [changetopic]
- transforms.changetopic.normalized.database.name = [orders]
- transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master
- transforms.changetopic.type = class com.clover.transformations.DebeziumShardedLogicalTableRouter
- value.converter = null
- (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig)
- [2019-02-19 15:46:38,482] INFO ProducerConfig values:
- acks = all
- batch.size = 16384
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 33554432
- client.id =
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- linger.ms = 0
- max.block.ms = 9223372036854775807
- max.in.flight.requests.per.connection = 1
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 20000
- retries = 2147483647
- retry.backoff.ms = 500
- sasl.jaas.config = [hidden]
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = PLAIN
- security.protocol = SASL_SSL
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = https
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:46:38,486] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,486] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,489] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,489] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,489] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-1.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO group.id = mysql_master_orders_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.whitelist = orders_shard_1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.server.id = 2 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO name = mysql-master-orders-1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,490] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,492] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO snapshot.locking.mode = minimal (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO transforms.changetopic.normalized.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO transforms.changetopic.normalized.database.name = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.history.kafka.topic = diff-115379-68670-0_mysql-master-orders-0.db (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO group.id = mysql_master_orders_0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO transforms = changetopic (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO snapshot.delay.ms = 2000 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.history.kafka.recovery.attempts = 100 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO include.schema.changes = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.history.kafka.recovery.poll.interval.ms = 500 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.history.skip.unparseable.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO table.ignore.builtin = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO binlog.buffer.size = 0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.whitelist = orders (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.user = debezium (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO transforms.changetopic.type = com.clover.transformations.DebeziumShardedLogicalTableRouter (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.server.id = 1 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.history.kafka.bootstrap.servers = pkc-e8xq4.us-central1.gcp.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO event.deserialization.failure.handling.mode = fail (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.server.name = diff-115379-68670-0_mysql_master (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,492] WARN The connection password is empty (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:46:38,493] INFO database.hostname = mysql-master-mms (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,493] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,493] INFO name = mysql-master-orders-0 (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,493] INFO database.history.store.only.monitored.tables.ddl = false (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,493] INFO connect.keep.alive = true (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,493] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask)
- [2019-02-19 15:46:38,498] INFO Successfully tested connection for jdbc:mysql://mysql-master-mms:3306/?useInformationSchema=true&nullCatalogMeansCurrent=false&useSSL=false&useUnicode=true&characterEncoding=UTF-8&characterSetResults=UTF-8&zeroDateTimeBehavior=convertToNull with user 'debezium' (io.debezium.connector.mysql.MySqlConnector)
- [2019-02-19 15:46:38,505] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-1-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-1-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:46:38,505] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-1-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:46:38,505] INFO KafkaDatabaseHistory Consumer config: {enable.auto.commit=false, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=mysql-master-orders-0-dbhistory, auto.offset.reset=earliest, session.timeout.ms=10000, bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, client.id=mysql-master-orders-0-dbhistory, key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, fetch.min.bytes=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:46:38,506] INFO KafkaDatabaseHistory Producer config: {bootstrap.servers=pkc-e8xq4.us-central1.gcp.confluent.cloud:9092, value.serializer=org.apache.kafka.common.serialization.StringSerializer, buffer.memory=1048576, retries=1, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=mysql-master-orders-0-dbhistory, linger.ms=0, batch.size=32768, max.block.ms=10000, acks=1} (io.debezium.relational.history.KafkaDatabaseHistory)
- [2019-02-19 15:46:38,506] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-1-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:46:38,507] INFO ProducerConfig values:
- acks = 1
- batch.size = 32768
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- buffer.memory = 1048576
- client.id = mysql-master-orders-0-dbhistory
- compression.type = none
- confluent.batch.expiry.ms = 30000
- connections.max.idle.ms = 540000
- enable.idempotence = false
- interceptor.classes = null
- key.serializer = class org.apache.kafka.common.serialization.StringSerializer
- linger.ms = 0
- max.block.ms = 10000
- max.in.flight.requests.per.connection = 5
- max.request.size = 1048576
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
- receive.buffer.bytes = 32768
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 30000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- transaction.timeout.ms = 60000
- transactional.id = null
- value.serializer = class org.apache.kafka.common.serialization.StringSerializer
- (org.apache.kafka.clients.producer.ProducerConfig)
- [2019-02-19 15:46:38,510] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,510] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,510] WARN Error registering AppInfo mbean (org.apache.kafka.common.utils.AppInfoParser)
- javax.management.InstanceAlreadyExistsException: kafka.producer:type=app-info,id=mysql-master-orders-0-dbhistory
- at com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:437)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1898)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:966)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:900)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:324)
- at com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)
- at org.apache.kafka.common.utils.AppInfoParser.registerAppInfo(AppInfoParser.java:62)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:438)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:298)
- at io.debezium.relational.history.KafkaDatabaseHistory.start(KafkaDatabaseHistory.java:171)
- at io.debezium.connector.mysql.MySqlSchema.start(MySqlSchema.java:155)
- at io.debezium.connector.mysql.MySqlTaskContext.start(MySqlTaskContext.java:257)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:63)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:46:38,510] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,511] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,511] WARN Error registering AppInfo mbean (org.apache.kafka.common.utils.AppInfoParser)
- javax.management.InstanceAlreadyExistsException: kafka.producer:type=app-info,id=mysql-master-orders-1-dbhistory
- at com.sun.jmx.mbeanserver.Repository.addMBean(Repository.java:437)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerWithRepository(DefaultMBeanServerInterceptor.java:1898)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerDynamicMBean(DefaultMBeanServerInterceptor.java:966)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerObject(DefaultMBeanServerInterceptor.java:900)
- at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.registerMBean(DefaultMBeanServerInterceptor.java:324)
- at com.sun.jmx.mbeanserver.JmxMBeanServer.registerMBean(JmxMBeanServer.java:522)
- at org.apache.kafka.common.utils.AppInfoParser.registerAppInfo(AppInfoParser.java:62)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:438)
- at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:298)
- at io.debezium.relational.history.KafkaDatabaseHistory.start(KafkaDatabaseHistory.java:171)
- at io.debezium.connector.mysql.MySqlSchema.start(MySqlSchema.java:155)
- at io.debezium.connector.mysql.MySqlTaskContext.start(MySqlTaskContext.java:257)
- at io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:63)
- at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:45)
- at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:182)
- at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:170)
- at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:214)
- at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
- at java.util.concurrent.FutureTask.run(FutureTask.java:266)
- at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
- at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
- at java.lang.Thread.run(Thread.java:745)
- [2019-02-19 15:46:38,541] INFO Connector mysql-master-orders-2 config updated (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:38,814] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-1-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,814] INFO AdminClientConfig values:
- bootstrap.servers = [pkc-e8xq4.us-central1.gcp.confluent.cloud:9092]
- client.id = mysql-master-orders-0-dbhistory
- connections.max.idle.ms = 300000
- metadata.max.age.ms = 300000
- metric.reporters = []
- metrics.num.samples = 2
- metrics.recording.level = INFO
- metrics.sample.window.ms = 30000
- receive.buffer.bytes = 65536
- reconnect.backoff.max.ms = 1000
- reconnect.backoff.ms = 50
- request.timeout.ms = 120000
- retries = 1
- retry.backoff.ms = 100
- sasl.jaas.config = null
- sasl.kerberos.kinit.cmd = /usr/bin/kinit
- sasl.kerberos.min.time.before.relogin = 60000
- sasl.kerberos.service.name = null
- sasl.kerberos.ticket.renew.jitter = 0.05
- sasl.kerberos.ticket.renew.window.factor = 0.8
- sasl.mechanism = GSSAPI
- security.protocol = PLAINTEXT
- send.buffer.bytes = 131072
- ssl.cipher.suites = null
- ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
- ssl.endpoint.identification.algorithm = null
- ssl.key.password = null
- ssl.keymanager.algorithm = SunX509
- ssl.keystore.location = null
- ssl.keystore.password = null
- ssl.keystore.type = JKS
- ssl.protocol = TLS
- ssl.provider = null
- ssl.secure.random.implementation = null
- ssl.trustmanager.algorithm = PKIX
- ssl.truststore.location = null
- ssl.truststore.password = null
- ssl.truststore.type = JKS
- (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,815] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,815] WARN The configuration 'value.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'batch.size' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,816] WARN The configuration 'max.block.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'acks' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'buffer.memory' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'key.serializer' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] WARN The configuration 'linger.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig)
- [2019-02-19 15:46:38,816] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,817] INFO Kafka version : 1.0.2-cp1 (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:38,817] INFO Kafka commitId : 129d4c93ba1371aa (org.apache.kafka.common.utils.AppInfoParser)
- [2019-02-19 15:46:39,081] INFO Rebalance started (org.apache.kafka.connect.runtime.distributed.DistributedHerder)
- [2019-02-19 15:46:39,081] INFO Stopping connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,081] INFO Stopping connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,082] INFO Stopping task mysql-master-orders-0-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,082] INFO Stopping task mysql-master-orders-1-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,678] INFO Stopped connector mysql-master-orders-0 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,678] INFO Stopped connector mysql-master-orders-1 (org.apache.kafka.connect.runtime.Worker)
- [2019-02-19 15:46:39,678] INFO 10.56.73.73 - - [19/Feb/2019:15:46:38 +0000] "PUT /connectors/mysql-master-orders-2/config HTTP/1.1" 201 1385 1323 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:39,755] INFO 10.56.73.73 - - [19/Feb/2019:15:46:39 +0000] "GET /connectors/mysql-master-orders-0/status HTTP/1.1" 200 220 72 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:41,759] INFO 10.56.73.73 - - [19/Feb/2019:15:46:41 +0000] "GET /connectors/mysql-master-orders-0/status HTTP/1.1" 200 220 3 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:45,794] INFO 10.56.73.73 - - [19/Feb/2019:15:46:45 +0000] "GET /connectors/mysql-master-orders-0/status HTTP/1.1" 200 220 25 (org.apache.kafka.connect.runtime.rest.RestServer)
- [2019-02-19 15:46:53,811] INFO 10.56.73.73 - - [19/Feb/2019:15:46:53 +0000] "GET /connectors/mysql-master-orders-0/status HTTP/1.1" 200 220 2 (org.apache.kafka.connect.runtime.rest.RestServer)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement