{"app":"vault-env","level":"info","msg":"received new Vault token","time":"2024-05-02T07:50:10Z"} {"app":"vault-env","level":"info","msg":"initial Vault token arrived","time":"2024-05-02T07:50:10Z"} {"app":"vault-env","level":"warning","msg":"secret/services/data-el/kafka_confluent/data-el: Endpoint ignored these unrecognized parameters: [version]","time":"2024-05-02T07:50:10Z"} {"app":"vault-env","level":"info","msg":"spawning process: [/etc/confluent/docker/run]","time":"2024-05-02T07:50:10Z"} ===> User uid=0(root) gid=0(root) groups=0(root) ===> Configuring ... ===> Running preflight checks ... ===> Check if Kafka is healthy ... Using log4j config /etc/cp-base-new/log4j.properties ===> Launching ... ===> Launching kafka-connect ... [2024-05-02 07:50:21,744] INFO Kafka Connect worker initializing ... (org.apache.kafka.connect.cli.AbstractConnectCli) [2024-05-02 07:50:21,747] INFO WorkerInfo values: jvm.args = -XX:+UseContainerSupport, -XX:MaxRAMPercentage=70.0, -XX:+UseG1GC, -XX:MaxGCPauseMillis=20, -XX:InitiatingHeapOccupancyPercent=35, -XX:+ExplicitGCInvokesConcurrent, -XX:MaxInlineLevel=15, -Djava.awt.headless=true, -Dcom.sun.management.jmxremote=true, -Dcom.sun.management.jmxremote.authenticate=false, -Dcom.sun.management.jmxremote.ssl=false, -Dcom.sun.management.jmxremote.local.only=false, -Dcom.sun.management.jmxremote.port=9853, -Dcom.sun.management.jmxremote.rmi.port=9853', -Djava.rmi.server.hostname=10.171.217.115, -Dcom.sun.management.jmxremote.local.only=false, -Dcom.sun.management.jmxremote.rmi.port=9853, -Dcom.sun.management.jmxremote.port=9853, -Dcom.sun.management.jmxremote.port=9853, -Dkafka.logs.dir=/var/log/kafka, -Dlog4j.configuration=file:/etc/kafka/connect-log4j.properties jvm.spec = Azul Systems, Inc., OpenJDK 64-Bit Server VM, 11.0.22, 11.0.22+7-LTS jvm.classpath = /etc/kafka-connect/jars/*:/usr/share/java/kafka/activation-1.1.1.jar:/usr/share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/share/java/kafka/argparse4j-0.7.0.jar:/usr/share/java/kafka/audience-annotations-0.12.0.jar:/usr/share/java/kafka/caffeine-2.9.3.jar:/usr/share/java/kafka/checker-qual-3.19.0.jar:/usr/share/java/kafka/commons-beanutils-1.9.4.jar:/usr/share/java/kafka/commons-cli-1.4.jar:/usr/share/java/kafka/commons-collections-3.2.2.jar:/usr/share/java/kafka/commons-digester-2.1.jar:/usr/share/java/kafka/commons-io-2.11.0.jar:/usr/share/java/kafka/commons-lang3-3.8.1.jar:/usr/share/java/kafka/commons-logging-1.2.jar:/usr/share/java/kafka/commons-validator-1.7.jar:/usr/share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/share/java/kafka/hk2-api-2.6.1.jar:/usr/share/java/kafka/hk2-locator-2.6.1.jar:/usr/share/java/kafka/hk2-utils-2.6.1.jar:/usr/share/java/kafka/jackson-annotations-2.13.5.jar:/usr/share/java/kafka/jackson-core-2.13.5.jar:/usr/share/java/kafka/jackson-databind-2.13.5.jar:/usr/share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/share/java/kafka/jakarta.inject-2.6.1.jar:/usr/share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/share/java/kafka/javassist-3.29.2-GA.jar:/usr/share/java/kafka/javax.activation-api-1.2.0.jar:/usr/share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/share/java/kafka/jaxb-api-2.3.1.jar:/usr/share/java/kafka/jersey-client-2.39.1.jar:/usr/share/java/kafka/jersey-common-2.39.1.jar:/usr/share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/share/java/kafka/jersey-hk2-2.39.1.jar:/usr/share/java/kafka/jersey-server-2.39.1.jar:/usr/share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/share/java/kafka/jline-3.25.1.jar:/usr/share/java/kafka/jopt-simple-5.0.4.jar:/usr/share/java/kafka/jose4j-0.9.4.jar:/usr/share/java/kafka/jsr305-3.0.2.jar:/usr/share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/share/java/kafka/kafka.jar:/usr/share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/share/java/kafka/lz4-java-1.8.0.jar:/usr/share/java/kafka/maven-artifact-3.8.8.jar:/usr/share/java/kafka/metrics-core-2.2.0.jar:/usr/share/java/kafka/metrics-core-4.1.12.1.jar:/usr/share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/share/java/kafka/netty-common-4.1.100.Final.jar:/usr/share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/share/java/kafka/paranamer-2.8.jar:/usr/share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/share/java/kafka/pcollections-4.0.1.jar:/usr/share/java/kafka/plexus-utils-3.3.1.jar:/usr/share/java/kafka/reflections-0.10.2.jar:/usr/share/java/kafka/reload4j-1.2.25.jar:/usr/share/java/kafka/rocksdbjni-7.9.2.jar:/usr/share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/share/java/kafka/scala-library-2.13.11.jar:/usr/share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/share/java/kafka/scala-reflect-2.13.11.jar:/usr/share/java/kafka/slf4j-api-1.7.36.jar:/usr/share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/share/java/kafka/snappy-java-1.1.10.5.jar:/usr/share/java/kafka/swagger-annotations-2.2.8.jar:/usr/share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/share/java/kafka/zookeeper-3.8.4.jar:/usr/share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/share/java/kafka/debezium-signalling-0.1.12.jar:/usr/share/java/confluent-common/build-tools-7.6.1.jar:/usr/share/java/confluent-common/common-config-7.6.1.jar:/usr/share/java/confluent-common/common-metrics-7.6.1.jar:/usr/share/java/confluent-common/common-utils-7.6.1.jar:/usr/share/java/confluent-common/slf4j-api-1.7.36.jar:/usr/share/java/kafka-serde-tools/JSONata4Java-2.4.5.jar:/usr/share/java/kafka-serde-tools/accessors-smart-2.4.9.jar:/usr/share/java/kafka-serde-tools/agrona-1.17.1.jar:/usr/share/java/kafka-serde-tools/annotations-13.0.jar:/usr/share/java/kafka-serde-tools/annotations-3.0.1.jar:/usr/share/java/kafka-serde-tools/antlr4-runtime-4.13.1.jar:/usr/share/java/kafka-serde-tools/asm-9.3.jar:/usr/share/java/kafka-serde-tools/auto-service-annotations-1.0.1.jar:/usr/share/java/kafka-serde-tools/auto-value-annotations-1.9.jar:/usr/share/java/kafka-serde-tools/avro-1.11.3.jar:/usr/share/java/kafka-serde-tools/aws-java-sdk-core-1.12.182.jar:/usr/share/java/kafka-serde-tools/aws-java-sdk-kms-1.12.182.jar:/usr/share/java/kafka-serde-tools/azure-core-1.46.0.jar:/usr/share/java/kafka-serde-tools/azure-core-http-netty-1.14.0.jar:/usr/share/java/kafka-serde-tools/azure-identity-1.11.2.jar:/usr/share/java/kafka-serde-tools/azure-json-1.1.0.jar:/usr/share/java/kafka-serde-tools/azure-security-keyvault-keys-4.6.1.jar:/usr/share/java/kafka-serde-tools/cel-core-0.3.12.jar:/usr/share/java/kafka-serde-tools/cel-generated-antlr-0.3.12.jar:/usr/share/java/kafka-serde-tools/cel-generated-pb-0.3.12.jar:/usr/share/java/kafka-serde-tools/cel-jackson-0.3.12.jar:/usr/share/java/kafka-serde-tools/cel-tools-0.3.12.jar:/usr/share/java/kafka-serde-tools/checker-qual-3.33.0.jar:/usr/share/java/kafka-serde-tools/classgraph-4.8.21.jar:/usr/share/java/kafka-serde-tools/commons-beanutils-1.9.4.jar:/usr/share/java/kafka-serde-tools/commons-codec-1.15.jar:/usr/share/java/kafka-serde-tools/commons-collections-3.2.2.jar:/usr/share/java/kafka-serde-tools/commons-compress-1.26.0.jar:/usr/share/java/kafka-serde-tools/commons-digester-2.1.jar:/usr/share/java/kafka-serde-tools/commons-io-2.15.1.jar:/usr/share/java/kafka-serde-tools/commons-lang3-3.12.0.jar:/usr/share/java/kafka-serde-tools/commons-logging-1.2.jar:/usr/share/java/kafka-serde-tools/commons-text-1.10.0.jar:/usr/share/java/kafka-serde-tools/commons-validator-1.7.jar:/usr/share/java/kafka-serde-tools/content-type-2.2.jar:/usr/share/java/kafka-serde-tools/dek-registry-client-7.6.1.jar:/usr/share/java/kafka-serde-tools/error_prone_annotations-2.18.0.jar:/usr/share/java/kafka-serde-tools/everit-json-schema-1.14.3.jar:/usr/share/java/kafka-serde-tools/failureaccess-1.0.1.jar:/usr/share/java/kafka-serde-tools/google-api-client-1.35.2.jar:/usr/share/java/kafka-serde-tools/google-api-services-cloudkms-v1-rev20221107-2.0.0.jar:/usr/share/java/kafka-serde-tools/google-auth-library-credentials-1.5.3.jar:/usr/share/java/kafka-serde-tools/google-auth-library-oauth2-http-1.5.3.jar:/usr/share/java/kafka-serde-tools/google-http-client-1.43.1.jar:/usr/share/java/kafka-serde-tools/google-http-client-apache-v2-1.42.0.jar:/usr/share/java/kafka-serde-tools/google-http-client-gson-1.43.1.jar:/usr/share/java/kafka-serde-tools/google-oauth-client-1.34.1.jar:/usr/share/java/kafka-serde-tools/grpc-context-1.27.2.jar:/usr/share/java/kafka-serde-tools/gson-2.9.0.jar:/usr/share/java/kafka-serde-tools/guava-32.0.1-jre.jar:/usr/share/java/kafka-serde-tools/handy-uri-templates-2.1.8.jar:/usr/share/java/kafka-serde-tools/httpclient-4.5.13.jar:/usr/share/java/kafka-serde-tools/httpcore-4.4.15.jar:/usr/share/java/kafka-serde-tools/ion-java-1.0.2.jar:/usr/share/java/kafka-serde-tools/j2objc-annotations-2.8.jar:/usr/share/java/kafka-serde-tools/jackson-annotations-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-core-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-databind-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-dataformat-cbor-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-dataformat-csv-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-dataformat-protobuf-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-dataformat-xml-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-dataformat-yaml-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-datatype-guava-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-datatype-jdk8-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-datatype-joda-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-datatype-jsr310-2.14.2.jar:/usr/share/java/kafka-serde-tools/jackson-datatype-protobuf-0.9.13.jar:/usr/share/java/kafka-serde-tools/jackson-module-parameter-names-2.14.2.jar:/usr/share/java/kafka-serde-tools/javapoet-1.13.0.jar:/usr/share/java/kafka-serde-tools/jcip-annotations-1.0-1.jar:/usr/share/java/kafka-serde-tools/jmespath-java-1.12.182.jar:/usr/share/java/kafka-serde-tools/jna-5.13.0.jar:/usr/share/java/kafka-serde-tools/jna-platform-5.6.0.jar:/usr/share/java/kafka-serde-tools/joda-time-2.10.8.jar:/usr/share/java/kafka-serde-tools/json-20231013.jar:/usr/share/java/kafka-serde-tools/json-sKema-0.14.0.jar:/usr/share/java/kafka-serde-tools/json-smart-2.4.10.jar:/usr/share/java/kafka-serde-tools/jsr305-3.0.2.jar:/usr/share/java/kafka-serde-tools/kafka-avro-serializer-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-connect-avro-converter-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-connect-avro-data-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-connect-json-schema-converter-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-connect-protobuf-converter-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-json-schema-provider-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-json-schema-serializer-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-json-serializer-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-protobuf-provider-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-protobuf-serializer-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-protobuf-types-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-converter-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-rules-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-aws-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-azure-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-gcp-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-hcvault-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-registry-client-encryption-tink-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-schema-serializer-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-streams-7.6.1-ccs.jar:/usr/share/java/kafka-serde-tools/kafka-streams-avro-serde-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-streams-json-schema-serde-7.6.1.jar:/usr/share/java/kafka-serde-tools/kafka-streams-protobuf-serde-7.6.1.jar:/usr/share/java/kafka-serde-tools/kotlin-reflect-1.9.22.jar:/usr/share/java/kafka-serde-tools/kotlin-script-runtime-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-scripting-common-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-scripting-compiler-embeddable-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-scripting-compiler-impl-embeddable-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-scripting-jvm-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-stdlib-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-stdlib-common-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-stdlib-jdk7-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlin-stdlib-jdk8-1.9.10.jar:/usr/share/java/kafka-serde-tools/kotlinpoet-jvm-1.16.0.jar:/usr/share/java/kafka-serde-tools/lang-tag-1.7.jar:/usr/share/java/kafka-serde-tools/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar:/usr/share/java/kafka-serde-tools/logredactor-1.0.12.jar:/usr/share/java/kafka-serde-tools/logredactor-metrics-1.0.12.jar:/usr/share/java/kafka-serde-tools/mbknor-jackson-jsonschema_2.13-1.0.39.jar:/usr/share/java/kafka-serde-tools/minimal-json-0.9.5.jar:/usr/share/java/kafka-serde-tools/msal4j-1.14.0.jar:/usr/share/java/kafka-serde-tools/msal4j-persistence-extension-1.2.0.jar:/usr/share/java/kafka-serde-tools/netty-buffer-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-codec-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-codec-dns-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-codec-http-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-codec-http2-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-codec-socks-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-common-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-handler-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-handler-proxy-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-resolver-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-resolver-dns-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-resolver-dns-classes-macos-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-resolver-dns-native-macos-4.1.100.Final-osx-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final-linux-aarch_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final-linux-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final-osx-aarch_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final-osx-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final-windows-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-boringssl-static-2.0.61.Final.jar:/usr/share/java/kafka-serde-tools/netty-tcnative-classes-2.0.61.Final.jar:/usr/share/java/kafka-serde-tools/netty-transport-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-transport-classes-kqueue-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/netty-transport-native-epoll-4.1.100.Final-linux-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-transport-native-kqueue-4.1.100.Final-osx-x86_64.jar:/usr/share/java/kafka-serde-tools/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/share/java/kafka-serde-tools/nimbus-jose-jwt-9.30.2.jar:/usr/share/java/kafka-serde-tools/oauth2-oidc-sdk-10.7.1.jar:/usr/share/java/kafka-serde-tools/okio-jvm-3.4.0.jar:/usr/share/java/kafka-serde-tools/opencensus-api-0.31.1.jar:/usr/share/java/kafka-serde-tools/opencensus-contrib-http-util-0.31.1.jar:/usr/share/java/kafka-serde-tools/picocli-4.7.5.jar:/usr/share/java/kafka-serde-tools/proto-google-common-protos-2.22.1.jar:/usr/share/java/kafka-serde-tools/protobuf-java-3.19.6.jar:/usr/share/java/kafka-serde-tools/protobuf-java-util-3.19.6.jar:/usr/share/java/kafka-serde-tools/protoparser-4.0.3.jar:/usr/share/java/kafka-serde-tools/re2j-1.6.jar:/usr/share/java/kafka-serde-tools/reactive-streams-1.0.4.jar:/usr/share/java/kafka-serde-tools/reactor-core-3.4.34.jar:/usr/share/java/kafka-serde-tools/reactor-netty-core-1.0.40.jar:/usr/share/java/kafka-serde-tools/reactor-netty-http-1.0.40.jar:/usr/share/java/kafka-serde-tools/rocksdbjni-7.9.2.jar:/usr/share/java/kafka-serde-tools/scala-library-2.13.10.jar:/usr/share/java/kafka-serde-tools/slf4j-api-1.7.36.jar:/usr/share/java/kafka-serde-tools/snakeyaml-2.0.jar:/usr/share/java/kafka-serde-tools/stax2-api-4.2.1.jar:/usr/share/java/kafka-serde-tools/swagger-annotations-2.1.10.jar:/usr/share/java/kafka-serde-tools/tink-1.12.0.jar:/usr/share/java/kafka-serde-tools/tink-awskms-1.9.1.jar:/usr/share/java/kafka-serde-tools/tink-gcpkms-1.9.0.jar:/usr/share/java/kafka-serde-tools/validation-api-2.0.1.Final.jar:/usr/share/java/kafka-serde-tools/vault-java-driver-5.4.0.jar:/usr/share/java/kafka-serde-tools/wire-runtime-jvm-4.9.7.jar:/usr/share/java/kafka-serde-tools/wire-schema-jvm-4.9.7.jar:/usr/share/java/kafka-serde-tools/woodstox-core-6.5.1.jar:/usr/share/java/monitoring-interceptors/monitoring-interceptors-7.6.1.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/checker-qual-3.19.0.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/commons-io-2.11.0.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.8.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/scala-library-2.13.11.jar:/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.11.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/bin/../share/java/kafka/debezium-signalling-0.1.12.jar:/usr/bin/../share/java/confluent-telemetry/confluent-metrics-7.6.1-ce.jar os.spec = Linux, amd64, 5.10.205-195.807.amzn2.x86_64 os.vcpus = 1 (org.apache.kafka.connect.runtime.WorkerInfo) [2024-05-02 07:50:21,747] INFO Scanning for plugin classes. This might take a moment ... (org.apache.kafka.connect.cli.AbstractConnectCli) [2024-05-02 07:50:21,934] INFO Loading plugin from: /usr/share/java/cp-base-new (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:22,450] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/cp-base-new/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:22,451] INFO Loading plugin from: /usr/share/java/kafka (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:22,741] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/kafka/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:22,741] INFO Loading plugin from: /usr/share/java/acl (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,060] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/acl/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,061] INFO Loading plugin from: /usr/share/java/confluent-common (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,123] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-common/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,123] INFO Loading plugin from: /usr/share/java/confluent-control-center (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,451] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-control-center/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,451] INFO Loading plugin from: /usr/share/java/confluent-hub-client (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,519] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-hub-client/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,519] INFO Loading plugin from: /usr/share/java/confluent-telemetry (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,526] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-telemetry/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,526] INFO Loading plugin from: /usr/share/java/kafka-serde-tools (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,550] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/kafka-serde-tools/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,550] INFO Loading plugin from: /usr/share/java/monitoring-interceptors (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,555] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/monitoring-interceptors/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,555] INFO Loading plugin from: /usr/share/java/rest-utils (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,636] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/rest-utils/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,636] INFO Loading plugin from: /usr/share/java/schema-registry (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,661] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/schema-registry/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,661] INFO Loading plugin from: /usr/share/confluent-hub-components/debezium-signalling-0.1.12.jar (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,721] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/debezium-signalling-0.1.12.jar} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,721] INFO Loading plugin from: /usr/share/confluent-hub-components/debezium-debezium-connector-mysql (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,728] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,733] INFO Loading plugin from: /usr/share/confluent-hub-components/data-el-smt-0.1.13.jar (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,736] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/data-el-smt-0.1.13.jar} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,736] INFO Loading plugin from: /usr/share/confluent-hub-components/confluentinc-kafka-connect-avro-converter (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,742] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/confluentinc-kafka-connect-avro-converter/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,742] INFO Loading plugin from: classpath (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,747] INFO Registered loader: jdk.internal.loader.ClassLoaders$AppClassLoader@2c13da15 (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,747] INFO Scanning plugins with ServiceLoaderScanner took 1814 ms (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:23,749] INFO Loading plugin from: /usr/share/java/cp-base-new (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:26,526] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/cp-base-new/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:26,526] INFO Loading plugin from: /usr/share/java/kafka (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:28,833] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/kafka/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:28,834] INFO Loading plugin from: /usr/share/java/acl (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:34,622] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/acl/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:34,622] INFO Loading plugin from: /usr/share/java/confluent-common (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:34,627] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-common/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:34,627] INFO Loading plugin from: /usr/share/java/confluent-control-center (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:41,821] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-control-center/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:41,821] INFO Loading plugin from: /usr/share/java/confluent-hub-client (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:42,325] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-hub-client/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:42,326] INFO Loading plugin from: /usr/share/java/confluent-telemetry (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:43,054] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/confluent-telemetry/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:43,054] INFO Loading plugin from: /usr/share/java/kafka-serde-tools (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,063] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/kafka-serde-tools/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,064] INFO Loading plugin from: /usr/share/java/monitoring-interceptors (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,264] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/monitoring-interceptors/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,264] INFO Loading plugin from: /usr/share/java/rest-utils (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,988] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/rest-utils/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:45,989] INFO Loading plugin from: /usr/share/java/schema-registry (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,755] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/java/schema-registry/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,755] INFO Loading plugin from: /usr/share/confluent-hub-components/debezium-signalling-0.1.12.jar (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,757] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/debezium-signalling-0.1.12.jar} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,758] INFO Loading plugin from: /usr/share/confluent-hub-components/debezium-debezium-connector-mysql (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,954] ERROR Failed to discover Converter in /usr/share/confluent-hub-components/debezium-debezium-connector-mysql: Unable to instantiate CloudEventsConverter: Failed to statically initialize plugin class (org.apache.kafka.connect.runtime.isolation.ReflectionScanner) java.lang.ExceptionInInitializerError at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.apache.kafka.connect.runtime.isolation.ReflectionScanner.versionFor(ReflectionScanner.java:73) at org.apache.kafka.connect.runtime.isolation.ReflectionScanner.getPluginDesc(ReflectionScanner.java:136) at org.apache.kafka.connect.runtime.isolation.ReflectionScanner.scanPlugins(ReflectionScanner.java:89) at org.apache.kafka.connect.runtime.isolation.PluginScanner.scanUrlsAndAddPlugins(PluginScanner.java:79) at org.apache.kafka.connect.runtime.isolation.PluginScanner.discoverPlugins(PluginScanner.java:67) at org.apache.kafka.connect.runtime.isolation.Plugins.initLoaders(Plugins.java:90) at org.apache.kafka.connect.runtime.isolation.Plugins.(Plugins.java:74) at org.apache.kafka.connect.runtime.isolation.Plugins.(Plugins.java:64) at org.apache.kafka.connect.cli.AbstractConnectCli.startConnect(AbstractConnectCli.java:121) at org.apache.kafka.connect.cli.AbstractConnectCli.run(AbstractConnectCli.java:94) at org.apache.kafka.connect.cli.ConnectDistributed.main(ConnectDistributed.java:116) Caused by: org.apache.kafka.connect.errors.DataException at io.debezium.converters.CloudEventsConverter.(CloudEventsConverter.java:111) ... 15 more [2024-05-02 07:50:47,970] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,970] INFO Loading plugin from: /usr/share/confluent-hub-components/data-el-smt-0.1.13.jar (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,972] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/data-el-smt-0.1.13.jar} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:47,972] INFO Loading plugin from: /usr/share/confluent-hub-components/confluentinc-kafka-connect-avro-converter (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:48,222] INFO Registered loader: PluginClassLoader{pluginLocation=file:/usr/share/confluent-hub-components/confluentinc-kafka-connect-avro-converter/} (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:48,222] INFO Loading plugin from: classpath (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:53,759] INFO Registered loader: jdk.internal.loader.ClassLoaders$AppClassLoader@2c13da15 (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:53,760] INFO Scanning plugins with ReflectionScanner took 30011 ms (org.apache.kafka.connect.runtime.isolation.PluginScanner) [2024-05-02 07:50:53,767] WARN One or more plugins are missing ServiceLoader manifests may not be usable with plugin.discovery=service_load: [ file:/usr/share/confluent-hub-components/confluentinc-kafka-connect-avro-converter/ io.confluent.connect.avro.AvroConverter converter undefined file:/usr/share/java/acl/ io.confluent.connect.avro.AvroConverter converter undefined file:/usr/share/java/confluent-control-center/ io.confluent.connect.avro.AvroConverter converter undefined file:/usr/share/java/kafka-serde-tools/ io.confluent.connect.avro.AvroConverter converter undefined classpath io.confluent.connect.avro.AvroConverter converter undefined file:/usr/share/java/acl/ io.confluent.connect.json.JsonSchemaConverter converter undefined file:/usr/share/java/confluent-control-center/ io.confluent.connect.json.JsonSchemaConverter converter undefined file:/usr/share/java/kafka-serde-tools/ io.confluent.connect.json.JsonSchemaConverter converter undefined classpath io.confluent.connect.json.JsonSchemaConverter converter undefined file:/usr/share/java/acl/ io.confluent.connect.protobuf.ProtobufConverter converter undefined file:/usr/share/java/confluent-control-center/ io.confluent.connect.protobuf.ProtobufConverter converter undefined file:/usr/share/java/kafka-serde-tools/ io.confluent.connect.protobuf.ProtobufConverter converter undefined classpath io.confluent.connect.protobuf.ProtobufConverter converter undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.connector.mysql.MySqlConnector source 2.2.1.Final file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.connector.mysql.transforms.ReadToInsertEvent transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.converters.BinaryDataConverter converter undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.converters.BinaryDataConverter header_converter undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.converters.ByteArrayConverter converter undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.converters.ByteArrayConverter header_converter undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.ByLogicalTableRouter transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.ExtractChangedRecordState transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.ExtractNewRecordState transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.HeaderToValue transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.outbox.EventRouter transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.partitions.ComputePartition transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.partitions.PartitionRouting transformation undefined file:/usr/share/confluent-hub-components/debezium-debezium-connector-mysql/ io.debezium.transforms.tracing.ActivateTracingSpan transformation undefined file:/usr/share/confluent-hub-components/data-el-smt-0.1.13.jar io.skai.dataelsmt.DataELTransformer transformation undefined ] Read the documentation at https://kafka.apache.org/documentation.html#connect_plugindiscovery for instructions on migrating your plugins to take advantage of the performance improvements of service_load mode. To silence this warning, set plugin.discovery=only_scan in the worker config. (org.apache.kafka.connect.runtime.isolation.Plugins) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.Filter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.mirror.MirrorSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.DropHeaders' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertHeader' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.confluent.connect.json.JsonSchemaConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.converters.BinaryDataConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.common.config.provider.DirectoryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.mirror.MirrorCheckpointConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.HeaderFrom$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.transforms.ExtractNewRecordState' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.skai.dataelsmt.DataELTransformer' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.transforms.partitions.PartitionRouting' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.predicates.TopicNameMatches' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.transforms.outbox.EventRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.predicates.RecordIsTombstone' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.transforms.partitions.ComputePartition' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.connector.mysql.MySqlConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.confluent.connect.avro.AvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.confluent.connect.security.ConnectSecurityExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.Cast$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.Flatten$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.rest.basic.auth.extension.BasicAuthSecurityRestExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,769] INFO Added plugin 'io.debezium.transforms.HeaderToValue' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'io.confluent.kafka.secretregistry.client.config.provider.SecretConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'io.debezium.transforms.ByLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.common.config.provider.FileConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'io.confluent.kafka.schemaregistry.client.config.provider.SchemaRegistryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.HeaderFrom$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'io.confluent.connect.protobuf.ProtobufConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.predicates.HasHeaderKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Value' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,770] INFO Added plugin 'org.apache.kafka.common.config.provider.EnvVarConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'io.debezium.transforms.ExtractChangedRecordState' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'io.debezium.transforms.tracing.ActivateTracingSpan' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.transforms.MaskField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'io.debezium.connector.mysql.transforms.ReadToInsertEvent' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.transforms.InsertField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.mirror.MirrorHeartbeatConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.transforms.HoistField$Key' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,819] INFO Added plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'JsonSchema' to plugin 'io.confluent.connect.json.JsonSchemaConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'RecordIsTombstone' to plugin 'org.apache.kafka.connect.transforms.predicates.RecordIsTombstone' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'MySql' to plugin 'io.debezium.connector.mysql.MySqlConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'EnvVar' to plugin 'org.apache.kafka.common.config.provider.EnvVarConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'EnvVarConfigProvider' to plugin 'org.apache.kafka.common.config.provider.EnvVarConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'MirrorCheckpointConnector' to plugin 'org.apache.kafka.connect.mirror.MirrorCheckpointConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'MySqlConnector' to plugin 'io.debezium.connector.mysql.MySqlConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,821] INFO Added alias 'AvroConverter' to plugin 'io.confluent.connect.avro.AvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'HeaderToValue' to plugin 'io.debezium.transforms.HeaderToValue' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'PartitionRouting' to plugin 'io.debezium.transforms.partitions.PartitionRouting' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'SchemaRegistryConfigProvider' to plugin 'io.confluent.kafka.schemaregistry.client.config.provider.SchemaRegistryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'StringConverter' to plugin 'org.apache.kafka.connect.storage.StringConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'IntegerConverter' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Float' to plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'LongConverter' to plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'DropHeaders' to plugin 'org.apache.kafka.connect.transforms.DropHeaders' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'SimpleHeaderConverter' to plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'BinaryData' to plugin 'io.debezium.converters.BinaryDataConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'DirectoryConfigProvider' to plugin 'org.apache.kafka.common.config.provider.DirectoryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ReadToInsertEvent' to plugin 'io.debezium.connector.mysql.transforms.ReadToInsertEvent' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ShortConverter' to plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'DataELTransformer' to plugin 'io.skai.dataelsmt.DataELTransformer' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'BasicAuthSecurityRestExtension' to plugin 'org.apache.kafka.connect.rest.basic.auth.extension.BasicAuthSecurityRestExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Avro' to plugin 'io.confluent.connect.avro.AvroConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Simple' to plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Protobuf' to plugin 'io.confluent.connect.protobuf.ProtobufConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'AllConnectorClientConfigOverridePolicy' to plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ExtractNewRecordState' to plugin 'io.debezium.transforms.ExtractNewRecordState' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'MirrorSource' to plugin 'org.apache.kafka.connect.mirror.MirrorSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Directory' to plugin 'org.apache.kafka.common.config.provider.DirectoryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'MirrorHeartbeat' to plugin 'org.apache.kafka.connect.mirror.MirrorHeartbeatConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'HasHeaderKey' to plugin 'org.apache.kafka.connect.transforms.predicates.HasHeaderKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ConnectSecurityExtension' to plugin 'io.confluent.connect.security.ConnectSecurityExtension' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'MirrorCheckpoint' to plugin 'org.apache.kafka.connect.mirror.MirrorCheckpointConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ExtractChangedRecordState' to plugin 'io.debezium.transforms.ExtractChangedRecordState' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'None' to plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'TimestampRouter' to plugin 'org.apache.kafka.connect.transforms.TimestampRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Principal' to plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'All' to plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'JsonConverter' to plugin 'org.apache.kafka.connect.json.JsonConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'RegexRouter' to plugin 'org.apache.kafka.connect.transforms.RegexRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ComputePartition' to plugin 'io.debezium.transforms.partitions.ComputePartition' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'NoneConnectorClientConfigOverridePolicy' to plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Short' to plugin 'org.apache.kafka.connect.converters.ShortConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ByLogicalTableRouter' to plugin 'io.debezium.transforms.ByLogicalTableRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Double' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'FileConfigProvider' to plugin 'org.apache.kafka.common.config.provider.FileConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Secret' to plugin 'io.confluent.kafka.secretregistry.client.config.provider.SecretConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'EventRouter' to plugin 'io.debezium.transforms.outbox.EventRouter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'Long' to plugin 'org.apache.kafka.connect.converters.LongConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'File' to plugin 'org.apache.kafka.common.config.provider.FileConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'FloatConverter' to plugin 'org.apache.kafka.connect.converters.FloatConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'ActivateTracingSpan' to plugin 'io.debezium.transforms.tracing.ActivateTracingSpan' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'DoubleConverter' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,822] INFO Added alias 'BinaryDataConverter' to plugin 'io.debezium.converters.BinaryDataConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'TopicNameMatches' to plugin 'org.apache.kafka.connect.transforms.predicates.TopicNameMatches' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'ProtobufConverter' to plugin 'io.confluent.connect.protobuf.ProtobufConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'MirrorHeartbeatConnector' to plugin 'org.apache.kafka.connect.mirror.MirrorHeartbeatConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'InsertHeader' to plugin 'org.apache.kafka.connect.transforms.InsertHeader' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'MirrorSourceConnector' to plugin 'org.apache.kafka.connect.mirror.MirrorSourceConnector' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'PrincipalConnectorClientConfigOverridePolicy' to plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'SecretConfigProvider' to plugin 'io.confluent.kafka.secretregistry.client.config.provider.SecretConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'ValueToKey' to plugin 'org.apache.kafka.connect.transforms.ValueToKey' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'Integer' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'JsonSchemaConverter' to plugin 'io.confluent.connect.json.JsonSchemaConverter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'Filter' to plugin 'org.apache.kafka.connect.transforms.Filter' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,823] INFO Added alias 'SchemaRegistry' to plugin 'io.confluent.kafka.schemaregistry.client.config.provider.SchemaRegistryConfigProvider' (org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader) [2024-05-02 07:50:53,853] INFO DistributedConfig values: access.control.allow.methods = access.control.allow.origin = admin.listeners = null auto.include.jmx.reporter = true bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] client.dns.lookup = use_all_dns_ips client.id = config.providers = [] config.storage.replication.factor = 3 config.storage.topic = eks-kafka-connect-prod-source-ks6008-config connect.protocol = sessioned connections.max.idle.ms = 540000 connector.client.config.override.policy = All exactly.once.source.support = disabled group.id = eks-kafka-connect-prod-source-ks6008 header.converter = class org.apache.kafka.connect.storage.SimpleHeaderConverter heartbeat.interval.ms = 3000 inter.worker.key.generation.algorithm = HmacSHA256 inter.worker.key.size = null inter.worker.key.ttl.ms = 3600000 inter.worker.signature.algorithm = HmacSHA256 inter.worker.verification.algorithms = [HmacSHA256] key.converter = class io.confluent.connect.avro.AvroConverter listeners = [http://:8083] metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 offset.flush.interval.ms = 60000 offset.flush.timeout.ms = 5000 offset.storage.partitions = 1 offset.storage.replication.factor = 3 offset.storage.topic = eks-kafka-connect-prod-source-ks6008-offset plugin.discovery = hybrid_warn plugin.path = [/usr/share/java/, /usr/share/confluent-hub-components/] rebalance.timeout.ms = 60000 receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 40000 response.http.headers.config = rest.advertised.host.name = 10.171.217.115 rest.advertised.listener = null rest.advertised.port = null rest.extension.classes = [] retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null scheduled.rebalance.max.delay.ms = 300000 security.protocol = SASL_SSL send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.client.auth = none ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS status.storage.partitions = 1 status.storage.replication.factor = 3 status.storage.topic = eks-kafka-connect-prod-source-ks6008-status task.shutdown.graceful.timeout.ms = 5000 topic.creation.enable = true topic.tracking.allow.reset = true topic.tracking.enable = true value.converter = class io.confluent.connect.avro.AvroConverter worker.sync.timeout.ms = 3000 worker.unsync.backoff.ms = 300000 (org.apache.kafka.connect.runtime.distributed.DistributedConfig) [2024-05-02 07:50:53,854] INFO Creating Kafka admin client (org.apache.kafka.connect.runtime.WorkerConfig) [2024-05-02 07:50:53,856] INFO AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] client.dns.lookup = use_all_dns_ips client.id = connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:50:54,022] INFO Successfully logged in. (org.apache.kafka.common.security.authenticator.AbstractLogin) [2024-05-02 07:50:54,142] INFO These configurations '[producer.sasl.jaas.config, group.id, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:50:54,142] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:54,142] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:54,142] INFO Kafka startTimeMs: 1714636254142 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,229] INFO [AdminClient clientId=adminclient-1] Node 3 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,229] WARN [AdminClient clientId=adminclient-1] Connection to node 3 (e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,258] INFO Kafka cluster ID: lkc-z6kvj7 (org.apache.kafka.connect.runtime.WorkerConfig) [2024-05-02 07:50:55,259] INFO App info kafka.admin.client for adminclient-1 unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,259] INFO [AdminClient clientId=adminclient-1] Metadata update failed (org.apache.kafka.clients.admin.internals.AdminMetadataManager) org.apache.kafka.common.errors.TimeoutException: The AdminClient thread has exited. Call: fetchMetadata [2024-05-02 07:50:55,259] INFO [AdminClient clientId=adminclient-1] Timed out 1 remaining operation(s) during close. (org.apache.kafka.clients.admin.KafkaAdminClient) [2024-05-02 07:50:55,264] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:50:55,264] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:50:55,264] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:50:55,321] INFO PublicConfig values: access.control.allow.methods = access.control.allow.origin = admin.listeners = null listeners = [http://:8083] response.http.headers.config = rest.advertised.host.name = 10.171.217.115 rest.advertised.listener = null rest.advertised.port = null rest.extension.classes = [] ssl.cipher.suites = null ssl.client.auth = none ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS topic.tracking.allow.reset = true topic.tracking.enable = true (org.apache.kafka.connect.runtime.rest.RestServerConfig$PublicConfig) [2024-05-02 07:50:55,327] INFO Logging initialized @34649ms to org.eclipse.jetty.util.log.Slf4jLog (org.eclipse.jetty.util.log) [2024-05-02 07:50:55,351] INFO Added connector for http://:8083 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,351] INFO Initializing REST server (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,366] INFO jetty-9.4.54.v20240208; built: 2024-02-08T19:42:39.027Z; git: cef3fbd6d736a21e7d541a5db490381d95a2047d; jvm 11.0.22+7-LTS (org.eclipse.jetty.server.Server) [2024-05-02 07:50:55,424] INFO Started http_8083@40f82d53{HTTP/1.1, (http/1.1)}{0.0.0.0:8083} (org.eclipse.jetty.server.AbstractConnector) [2024-05-02 07:50:55,424] INFO Started @34746ms (org.eclipse.jetty.server.Server) [2024-05-02 07:50:55,444] INFO Advertised URI: http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,444] INFO REST server listening at http://10.171.217.115:8083/, advertising URL http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,444] INFO Advertised URI: http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,444] INFO REST admin endpoints at http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,444] INFO Advertised URI: http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,445] INFO Setting up All Policy for ConnectorClientConfigOverride. This will allow all client configurations to be overridden (org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy) [2024-05-02 07:50:55,449] INFO JsonConverterConfig values: converter.type = key decimal.format = BASE64 replace.null.with.default = true schemas.cache.size = 1000 schemas.enable = false (org.apache.kafka.connect.json.JsonConverterConfig) [2024-05-02 07:50:55,463] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,463] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,463] INFO Kafka startTimeMs: 1714636255463 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,519] INFO JsonConverterConfig values: converter.type = key decimal.format = BASE64 replace.null.with.default = true schemas.cache.size = 1000 schemas.enable = false (org.apache.kafka.connect.json.JsonConverterConfig) [2024-05-02 07:50:55,520] INFO JsonConverterConfig values: converter.type = value decimal.format = BASE64 replace.null.with.default = true schemas.cache.size = 1000 schemas.enable = false (org.apache.kafka.connect.json.JsonConverterConfig) [2024-05-02 07:50:55,527] WARN Ignoring 'config.storage.partitions=1' setting, since config topic partitions is always 1 (org.apache.kafka.connect.runtime.distributed.DistributedConfig) [2024-05-02 07:50:55,533] INFO Advertised URI: http://10.171.217.115:8083/ (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,539] INFO Successfully logged in. (org.apache.kafka.common.security.authenticator.AbstractLogin) [2024-05-02 07:50:55,554] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,554] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,554] INFO Kafka startTimeMs: 1714636255554 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,556] INFO Kafka Connect worker initialization took 33811ms (org.apache.kafka.connect.cli.AbstractConnectCli) [2024-05-02 07:50:55,556] INFO Kafka Connect starting (org.apache.kafka.connect.runtime.Connect) [2024-05-02 07:50:55,557] INFO Initializing REST resources (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,557] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Herder starting (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:50:55,558] INFO Worker starting (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:50:55,558] INFO Starting KafkaOffsetBackingStore (org.apache.kafka.connect.storage.KafkaOffsetBackingStore) [2024-05-02 07:50:55,558] INFO Starting KafkaBasedLog with topic eks-kafka-connect-prod-source-ks6008-offset (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:55,558] INFO AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-shared-admin connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:50:55,624] INFO These configurations '[producer.sasl.jaas.config, group.id, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:50:55,624] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,624] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,624] INFO Kafka startTimeMs: 1714636255624 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:55,645] INFO Adding admin resources to main listener (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:55,734] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,734] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 1 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,743] INFO DefaultSessionIdManager workerName=node0 (org.eclipse.jetty.server.session) [2024-05-02 07:50:55,743] INFO No SessionScavenger set, using defaults (org.eclipse.jetty.server.session) [2024-05-02 07:50:55,744] INFO node0 Scavenging every 600000ms (org.eclipse.jetty.server.session) [2024-05-02 07:50:55,825] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,827] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 1 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,844] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 2 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:55,844] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 2 (e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:56,047] INFO ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-offsets compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:56,136] INFO These configurations '[producer.sasl.jaas.config, group.id, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:56,136] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,136] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,136] INFO Kafka startTimeMs: 1714636256136 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,227] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-offsets client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = eks-kafka-connect-prod-source-ks6008 group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:56,352] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-offsets] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:56,436] INFO These configurations '[producer.sasl.jaas.config, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:56,436] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,436] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,437] INFO Kafka startTimeMs: 1714636256436 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:56,544] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:56,547] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Assigned to partition(s): eks-kafka-connect-prod-source-ks6008-offset-0 (org.apache.kafka.clients.consumer.KafkaConsumer) [2024-05-02 07:50:56,550] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Seeking to earliest offset of partition eks-kafka-connect-prod-source-ks6008-offset-0 (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:56,649] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:56,650] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 0 (e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:56,825] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:56,825] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 1 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,029] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,029] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 0 (e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,140] INFO Started o.e.j.s.ServletContextHandler@50b336cb{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler) [2024-05-02 07:50:57,140] INFO REST resources initialized; server is started and ready to handle requests (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:50:57,140] INFO Kafka Connect started (org.apache.kafka.connect.runtime.Connect) [2024-05-02 07:50:57,221] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-offsets, groupId=eks-kafka-connect-prod-source-ks6008] Resetting offset for partition eks-kafka-connect-prod-source-ks6008-offset-0 to position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 0 rack: use1-az4)], epoch=7}}. (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:57,429] INFO Finished reading KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-offset (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:57,429] INFO Started KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-offset (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:57,429] INFO Finished reading offsets topic and starting KafkaOffsetBackingStore (org.apache.kafka.connect.storage.KafkaOffsetBackingStore) [2024-05-02 07:50:57,432] INFO Worker started (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:50:57,432] INFO Starting KafkaBasedLog with topic eks-kafka-connect-prod-source-ks6008-status (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:57,441] INFO ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-statuses compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 0 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:57,445] INFO These configurations '[producer.sasl.jaas.config, group.id, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:57,446] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,446] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,446] INFO Kafka startTimeMs: 1714636257445 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,446] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-statuses client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = eks-kafka-connect-prod-source-ks6008 group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:57,450] INFO These configurations '[producer.sasl.jaas.config, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:57,450] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,453] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,453] INFO Kafka startTimeMs: 1714636257450 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:57,526] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:57,536] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:57,536] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Assigned to partition(s): eks-kafka-connect-prod-source-ks6008-status-0 (org.apache.kafka.clients.consumer.KafkaConsumer) [2024-05-02 07:50:57,536] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Seeking to earliest offset of partition eks-kafka-connect-prod-source-ks6008-status-0 (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:57,577] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,577] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,623] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,623] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,763] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,764] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,964] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:57,965] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,092] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-statuses, groupId=eks-kafka-connect-prod-source-ks6008] Resetting offset for partition eks-kafka-connect-prod-source-ks6008-status-0 to position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 5 rack: use1-az6)], epoch=6}}. (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:58,201] INFO Finished reading KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-status (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:58,201] INFO Started KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-status (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:58,204] INFO Starting KafkaConfigBackingStore (org.apache.kafka.connect.storage.KafkaConfigBackingStore) [2024-05-02 07:50:58,204] INFO Starting KafkaBasedLog with topic eks-kafka-connect-prod-source-ks6008-config (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:58,248] INFO ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-configs compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:58,252] INFO These configurations '[producer.sasl.jaas.config, group.id, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:50:58,252] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,253] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,253] INFO Kafka startTimeMs: 1714636258252 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,253] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = eks-kafka-connect-prod-source-ks6008-configs client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = eks-kafka-connect-prod-source-ks6008 group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:58,259] INFO These configurations '[producer.sasl.jaas.config, value.converter.schema.registry.basic.auth.user.info, plugin.path, status.storage.partitions, metrics.context.connect.kafka.cluster.id, offset.storage.partitions, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, config.storage.topic, producer.security.protocol, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, value.converter.schema.registry.basic.auth.credentials.source, producer.sasl.mechanism, rest.port, config.storage.partitions, config.storage.replication.factor, value.converter.schema.registry.url, offset.storage.replication.factor, key.converter.schema.registry.basic.auth.user.info, log4j.root.loglevel, key.converter.schema.registry.basic.auth.credentials.source, key.converter.schema.registry.url]' were supplied but are not used yet. (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:50:58,259] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,259] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,259] INFO Kafka startTimeMs: 1714636258259 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:50:58,283] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-configs] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:58,287] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:58,288] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Assigned to partition(s): eks-kafka-connect-prod-source-ks6008-config-0 (org.apache.kafka.clients.consumer.KafkaConsumer) [2024-05-02 07:50:58,288] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Seeking to earliest offset of partition eks-kafka-connect-prod-source-ks6008-config-0 (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:58,323] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 3 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,323] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 3 (e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,386] INFO [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Node 3 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,386] WARN [AdminClient clientId=eks-kafka-connect-prod-source-ks6008-shared-admin] Connection to node 3 (e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,531] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,532] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 0 (e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,636] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,636] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 1 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,742] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Node 2 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,742] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 2 (e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,847] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Node 3 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,847] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 3 (e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,973] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Node 3 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:58,974] WARN [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 3 (e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:59,195] INFO [Consumer clientId=eks-kafka-connect-prod-source-ks6008-configs, groupId=eks-kafka-connect-prod-source-ks6008] Resetting offset for partition eks-kafka-connect-prod-source-ks6008-config-0 to position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[e-0cd3.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 3 rack: use1-az4)], epoch=6}}. (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:50:59,295] INFO Finished reading KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-config (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:59,295] INFO Started KafkaBasedLog for topic eks-kafka-connect-prod-source-ks6008-config (org.apache.kafka.connect.util.KafkaBasedLog) [2024-05-02 07:50:59,295] INFO Started KafkaConfigBackingStore (org.apache.kafka.connect.storage.KafkaConfigBackingStore) [2024-05-02 07:50:59,295] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Herder started (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:50:59,425] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:50:59,426] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Discovered group coordinator e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483645 rack: null) (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,428] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Rebalance started (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,428] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] (Re-)joining group (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,436] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Node 2147483645 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:59,436] WARN [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 2147483645 (e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:59,437] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Group coordinator e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483645 rack: null) is unavailable or invalid due to cause: null. isDisconnected: true. Rediscovery will be attempted. (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,437] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Request joining group due to: rebalance failed due to 'null' (DisconnectException) (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,543] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:59,543] WARN [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Connection to node 0 (e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:50:59,722] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Discovered group coordinator e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483645 rack: null) (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,723] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] (Re-)joining group (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,747] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:50:59,747] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] (Re-)joining group (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:02,470] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Successfully joined group with generation Generation{generationId=46, memberId='connect-1-41fae91d-cbcc-41ab-a55f-480123080f25', protocol='sessioned'} (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:02,477] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Successfully synced group in generation Generation{generationId=46, memberId='connect-1-41fae91d-cbcc-41ab-a55f-480123080f25', protocol='sessioned'} (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:02,478] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Joined group at generation 46 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-16dddc1b-f922-491a-befc-eb74bde126c4', leaderUrl='http://10.171.221.38:8083/', offset=1058, connectorIds=[], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:02,478] WARN [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Catching up to assignment's config offset. (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:02,478] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Current config state offset -1 is behind group assignment 1058, reading to end of config log (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:02,482] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Finished reading to end of log and updated config snapshot, new config log offset: 1058 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:02,482] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Starting connectors and tasks using config offset 1058 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:02,482] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:12,529] INFO 10.171.220.172 - - [02/May/2024:07:51:12 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 157 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:14,474] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Rebalance started (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:14,475] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] (Re-)joining group (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:14,479] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Successfully joined group with generation Generation{generationId=47, memberId='connect-1-41fae91d-cbcc-41ab-a55f-480123080f25', protocol='sessioned'} (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:14,495] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Successfully synced group in generation Generation{generationId=47, memberId='connect-1-41fae91d-cbcc-41ab-a55f-480123080f25', protocol='sessioned'} (org.apache.kafka.connect.runtime.distributed.WorkerCoordinator) [2024-05-02 07:51:14,495] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Joined group at generation 47 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-41fae91d-cbcc-41ab-a55f-480123080f25', leaderUrl='http://10.171.217.115:8083/', offset=1058, connectorIds=[data-el-6008], taskIds=[data-el-6008-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:14,496] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Starting connectors and tasks using config offset 1058 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:14,497] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Starting task data-el-6008-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:14,497] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Starting connector data-el-6008 (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:14,501] INFO Creating connector data-el-6008 of type io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,502] INFO Creating task data-el-6008-0 (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,503] INFO ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 predicates = [] tasks.max = 1 transforms = [dataelsmt] value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.ConnectorConfig) [2024-05-02 07:51:14,503] INFO SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.SourceConnectorConfig) [2024-05-02 07:51:14,506] INFO EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] transforms.dataelsmt.negate = false transforms.dataelsmt.predicate = null transforms.dataelsmt.source.name = KS6008 transforms.dataelsmt.tables.heartbeat.values = default=-1050505050,ad_creative.creative_id=0,ad_criterias.criteria_id=0,campaign_ads.ad_id=0,dismissed_topics.topic_id=1,element_locations_connection.row_id=0 transforms.dataelsmt.type = class io.skai.dataelsmt.DataELTransformer value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig) [2024-05-02 07:51:14,506] INFO EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 predicates = [] tasks.max = 1 transforms = [dataelsmt] transforms.dataelsmt.negate = false transforms.dataelsmt.predicate = null transforms.dataelsmt.source.name = KS6008 transforms.dataelsmt.tables.heartbeat.values = default=-1050505050,ad_creative.creative_id=0,ad_criterias.criteria_id=0,campaign_ads.ad_id=0,dismissed_topics.topic_id=1,element_locations_connection.row_id=0 transforms.dataelsmt.type = class io.skai.dataelsmt.DataELTransformer value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig) [2024-05-02 07:51:14,511] INFO Instantiated connector data-el-6008 with version 2.2.1.Final of type class io.debezium.connector.mysql.MySqlConnector (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,512] INFO Finished creating connector data-el-6008 (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,516] INFO TaskConfig values: task.class = class io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.TaskConfig) [2024-05-02 07:51:14,517] INFO Instantiated task data-el-6008-0 with version 2.2.1.Final of type io.debezium.connector.mysql.MySqlConnectorTask (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,520] INFO AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.connect.avro.AvroConverterConfig) [2024-05-02 07:51:14,521] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,521] WARN [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,560] INFO KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroSerializerConfig) [2024-05-02 07:51:14,638] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Node 2 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,638] WARN [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Connection to node 2 (e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,744] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,744] WARN [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Connection to node 0 (e-0071.use1-az4.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,754] INFO KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroDeserializerConfig) [2024-05-02 07:51:14,824] INFO AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false (io.confluent.connect.avro.AvroDataConfig) [2024-05-02 07:51:14,825] INFO AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.connect.avro.AvroConverterConfig) [2024-05-02 07:51:14,826] INFO KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroSerializerConfig) [2024-05-02 07:51:14,828] INFO KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://psrc-y9ym7.us-east-1.aws.confluent.cloud] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroDeserializerConfig) [2024-05-02 07:51:14,829] INFO AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false (io.confluent.connect.avro.AvroDataConfig) [2024-05-02 07:51:14,829] INFO Set up the key converter class io.confluent.connect.avro.AvroConverter for task data-el-6008-0 using the connector config (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,829] INFO Set up the value converter class io.confluent.connect.avro.AvroConverter for task data-el-6008-0 using the connector config (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,830] INFO Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task data-el-6008-0 using the worker config (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,833] INFO SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.SourceConnectorConfig) [2024-05-02 07:51:14,833] INFO EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] transforms.dataelsmt.negate = false transforms.dataelsmt.predicate = null transforms.dataelsmt.source.name = KS6008 transforms.dataelsmt.tables.heartbeat.values = default=-1050505050,ad_creative.creative_id=0,ad_criterias.criteria_id=0,campaign_ads.ad_id=0,dismissed_topics.topic_id=1,element_locations_connection.row_id=0 transforms.dataelsmt.type = class io.skai.dataelsmt.DataELTransformer value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig) [2024-05-02 07:51:14,835] INFO Initializing: org.apache.kafka.connect.runtime.TransformationChain{io.skai.dataelsmt.DataELTransformer} (org.apache.kafka.connect.runtime.Worker) [2024-05-02 07:51:14,835] INFO ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = connector-producer-data-el-6008-0 compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 9223372036854775807 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:51:14,838] INFO These configurations '[metrics.context.connect.group.id, metrics.context.connect.kafka.cluster.id]' were supplied but are not used yet. (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:51:14,838] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:14,838] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:14,838] INFO Kafka startTimeMs: 1714636274838 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:14,847] INFO [Worker clientId=connect-1, groupId=eks-kafka-connect-prod-source-ks6008] Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder) [2024-05-02 07:51:14,851] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,851] WARN [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Connection to node 1 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:14,922] INFO [Producer clientId=connector-producer-data-el-6008-0] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:14,935] INFO Starting MySqlConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO connector.class = io.debezium.connector.mysql.MySqlConnector (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO snapshot.locking.mode = none (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO schema.history.internal.consumer.sasl.jaas.config = ******** (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO transforms.dataelsmt.source.name = KS6008 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO transforms = dataelsmt (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO signal.enabled.channels = source,kafka,kafkaCustom (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO include.schema.changes = false (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO schema.history.internal.producer.security.protocol = SASL_SSL (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO log.cleanup.policy = compact (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO converters = datetime,boolean (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO key.converter = io.confluent.connect.avro.AvroConverter (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO schema.history.internal.producer.sasl.mechanism = PLAIN (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO database.user = data_el (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO boolean.type = io.debezium.connector.mysql.converters.TinyIntOneToBooleanConverter (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,935] INFO value.converter.schema.registry.basic.auth.credentials.source = USER_INFO (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.kafka.bootstrap.servers = lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO transforms.dataelsmt.type = io.skai.dataelsmt.DataELTransformer (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO value.converter.schema.registry.url = https://psrc-y9ym7.us-east-1.aws.confluent.cloud (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.password = ******** (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO name = data-el-6008 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.consumer.sasl.mechanism = PLAIN (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.producer.sasl.jaas.config = ******** (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO key.converter.schema.registry.basic.auth.user.info = ******** (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO snapshot.mode = schema_only (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.consumer.security.protocol = SASL_SSL (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO log.cleaner.enable = true (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO tasks.max = 1 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO value.converter.schema.registry.basic.auth.user.info = ******** (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.store.only.captured.tables.ddl = true (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO tombstones.on.delete = false (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO topic.prefix = data_el (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO decimal.handling.mode = double (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO schema.history.internal.kafka.topic = data-el-6008.kazaam-history-docker (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO datetime.type = io.skai.customconverters.DateTimeConverter (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO value.converter = io.confluent.connect.avro.AvroConverter (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO transforms.dataelsmt.tables.heartbeat.values = default=-1050505050,ad_creative.creative_id=0,ad_criterias.criteria_id=0,campaign_ads.ad_id=0,dismissed_topics.topic_id=1,element_locations_connection.row_id=0 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.whitelist = kazaam (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO bigint.unsigned.handling.mode = precise (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.server.id = 60081 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.server.name = ks-60081 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.port = 3306 (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.serverTimezone = UTC (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO task.class = io.debezium.connector.mysql.MySqlConnectorTask (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.hostname = eclidb6008.kenshooprd.local (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.connectionTimeZone = UTC (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO table.include.list = kazaam.ad_criterias,kazaam.heartbeat,kazaam.parameters (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO key.converter.schema.registry.url = https://psrc-y9ym7.us-east-1.aws.confluent.cloud (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO key.converter.schema.registry.basic.auth.credentials.source = USER_INFO (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO database.include.list = kazaam (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:14,936] INFO WorkerSourceTask{id=data-el-6008-0} Source task finished initialization and start (org.apache.kafka.connect.runtime.AbstractWorkerSourceTask) [2024-05-02 07:51:14,962] INFO Attempting to start task (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:15,020] INFO SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.SourceConnectorConfig) [2024-05-02 07:51:15,021] INFO EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mysql.MySqlConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = data-el-6008 offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [dataelsmt] transforms.dataelsmt.negate = false transforms.dataelsmt.predicate = null transforms.dataelsmt.source.name = KS6008 transforms.dataelsmt.tables.heartbeat.values = default=-1050505050,ad_creative.creative_id=0,ad_criterias.criteria_id=0,campaign_ads.ad_id=0,dismissed_topics.topic_id=1,element_locations_connection.row_id=0 transforms.dataelsmt.type = class io.skai.dataelsmt.DataELTransformer value.converter = class io.confluent.connect.avro.AvroConverter (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig) [2024-05-02 07:51:15,022] INFO [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:15,022] WARN [Producer clientId=eks-kafka-connect-prod-source-ks6008-statuses] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:15,023] INFO Initializing io.skai.customconverters.DateTimeConverter (io.skai.customconverters.DateTimeConverter) [2024-05-02 07:51:15,033] INFO Loading the custom topic naming strategy plugin: io.debezium.schema.DefaultTopicNamingStrategy (io.debezium.config.CommonConnectorConfig) [2024-05-02 07:51:15,451] INFO Found previous partition offset MySqlPartition [sourcePartition={server=data_el}]: {transaction_id=null, file=mysql6008-bin.5073195, pos=71034388, row=1, event=2} (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:15,535] INFO KafkaSchemaHistory Consumer config: {key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, group.id=data_el-schemahistory, bootstrap.servers=lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092, security.protocol=SASL_SSL, enable.auto.commit=false, sasl.mechanism=PLAIN, sasl.jaas.config=********, fetch.min.bytes=1, session.timeout.ms=10000, auto.offset.reset=earliest, client.id=data_el-schemahistory} (io.debezium.storage.kafka.history.KafkaSchemaHistory) [2024-05-02 07:51:15,536] INFO KafkaSchemaHistory Producer config: {batch.size=32768, acks=1, bootstrap.servers=lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092, buffer.memory=1048576, key.serializer=org.apache.kafka.common.serialization.StringSerializer, security.protocol=SASL_SSL, retries=1, value.serializer=org.apache.kafka.common.serialization.StringSerializer, sasl.mechanism=PLAIN, sasl.jaas.config=********, max.block.ms=10000, client.id=data_el-schemahistory, linger.ms=0} (io.debezium.storage.kafka.history.KafkaSchemaHistory) [2024-05-02 07:51:15,536] INFO Requested thread factory for connector MySqlConnector, id = data_el named = db-history-config-check (io.debezium.util.Threads) [2024-05-02 07:51:15,538] INFO Idempotence will be disabled because acks is set to 1, not set to 'all'. (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:51:15,538] INFO ProducerConfig values: acks = 1 auto.include.jmx.reporter = true batch.size = 32768 bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] buffer.memory = 1048576 client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 10000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.StringSerializer (org.apache.kafka.clients.producer.ProducerConfig) [2024-05-02 07:51:15,541] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,541] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,541] INFO Kafka startTimeMs: 1714636275541 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,624] INFO Closing connection before starting schema recovery (io.debezium.connector.mysql.MySqlConnectorTask) [2024-05-02 07:51:15,630] INFO Connection gracefully closed (io.debezium.jdbc.JdbcConnection) [2024-05-02 07:51:15,632] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = data_el-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:51:15,634] INFO [Producer clientId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:15,635] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,635] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,635] INFO Kafka startTimeMs: 1714636275635 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,658] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:15,735] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,735] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,736] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,736] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,736] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,737] INFO App info kafka.consumer for data_el-schemahistory unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,738] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = data_el-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:51:15,740] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,740] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,740] INFO Kafka startTimeMs: 1714636275740 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,740] INFO Creating thread debezium-mysqlconnector-data_el-db-history-config-check (io.debezium.util.Threads) [2024-05-02 07:51:15,741] INFO AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory-topic-check connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:51:15,743] INFO These configurations '[batch.size, acks, buffer.memory, key.serializer, value.serializer, max.block.ms, linger.ms]' were supplied but are not used yet. (org.apache.kafka.clients.admin.AdminClientConfig) [2024-05-02 07:51:15,744] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,744] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,744] INFO Kafka startTimeMs: 1714636275743 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,828] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:15,837] INFO [AdminClient clientId=data_el-schemahistory-topic-check] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:15,837] WARN [AdminClient clientId=data_el-schemahistory-topic-check] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:15,853] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,853] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,854] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,854] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,854] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,855] INFO App info kafka.consumer for data_el-schemahistory unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,856] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = data_el-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:51:15,860] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,860] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,860] INFO Kafka startTimeMs: 1714636275860 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,919] INFO Database schema history topic 'data-el-6008.kazaam-history-docker' has correct settings (io.debezium.storage.kafka.history.KafkaSchemaHistory) [2024-05-02 07:51:15,921] INFO App info kafka.admin.client for data_el-schemahistory-topic-check unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,923] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,923] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,923] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,937] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:15,973] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,973] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:15,973] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,973] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,973] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:15,974] INFO App info kafka.consumer for data_el-schemahistory unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,975] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = data_el-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:51:15,977] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,977] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:15,977] INFO Kafka startTimeMs: 1714636275977 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:16,030] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:16,050] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,050] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,050] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:16,050] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:16,051] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:16,051] INFO App info kafka.consumer for data_el-schemahistory unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:16,052] INFO Started database schema history recovery (io.debezium.relational.history.SchemaHistoryMetrics) [2024-05-02 07:51:16,057] INFO ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [lkc-z6kvj7.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = data_el-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = data_el-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = [hidden] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = PLAIN sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_SSL security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer (org.apache.kafka.clients.consumer.ConsumerConfig) [2024-05-02 07:51:16,059] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:16,059] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:16,059] INFO Kafka startTimeMs: 1714636276059 (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:16,060] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Subscribed to topic(s): data-el-6008.kazaam-history-docker (org.apache.kafka.clients.consumer.KafkaConsumer) [2024-05-02 07:51:16,080] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Cluster ID: lkc-z6kvj7 (org.apache.kafka.clients.Metadata) [2024-05-02 07:51:16,137] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Discovered group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,138] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] (Re-)joining group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,145] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Node 2147483646 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:16,145] WARN [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Connection to node 2147483646 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.15.232:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:16,145] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) is unavailable or invalid due to cause: null. isDisconnected: true. Rediscovery will be attempted. (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,145] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: rebalance failed due to 'null' (DisconnectException) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,247] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Discovered group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,248] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] (Re-)joining group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,252] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Node 2147483646 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:16,252] WARN [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Connection to node 2147483646 (e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:16,252] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) is unavailable or invalid due to cause: null. isDisconnected: true. Rediscovery will be attempted. (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,252] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: rebalance failed due to 'null' (DisconnectException) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,354] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Discovered group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,355] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) is unavailable or invalid due to cause: coordinator unavailable. isDisconnected: false. Rediscovery will be attempted. (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,355] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Requesting disconnect from last known coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,457] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Discovered group coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,458] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] (Re-)joining group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,486] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: need to re-join with the given member-id: data_el-schemahistory-fcb4fe94-fb22-4349-bcbb-f989169d96b3 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,486] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,486] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] (Re-)joining group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,498] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Successfully joined group with generation Generation{generationId=1, memberId='data_el-schemahistory-fcb4fe94-fb22-4349-bcbb-f989169d96b3', protocol='range'} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,505] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Finished assignment for group at generation 1: {data_el-schemahistory-fcb4fe94-fb22-4349-bcbb-f989169d96b3=Assignment(partitions=[data-el-6008.kazaam-history-docker-0])} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,521] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Successfully synced group in generation Generation{generationId=1, memberId='data_el-schemahistory-fcb4fe94-fb22-4349-bcbb-f989169d96b3', protocol='range'} (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,521] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Notifying assignor about the new Assignment(partitions=[data-el-6008.kazaam-history-docker-0]) (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,521] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Adding newly assigned partitions: data-el-6008.kazaam-history-docker-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,530] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Found no committed offset for partition data-el-6008.kazaam-history-docker-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:16,532] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting offset for partition data-el-6008.kazaam-history-docker-0 to position FetchPosition{offset=15, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[e-1b67.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 4 rack: use1-az2)], epoch=6}}. (org.apache.kafka.clients.consumer.internals.SubscriptionState) [2024-05-02 07:51:17,340] INFO 10.171.220.172 - - [02/May/2024:07:51:17 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 3 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:17,434] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Revoke previously assigned partitions data-el-6008.kazaam-history-docker-0 (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:17,435] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Member data_el-schemahistory-fcb4fe94-fb22-4349-bcbb-f989169d96b3 sending LeaveGroup request to coordinator e-1bb8.use1-az2.dom8pm0q73g.us-east-1.aws.confluent.cloud:9092 (id: 2147483646 rack: null) due to the consumer is being closed (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:17,435] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:17,435] INFO [Consumer clientId=data_el-schemahistory, groupId=data_el-schemahistory] Request joining group due to: consumer pro-actively leaving the group (org.apache.kafka.clients.consumer.internals.ConsumerCoordinator) [2024-05-02 07:51:17,445] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:17,445] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:17,446] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) [2024-05-02 07:51:17,446] INFO App info kafka.consumer for data_el-schemahistory unregistered (org.apache.kafka.common.utils.AppInfoParser) [2024-05-02 07:51:17,447] INFO Finished database schema history recovery of 4 change(s) in 1394 ms (io.debezium.relational.history.SchemaHistoryMetrics) [2024-05-02 07:51:17,462] INFO Reconnecting after finishing schema recovery (io.debezium.connector.mysql.MySqlConnectorTask) [2024-05-02 07:51:17,525] INFO Get all known binlogs from MySQL (io.debezium.connector.mysql.MySqlConnection) [2024-05-02 07:51:17,566] INFO MySQL has the binlog file 'mysql6008-bin.5073195' required by the connector (io.debezium.connector.mysql.MySqlConnectorTask) [2024-05-02 07:51:17,629] INFO Requested thread factory for connector MySqlConnector, id = data_el named = change-event-source-coordinator (io.debezium.util.Threads) [2024-05-02 07:51:17,631] INFO Creating thread debezium-mysqlconnector-data_el-change-event-source-coordinator (io.debezium.util.Threads) [2024-05-02 07:51:17,631] INFO Successfully started task (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:17,633] INFO Metrics registered (io.debezium.pipeline.ChangeEventSourceCoordinator) [2024-05-02 07:51:17,633] INFO Context created (io.debezium.pipeline.ChangeEventSourceCoordinator) [2024-05-02 07:51:17,638] INFO A previous offset indicating a completed snapshot has been found. Neither schema nor data will be snapshotted. (io.debezium.connector.mysql.MySqlSnapshotChangeEventSource) [2024-05-02 07:51:17,748] INFO Snapshot ended with SnapshotResult [status=SKIPPED, offset=MySqlOffsetContext [sourceInfoSchema=Schema{io.debezium.connector.mysql.Source:STRUCT}, sourceInfo=SourceInfo [currentGtid=null, currentBinlogFilename=mysql6008-bin.5073195, currentBinlogPosition=71034388, currentRowNumber=0, serverId=0, sourceTime=null, threadId=-1, currentQuery=null, tableIds=[], databaseName=null], snapshotCompleted=false, transactionContext=TransactionContext [currentTransactionId=null, perTableEventCount={}, totalEventCount=0], restartGtidSet=null, currentGtidSet=null, restartBinlogFilename=mysql6008-bin.5073195, restartBinlogPosition=71034388, restartRowsToSkip=1, restartEventsToSkip=2, currentEventLengthInBytes=0, inTransaction=false, transactionId=null, incrementalSnapshotContext =IncrementalSnapshotContext [windowOpened=false, chunkEndPosition=null, dataCollectionsToSnapshot=[], lastEventKeySent=null, maximumKey=null]]] (io.debezium.pipeline.ChangeEventSourceCoordinator) [2024-05-02 07:51:17,752] INFO Requested thread factory for connector MySqlConnector, id = data_el named = binlog-client (io.debezium.util.Threads) [2024-05-02 07:51:17,753] INFO Starting streaming (io.debezium.pipeline.ChangeEventSourceCoordinator) [2024-05-02 07:51:17,759] INFO Skip 2 events on streaming start (io.debezium.connector.mysql.MySqlStreamingChangeEventSource) [2024-05-02 07:51:17,759] INFO Skip 1 rows on streaming start (io.debezium.connector.mysql.MySqlStreamingChangeEventSource) [2024-05-02 07:51:17,759] INFO Creating thread debezium-mysqlconnector-data_el-binlog-client (io.debezium.util.Threads) [2024-05-02 07:51:17,822] INFO Creating thread debezium-mysqlconnector-data_el-binlog-client (io.debezium.util.Threads) May 02, 2024 7:51:17 AM com.github.shyiko.mysql.binlog.BinaryLogClient connect INFO: Connected to eclidb6008.kenshooprd.local:3306 at mysql6008-bin.5073195/71034388 (sid:60081, cid:12808668) [2024-05-02 07:51:17,834] INFO Connected to MySQL binlog at eclidb6008.kenshooprd.local:3306, starting at MySqlOffsetContext [sourceInfoSchema=Schema{io.debezium.connector.mysql.Source:STRUCT}, sourceInfo=SourceInfo [currentGtid=null, currentBinlogFilename=mysql6008-bin.5073195, currentBinlogPosition=71034388, currentRowNumber=0, serverId=0, sourceTime=null, threadId=-1, currentQuery=null, tableIds=[], databaseName=null], snapshotCompleted=false, transactionContext=TransactionContext [currentTransactionId=null, perTableEventCount={}, totalEventCount=0], restartGtidSet=null, currentGtidSet=null, restartBinlogFilename=mysql6008-bin.5073195, restartBinlogPosition=71034388, restartRowsToSkip=1, restartEventsToSkip=2, currentEventLengthInBytes=0, inTransaction=false, transactionId=null, incrementalSnapshotContext =IncrementalSnapshotContext [windowOpened=false, chunkEndPosition=null, dataCollectionsToSnapshot=[], lastEventKeySent=null, maximumKey=null]] (io.debezium.connector.mysql.MySqlStreamingChangeEventSource) [2024-05-02 07:51:17,834] INFO Waiting for keepalive thread to start (io.debezium.connector.mysql.MySqlStreamingChangeEventSource) [2024-05-02 07:51:17,834] INFO Creating thread debezium-mysqlconnector-data_el-binlog-client (io.debezium.util.Threads) [2024-05-02 07:51:17,934] INFO Keepalive thread is running (io.debezium.connector.mysql.MySqlStreamingChangeEventSource) [2024-05-02 07:51:21,643] INFO 2 records sent during previous 00:00:07.125, last recorded offset of {server=data_el} partition is {transaction_id=null, ts_sec=1714636281, file=mysql6008-bin.5073195, pos=89206424, row=1, server_id=6008, event=2} (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:22,137] INFO [Producer clientId=connector-producer-data-el-6008-0] Node 2 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,138] WARN [Producer clientId=connector-producer-data-el-6008-0] Connection to node 2 (e-20f0.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,267] INFO [Producer clientId=connector-producer-data-el-6008-0] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,267] WARN [Producer clientId=connector-producer-data-el-6008-0] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.2.159:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,323] INFO [Producer clientId=connector-producer-data-el-6008-0] Node 5 disconnected. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,324] WARN [Producer clientId=connector-producer-data-el-6008-0] Connection to node 5 (e-2040.use1-az6.dom8pm0q73g.us-east-1.aws.confluent.cloud/10.181.25.145:9092) terminated during authentication. This may happen due to any of the following reasons: (1) Authentication failed due to invalid credentials with brokers older than 1.0.0, (2) Firewall blocking Kafka TLS traffic (eg it may only allow HTTPS traffic), (3) Transient network issue. (org.apache.kafka.clients.NetworkClient) [2024-05-02 07:51:22,339] INFO 10.171.220.172 - - [02/May/2024:07:51:22 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 3 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:27,337] INFO 10.171.220.172 - - [02/May/2024:07:51:27 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 2 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:31,781] INFO 2 records sent during previous 00:00:10.138, last recorded offset of {server=data_el} partition is {transaction_id=null, ts_sec=1714636291, file=mysql6008-bin.5073196, pos=46529036, row=1, server_id=6008, event=2} (io.debezium.connector.common.BaseSourceTask) [2024-05-02 07:51:32,336] INFO 10.171.220.172 - - [02/May/2024:07:51:32 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 2 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:37,338] INFO 10.171.220.172 - - [02/May/2024:07:51:37 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 4 (org.apache.kafka.connect.runtime.rest.RestServer) [2024-05-02 07:51:42,337] INFO 10.171.220.172 - - [02/May/2024:07:51:42 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "kube-probe/1.27+" 2 (org.apache.kafka.connect.runtime.rest.RestServer)