2023-09-13 20:52:45,658 INFO || [Producer clientId=naument--statuses] Node 3 disconnected. [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:53:39,844 INFO || [AdminClient clientId=naument--shared-admin] Node 2 disconnected. [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:20,790 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:20 +0000] "GET /connectors HTTP/1.1" 200 2 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,007 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors HTTP/1.1" 200 2 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,013 INFO || Loading the custom source info struct maker plugin: io.debezium.connector.oracle.OracleSourceInfoStructMaker [io.debezium.config.CommonConnectorConfig] 2023-09-13 20:56:39,077 INFO || Database Version: Oracle Database 12c Enterprise Edition Release 12.2.0.1.0 - 64bit Production [io.debezium.connector.oracle.OracleConnection] 2023-09-13 20:56:39,079 INFO || Connection gracefully closed [io.debezium.jdbc.JdbcConnection] 2023-09-13 20:56:39,080 INFO || AbstractConfig values: [org.apache.kafka.common.config.AbstractConfig] 2023-09-13 20:56:39,085 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau55_src config updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,085 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,085 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,086 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "POST /connectors HTTP/1.1" 201 1491 "-" "ReactorNetty/1.1.6" 76 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,086 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=141, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,088 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=141, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,088 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 141 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', leaderUrl='http://172.18.0.6:8083/', offset=2978, connectorIds=[vk_nau55_src], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,089 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 2978 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,089 INFO || [Worker clientId=connect-1, groupId=naument] Starting connector vk_nau55_src [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,089 INFO || Creating connector vk_nau55_src of type io.debezium.connector.oracle.OracleConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,089 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-13 20:56:39,089 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,089 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-13 20:56:39,089 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,090 INFO || Instantiated connector vk_nau55_src with version 2.4.0.Beta1 of type class io.debezium.connector.oracle.OracleConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,090 INFO || Finished creating connector vk_nau55_src [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,090 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,091 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1491 "-" "ReactorNetty/1.1.6" 4 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,092 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-13 20:56:39,092 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,093 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-13 20:56:39,093 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,094 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 404 73 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,100 INFO || [Worker clientId=connect-1, groupId=naument] Tasks [vk_nau55_src-0] configs updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,102 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,102 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,103 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=142, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,105 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=142, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:56:39,105 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 142 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', leaderUrl='http://172.18.0.6:8083/', offset=2980, connectorIds=[vk_nau55_src], taskIds=[vk_nau55_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,105 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 2980 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,105 INFO || [Worker clientId=connect-1, groupId=naument] Starting task vk_nau55_src-0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,105 INFO || Creating task vk_nau55_src-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,106 INFO || ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig] 2023-09-13 20:56:39,106 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,106 INFO || TaskConfig values: task.class = class io.debezium.connector.oracle.OracleConnectorTask [org.apache.kafka.connect.runtime.TaskConfig] 2023-09-13 20:56:39,107 INFO || Instantiated task vk_nau55_src-0 with version 2.4.0.Beta1 of type io.debezium.connector.oracle.OracleConnectorTask [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,107 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-13 20:56:39,107 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-13 20:56:39,108 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-13 20:56:39,108 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-13 20:56:39,108 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-13 20:56:39,109 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-13 20:56:39,109 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-13 20:56:39,109 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-13 20:56:39,109 INFO || Set up the key converter class io.confluent.connect.avro.AvroConverter for task vk_nau55_src-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,109 INFO || Set up the value converter class io.confluent.connect.avro.AvroConverter for task vk_nau55_src-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,109 INFO || Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task vk_nau55_src-0 using the worker config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,110 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-13 20:56:39,110 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,110 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-13 20:56:39,110 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:56:39,110 INFO || Initializing: org.apache.kafka.connect.runtime.TransformationChain{} [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:56:39,110 INFO || ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = connector-producer-vk_nau55_src-0 compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 9223372036854775807 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-13 20:56:39,113 INFO || These configurations '[metrics.context.connect.kafka.cluster.id, metrics.context.connect.group.id]' were supplied but are not used yet. [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-13 20:56:39,113 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,113 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,113 INFO || Kafka startTimeMs: 1694627799113 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,120 INFO || AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] client.dns.lookup = use_all_dns_ips client.id = connector-adminclient-vk_nau55_src-0 connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-13 20:56:39,122 INFO || These configurations '[group.id, max.partition.fetch.bytes, rest.advertised.port, task.shutdown.graceful.timeout.ms, plugin.path, metrics.context.connect.kafka.cluster.id, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, message.max.bytes, config.storage.topic, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, rest.host.name, offset.flush.timeout.ms, config.storage.replication.factor, offset.flush.interval.ms, rest.port, key.converter.schemas.enable, max.request.size, replica.fetch.max.bytes, value.converter.schemas.enable, offset.storage.replication.factor]' were supplied but are not used yet. [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-13 20:56:39,122 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,122 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,122 INFO || Kafka startTimeMs: 1694627799122 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,122 INFO || [Producer clientId=connector-producer-vk_nau55_src-0] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-13 20:56:39,123 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:56:39,123 INFO || Starting OracleConnectorTask with configuration: [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || connector.class = io.debezium.connector.oracle.OracleConnector [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.default.partitions = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || tasks.max = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || schema.history.internal.store.only.captured.tables.ddl = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || schema.history.internal.store.only.captured.databases.ddl = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || include.schema.changes = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.prefix = vk_nau55 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || decimal.handling.mode = precise [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || schema.history.internal.kafka.topic = vk_nau55_src.schema-changes [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.default.include = vk_nau55\.* [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.default.replication.factor = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || lob.enabled = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || value.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || errors.log.enable = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || key.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || snapshot.lock.timeout.ms = 5000 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.user = debezium [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.dbname = NAUMENT1 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || datatype.propagate.source.type = .* [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.default.compression.type = lz4 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.connection.adapter = logminer [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || schema.history.internal.kafka.bootstrap.servers = broker1:29092,broker3:29092,broker3:29092 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.default.retention.ms = 432000000 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.port = 1521 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || topic.creation.enable = true [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || value.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || log.mining.session.max.ms = 120000 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || task.class = io.debezium.connector.oracle.OracleConnectorTask [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.hostname = naumen-db-test.rgs.ru [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || database.password = ******** [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || name = vk_nau55_src [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || table.include.list = DEBEZIUM.GBC_TBL_SERVICECALL_NC55 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || key.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,124 INFO || snapshot.mode = always [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,125 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 16 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,125 INFO || Loading the custom source info struct maker plugin: io.debezium.connector.oracle.OracleSourceInfoStructMaker [io.debezium.config.CommonConnectorConfig] 2023-09-13 20:56:39,125 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 16 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,125 INFO || Loading the custom topic naming strategy plugin: io.debezium.schema.SchemaTopicNamingStrategy [io.debezium.config.CommonConnectorConfig] 2023-09-13 20:56:39,129 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 404 70 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,130 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:39 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 112 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:39,186 INFO || Database Version: Oracle Database 12c Enterprise Edition Release 12.2.0.1.0 - 64bit Production [io.debezium.connector.oracle.OracleConnection] 2023-09-13 20:56:39,198 INFO || KafkaSchemaHistory Consumer config: {key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, enable.auto.commit=false, group.id=vk_nau55-schemahistory, bootstrap.servers=broker1:29092,broker3:29092,broker3:29092, fetch.min.bytes=1, session.timeout.ms=10000, auto.offset.reset=earliest, client.id=vk_nau55-schemahistory} [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-13 20:56:39,198 INFO || KafkaSchemaHistory Producer config: {retries=1, value.serializer=org.apache.kafka.common.serialization.StringSerializer, acks=1, batch.size=32768, max.block.ms=10000, bootstrap.servers=broker1:29092,broker3:29092,broker3:29092, buffer.memory=1048576, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=vk_nau55-schemahistory, linger.ms=0} [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-13 20:56:39,198 INFO || Requested thread factory for connector OracleConnector, id = vk_nau55 named = db-history-config-check [io.debezium.util.Threads] 2023-09-13 20:56:39,198 INFO || Idempotence will be disabled because acks is set to 1, not set to 'all'. [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-13 20:56:39,198 INFO || ProducerConfig values: acks = 1 auto.include.jmx.reporter = true batch.size = 32768 bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] buffer.memory = 1048576 client.dns.lookup = use_all_dns_ips client.id = vk_nau55-schemahistory compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 10000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.StringSerializer [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-13 20:56:39,200 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,200 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,200 INFO || Kafka startTimeMs: 1694627799200 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,202 INFO || [Producer clientId=vk_nau55-schemahistory] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-13 20:56:39,203 INFO || No previous offsets found [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:39,203 INFO || Connector started for the first time, database schema history recovery will not be executed [io.debezium.connector.oracle.OracleConnectorTask] 2023-09-13 20:56:39,204 INFO || ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = vk_nau55-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = vk_nau55-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-13 20:56:39,205 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,205 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,205 INFO || Kafka startTimeMs: 1694627799205 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,207 INFO || [Consumer clientId=vk_nau55-schemahistory, groupId=vk_nau55-schemahistory] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-13 20:56:39,209 INFO || [Consumer clientId=vk_nau55-schemahistory, groupId=vk_nau55-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:56:39,209 INFO || [Consumer clientId=vk_nau55-schemahistory, groupId=vk_nau55-schemahistory] Request joining group due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:56:39,209 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,209 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,209 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,210 INFO || App info kafka.consumer for vk_nau55-schemahistory unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,210 INFO || AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] client.dns.lookup = use_all_dns_ips client.id = vk_nau55-schemahistory connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-13 20:56:39,211 INFO || These configurations '[value.serializer, acks, batch.size, max.block.ms, buffer.memory, key.serializer, linger.ms]' were supplied but are not used yet. [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-13 20:56:39,211 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,211 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,211 INFO || Kafka startTimeMs: 1694627799211 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,244 INFO || Database schema history topic '(name=vk_nau55_src.schema-changes, numPartitions=1, replicationFactor=default, replicasAssignments=null, configs={cleanup.policy=delete, retention.ms=9223372036854775807, retention.bytes=-1})' created [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-13 20:56:39,244 INFO || App info kafka.admin.client for vk_nau55-schemahistory unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:56:39,245 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,245 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,245 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 20:56:39,245 INFO || Requested thread factory for connector OracleConnector, id = vk_nau55 named = SignalProcessor [io.debezium.util.Threads] 2023-09-13 20:56:39,246 INFO || Requested thread factory for connector OracleConnector, id = vk_nau55 named = change-event-source-coordinator [io.debezium.util.Threads] 2023-09-13 20:56:39,246 INFO || Requested thread factory for connector OracleConnector, id = vk_nau55 named = blocking-snapshot [io.debezium.util.Threads] 2023-09-13 20:56:39,246 INFO || Creating thread debezium-oracleconnector-vk_nau55-change-event-source-coordinator [io.debezium.util.Threads] 2023-09-13 20:56:39,246 INFO Oracle|vk_nau55|snapshot Metrics registered [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot Context created [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot Snapshot mode is set to ALWAYS, not checking exiting offset. [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot According to the connector configuration both schema and data will be snapshot. [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-13 20:56:39,247 INFO || SignalProcessor started. Scheduling it every 5000ms [io.debezium.pipeline.signal.SignalProcessor] 2023-09-13 20:56:39,247 INFO || Creating thread debezium-oracleconnector-vk_nau55-SignalProcessor [io.debezium.util.Threads] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot Snapshot step 1 - Preparing [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot Snapshot step 2 - Determining captured tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:39,247 INFO Oracle|vk_nau55|snapshot WorkerSourceTask{id=vk_nau55_src-0} Source task finished initialization and start [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:40,327 INFO Oracle|vk_nau55|snapshot Adding table NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 to the list of capture schema tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:40,332 INFO Oracle|vk_nau55|snapshot Created connection pool with 1 threads [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:40,332 INFO Oracle|vk_nau55|snapshot Snapshot step 3 - Locking captured tables [NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55] [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:40,335 INFO Oracle|vk_nau55|snapshot Snapshot step 4 - Determining snapshot offset [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:40,476 INFO Oracle|vk_nau55|snapshot No in-progress transactions will be captured. [io.debezium.connector.oracle.logminer.LogMinerAdapter] 2023-09-13 20:56:40,478 INFO Oracle|vk_nau55|snapshot Connection gracefully closed [io.debezium.jdbc.JdbcConnection] 2023-09-13 20:56:40,478 INFO Oracle|vk_nau55|snapshot Snapshot step 5 - Reading structure of captured tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:40,478 INFO Oracle|vk_nau55|snapshot Only captured tables schema should be captured, capturing: [NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55] [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-13 20:56:41,153 INFO Oracle|vk_nau55|snapshot Snapshot step 6 - Persisting schema history [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:41,153 INFO Oracle|vk_nau55|snapshot Capturing structure of table NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:42,439 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,441 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,442 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 0 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,444 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,446 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,448 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:42,449 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:42 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 30 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:43,878 INFO Oracle|vk_nau55|snapshot Already applied 1 database changes [io.debezium.relational.history.SchemaHistoryMetrics] 2023-09-13 20:56:43,880 INFO Oracle|vk_nau55|snapshot Snapshot step 7 - Snapshotting data [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:43,880 INFO Oracle|vk_nau55|snapshot Creating snapshot worker pool with 1 worker thread(s) [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:43,880 INFO Oracle|vk_nau55|snapshot For table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' using select statement: 'SELECT "ID", "CREATION_DATE", "CLAIM_TRANSFERDATE", "TITLE", "CLIENT_EMAIL", "CLAIM_SUMRETURN" FROM "DEBEZIUM"."GBC_TBL_SERVICECALL_NC55" AS OF SCN 290260831995' [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:43,880 INFO Oracle|vk_nau55|snapshot Exporting data from table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' (1 of 1 tables) [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:43,898 INFO Oracle|vk_nau55|snapshot Finished exporting 2 records for table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' (1 of 1 tables); total duration '00:00:00.018' [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-13 20:56:43,899 INFO Oracle|vk_nau55|snapshot Snapshot - Final stage [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2023-09-13 20:56:43,899 INFO Oracle|vk_nau55|snapshot Snapshot completed [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2023-09-13 20:56:43,899 INFO Oracle|vk_nau55|snapshot Snapshot ended with SnapshotResult [status=COMPLETED, offset=OracleOffsetContext [scn=290260831995, commit_scn=[]]] [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-13 20:56:43,899 INFO Oracle|vk_nau55|streaming Connected metrics set to 'true' [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-13 20:56:43,899 INFO Oracle|vk_nau55|streaming Starting streaming [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Redo Log Group Sizes: [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #1: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #2: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #3: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #4: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #5: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #6: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,166 INFO Oracle|vk_nau55|streaming Group #7: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:56:44,257 INFO Oracle|vk_nau55|snapshot 3 records sent during previous 00:00:05.15, last recorded offset of {server=vk_nau55} partition is {snapshot_scn=290260831995, snapshot=true, scn=290260831995, snapshot_completed=true} [io.debezium.connector.common.BaseSourceTask] 2023-09-13 20:56:44,277 INFO Oracle|vk_nau55|snapshot The task will send records to topic 'vk_nau55' for the first time. Checking whether topic exists [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,279 INFO Oracle|vk_nau55|snapshot Creating topic 'vk_nau55' [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,309 INFO Oracle|vk_nau55|snapshot Created topic (name=vk_nau55, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000}) on brokers at broker1:29092,broker2:29092,broker3:29092 [org.apache.kafka.connect.util.TopicAdmin] 2023-09-13 20:56:44,309 INFO Oracle|vk_nau55|snapshot Created topic '(name=vk_nau55, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000})' using creation group TopicCreationGroup{name='default', inclusionPattern=.*, exclusionPattern=, numPartitions=1, replicationFactor=1, otherConfigs={compression.type=lz4, retention.ms=432000000}} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,311 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 3 : {vk_nau55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,412 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 4 : {vk_nau55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,528 INFO Oracle|vk_nau55|snapshot The task will send records to topic 'vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' for the first time. Checking whether topic exists [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,528 INFO Oracle|vk_nau55|snapshot Creating topic 'vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,557 INFO Oracle|vk_nau55|snapshot Created topic (name=vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000}) on brokers at broker1:29092,broker2:29092,broker3:29092 [org.apache.kafka.connect.util.TopicAdmin] 2023-09-13 20:56:44,557 INFO Oracle|vk_nau55|snapshot Created topic '(name=vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000})' using creation group TopicCreationGroup{name='default', inclusionPattern=.*, exclusionPattern=, numPartitions=1, replicationFactor=1, otherConfigs={compression.type=lz4, retention.ms=432000000}} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-13 20:56:44,558 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 8 : {vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,659 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 9 : {vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,659 INFO || [Producer clientId=connector-producer-vk_nau55_src-0] Resetting the last seen epoch of partition vk_nau55-0 to 0 since the associated topicId changed from null to psuBgggpQauRZPlntPOHjw [org.apache.kafka.clients.Metadata] 2023-09-13 20:56:44,760 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 10 : {vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,861 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 11 : {vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:44,962 WARN || [Producer clientId=connector-producer-vk_nau55_src-0] Error while fetching metadata with correlation id 12 : {vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-13 20:56:45,048 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:45 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:45,048 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:45 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:45,050 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:45 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:45,051 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:45 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:47,282 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:47 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:47,285 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:47 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:47,285 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:47 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:47,287 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:47 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,626 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,628 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,630 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,632 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,634 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,635 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:56:48,636 INFO || 10.0.2.3 - - [13/Sep/2023:17:56:48 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 0 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:39,123 INFO || WorkerSourceTask{id=vk_nau55_src-0} Committing offsets for 3 acknowledged messages [org.apache.kafka.connect.runtime.WorkerSourceTask] 2023-09-13 20:57:53,844 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,848 INFO || AbstractConfig values: [org.apache.kafka.common.config.AbstractConfig] 2023-09-13 20:57:53,851 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau55_sink config updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,851 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,851 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,852 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "POST /connectors HTTP/1.1" 201 868 "-" "ReactorNetty/1.1.6" 6 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,852 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=143, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,854 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=143, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,854 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 143 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', leaderUrl='http://172.18.0.6:8083/', offset=2981, connectorIds=[vk_nau55_sink, vk_nau55_src], taskIds=[vk_nau55_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,854 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 2981 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,855 INFO || [Worker clientId=connect-1, groupId=naument] Starting connector vk_nau55_sink [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,855 INFO || Creating connector vk_nau55_sink of type io.debezium.connector.jdbc.JdbcSinkConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,855 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-13 20:57:53,855 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:57:53,855 INFO || Instantiated connector vk_nau55_sink with version 2.4.0.Beta1 of type class io.debezium.connector.jdbc.JdbcSinkConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,855 INFO || Finished creating connector vk_nau55_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,855 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,856 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-13 20:57:53,856 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:57:53,856 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 868 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,860 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 404 74 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,863 INFO || [Worker clientId=connect-1, groupId=naument] Tasks [vk_nau55_sink-0] configs updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,863 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,863 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,864 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=144, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,865 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=144, memberId='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-13 20:57:53,866 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 144 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-e6c3877a-a74d-4220-9079-58bada7b10b7', leaderUrl='http://172.18.0.6:8083/', offset=2983, connectorIds=[vk_nau55_sink, vk_nau55_src], taskIds=[vk_nau55_sink-0, vk_nau55_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,866 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 2983 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,866 INFO || [Worker clientId=connect-1, groupId=naument] Starting task vk_nau55_sink-0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,866 INFO || Creating task vk_nau55_sink-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,866 INFO || ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig] 2023-09-13 20:57:53,866 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:57:53,866 INFO || TaskConfig values: task.class = class io.debezium.connector.jdbc.JdbcSinkConnectorTask [org.apache.kafka.connect.runtime.TaskConfig] 2023-09-13 20:57:53,866 INFO || Instantiated task vk_nau55_sink-0 with version 2.4.0.Beta1 of type io.debezium.connector.jdbc.JdbcSinkConnectorTask [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,866 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-13 20:57:53,867 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-13 20:57:53,868 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-13 20:57:53,868 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-13 20:57:53,868 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-13 20:57:53,869 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-13 20:57:53,869 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-13 20:57:53,869 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-13 20:57:53,869 INFO || Set up the key converter class io.confluent.connect.avro.AvroConverter for task vk_nau55_sink-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,869 INFO || Set up the value converter class io.confluent.connect.avro.AvroConverter for task vk_nau55_sink-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,869 INFO || Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task vk_nau55_sink-0 using the worker config [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,870 INFO || Initializing: org.apache.kafka.connect.runtime.TransformationChain{} [org.apache.kafka.connect.runtime.Worker] 2023-09-13 20:57:53,870 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-13 20:57:53,870 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau55_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-13 20:57:53,870 INFO || ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = connector-consumer-vk_nau55_sink-0 client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = connect-vk_nau55_sink group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-13 20:57:53,872 INFO || These configurations '[metrics.context.connect.kafka.cluster.id, metrics.context.connect.group.id]' were supplied but are not used yet. [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-13 20:57:53,872 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:57:53,872 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:57:53,872 INFO || Kafka startTimeMs: 1694627873872 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 20:57:53,873 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-13 20:57:53,873 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Subscribed to pattern: 'vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55' [org.apache.kafka.clients.consumer.KafkaConsumer] 2023-09-13 20:57:53,874 INFO || Starting JdbcSinkConnectorConfig with configuration: [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || connector.class = io.debezium.connector.jdbc.JdbcSinkConnector [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || table.name.format = vk_nau55_tbl_servicecall [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || connection.password = ******** [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || primary.key.mode = record_key [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || tasks.max = 1 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || connection.username = debeziumt [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || quote.identifiers = false [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || topics.regex = vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || value.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || delete.enabled = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || schema.evolution = basic [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || auto.evolve = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || task.class = io.debezium.connector.jdbc.JdbcSinkConnectorTask [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || name = vk_nau55_sink [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || auto.create = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || connection.url = jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || value.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || insert.mode = upsert [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || key.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,874 INFO || key.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-13 20:57:53,875 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,875 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,878 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 404 71 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,880 INFO || 10.0.2.3 - - [13/Sep/2023:17:57:53 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 111 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:57:53,881 INFO || HHH000130: Instantiating explicit connection provider: org.hibernate.c3p0.internal.C3P0ConnectionProvider [org.hibernate.engine.jdbc.connections.internal.ConnectionProviderInitiator] 2023-09-13 20:57:53,881 INFO || HHH010002: C3P0 using driver: null at URL: jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1 [org.hibernate.orm.connections.pooling.c3p0] 2023-09-13 20:57:53,881 INFO || HHH10001001: Connection properties: {password=****, user=debeziumt} [org.hibernate.orm.connections.pooling.c3p0] 2023-09-13 20:57:53,881 INFO || HHH10001003: Autocommit mode: false [org.hibernate.orm.connections.pooling.c3p0] 2023-09-13 20:57:53,881 WARN || HHH10001006: No JDBC Driver class was specified by property hibernate.connection.driver_class [org.hibernate.orm.connections.pooling.c3p0] 2023-09-13 20:57:53,905 INFO || HHH10001007: JDBC isolation level: [org.hibernate.orm.connections.pooling.c3p0] 2023-09-13 20:57:53,906 INFO || Initializing c3p0 pool... com.mchange.v2.c3p0.PoolBackedDataSource@500940b5 [ connectionPoolDataSource -> com.mchange.v2.c3p0.WrapperConnectionPoolDataSource@820b1718 [ acquireIncrement -> 32, acquireRetryAttempts -> 30, acquireRetryDelay -> 1000, autoCommitOnClose -> false, automaticTestTable -> null, breakAfterAcquireFailure -> false, checkoutTimeout -> 0, connectionCustomizerClassName -> null, connectionTesterClassName -> com.mchange.v2.c3p0.impl.DefaultConnectionTester, contextClassLoaderSource -> caller, debugUnreturnedConnectionStackTraces -> false, factoryClassLocation -> null, forceIgnoreUnresolvedTransactions -> false, forceSynchronousCheckins -> false, identityToken -> 2rvy88ayj9hfi81jom5bf|411391cb, idleConnectionTestPeriod -> 0, initialPoolSize -> 5, maxAdministrativeTaskTime -> 0, maxConnectionAge -> 0, maxIdleTime -> 0, maxIdleTimeExcessConnections -> 0, maxPoolSize -> 32, maxStatements -> 0, maxStatementsPerConnection -> 0, minPoolSize -> 5, nestedDataSource -> com.mchange.v2.c3p0.DriverManagerDataSource@f814a92d [ description -> null, driverClass -> null, factoryClassLocation -> null, forceUseNamedDriverClass -> false, identityToken -> 2rvy88ayj9hfi81jom5bf|4d151806, jdbcUrl -> jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1, properties -> {password=******, user=******} ], preferredTestQuery -> null, privilegeSpawnedThreads -> false, propertyCycle -> 0, statementCacheNumDeferredCloseThreads -> 0, testConnectionOnCheckin -> false, testConnectionOnCheckout -> false, unreturnedConnectionTimeout -> 0, usesTraditionalReflectiveProxies -> false; userOverrides: {} ], dataSourceName -> null, extensions -> {}, factoryClassLocation -> null, identityToken -> 2rvy88ayj9hfi81jom5bf|5e6ecc93, numHelperThreads -> 3 ] [com.mchange.v2.c3p0.impl.AbstractPoolBackedDataSource] 2023-09-13 20:57:53,929 INFO || HHH000400: Using dialect: org.hibernate.dialect.PostgreSQLDialect [SQL dialect] 2023-09-13 20:57:53,942 INFO || HHH000490: Using JtaPlatform implementation: [org.hibernate.engine.transaction.jta.platform.internal.NoJtaPlatform] [org.hibernate.engine.transaction.jta.platform.internal.JtaPlatformInitiator] 2023-09-13 20:57:53,942 INFO || Using dialect io.debezium.connector.jdbc.dialect.postgres.PostgresDatabaseDialect [io.debezium.connector.jdbc.dialect.DatabaseDialectResolver] 2023-09-13 20:57:53,946 INFO || Database TimeZone: Europe/Moscow [io.debezium.connector.jdbc.dialect.GeneralDatabaseDialect] 2023-09-13 20:57:53,946 INFO || Database version 13.2.0 [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-13 20:57:53,946 INFO || WorkerSinkTask{id=vk_nau55_sink-0} Sink task finished initialization and start [org.apache.kafka.connect.runtime.WorkerSinkTask] 2023-09-13 20:57:53,946 INFO || WorkerSinkTask{id=vk_nau55_sink-0} Executing sink task [org.apache.kafka.connect.runtime.WorkerSinkTask] 2023-09-13 20:57:53,949 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-13 20:57:53,949 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Discovered group coordinator broker2:29092 (id: 2147483645 rack: null) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,950 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] (Re-)joining group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,953 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Request joining group due to: need to re-join with the given member-id: connector-consumer-vk_nau55_sink-0-63d1c2d6-ab64-4c58-a83e-ab20c0125af1 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,954 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,954 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] (Re-)joining group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,955 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Successfully joined group with generation Generation{generationId=1, memberId='connector-consumer-vk_nau55_sink-0-63d1c2d6-ab64-4c58-a83e-ab20c0125af1', protocol='range'} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,956 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Finished assignment for group at generation 1: {connector-consumer-vk_nau55_sink-0-63d1c2d6-ab64-4c58-a83e-ab20c0125af1=Assignment(partitions=[vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0])} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,958 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Successfully synced group in generation Generation{generationId=1, memberId='connector-consumer-vk_nau55_sink-0-63d1c2d6-ab64-4c58-a83e-ab20c0125af1', protocol='range'} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,958 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Notifying assignor about the new Assignment(partitions=[vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0]) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,958 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Adding newly assigned partitions: vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,959 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Found no committed offset for partition vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 20:57:53,961 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Resetting offset for partition vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0 to position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[broker1:29092 (id: 1 rack: null)], epoch=0}}. [org.apache.kafka.clients.consumer.internals.SubscriptionState] Hibernate: CREATE TABLE vk_nau55_tbl_servicecall (ID decimal(19,0) NOT NULL, CREATION_DATE timestamp(6) NOT NULL, CLAIM_TRANSFERDATE timestamp(6) NULL, TITLE varchar(4000) NULL, CLIENT_EMAIL varchar(255) NULL, CLAIM_SUMRETURN double precision NULL, PRIMARY KEY(ID)) Hibernate: INSERT INTO naument1.vk_nau55_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN) VALUES (?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN Hibernate: INSERT INTO naument1.vk_nau55_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN) VALUES (?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN 2023-09-13 20:58:08,878 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,880 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,881 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,882 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,883 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,885 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,885 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,887 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,888 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,889 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,890 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,891 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:08,891 INFO || 10.0.2.3 - - [13/Sep/2023:17:58:08 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:58:47,717 INFO Oracle|vk_nau55|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-13 20:59:05,804 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,808 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,808 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,810 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,810 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,812 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,812 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,814 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,815 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,816 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,817 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,818 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:05,818 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:05 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 0 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,142 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,145 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,145 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,147 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,148 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,149 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,150 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,151 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,152 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,152 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,154 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,154 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:08,158 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:08 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:14,862 INFO Oracle|vk_nau55|snapshot 1 records sent during previous 00:02:30.605, last recorded offset of {server=vk_nau55} partition is {commit_scn=290260873266:1:09001200cc2c3400, transaction_id=null, snapshot_scn=290260831995, scn=290260873265} [io.debezium.connector.common.BaseSourceTask] Hibernate: INSERT INTO naument1.vk_nau55_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN) VALUES (?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN 2023-09-13 20:59:39,127 INFO || WorkerSourceTask{id=vk_nau55_src-0} Committing offsets for 1 acknowledged messages [org.apache.kafka.connect.runtime.WorkerSourceTask] Hibernate: DELETE FROM naument1.vk_nau55_tbl_servicecall WHERE ID=? 2023-09-13 20:59:55,957 ERROR || Failed to process record: Failed to process a sink record [io.debezium.connector.jdbc.JdbcSinkConnectorTask] org.apache.kafka.connect.errors.ConnectException: Failed to process a sink record at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:82) at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:93) at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:587) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:336) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:237) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:206) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:204) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:259) at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:181) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.NullPointerException at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.isFlattened(SinkRecordDescriptor.java:281) at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.build(SinkRecordDescriptor.java:273) at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:70) ... 13 more 2023-09-13 20:59:58,302 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,306 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,306 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,308 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,309 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,311 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,311 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,313 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,314 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,315 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,316 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,317 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:58,317 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:58 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,794 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,796 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,796 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,798 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,798 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,800 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,801 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,802 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,803 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,804 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,805 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,806 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 20:59:59,806 INFO || 10.0.2.5 - - [13/Sep/2023:17:59:59 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,595 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,598 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,598 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,600 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,601 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,602 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,603 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,605 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,605 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,606 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 0 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,607 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,608 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:00,609 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:00 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,510 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,513 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,513 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,515 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,515 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,517 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,518 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,520 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,521 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,522 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,523 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,524 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:01,524 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:01 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,657 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,660 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,660 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,662 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,662 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,664 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,665 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,666 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,667 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,668 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,669 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,669 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:02,671 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:02 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:33,019 ERROR || WorkerSinkTask{id=vk_nau55_sink-0} Task threw an uncaught and unrecoverable exception. Task is being killed and will not recover until manually restarted. Error: JDBC sink connector failure [org.apache.kafka.connect.runtime.WorkerSinkTask] org.apache.kafka.connect.errors.ConnectException: JDBC sink connector failure at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:83) at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:587) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:336) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:237) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:206) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:204) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:259) at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:181) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: org.apache.kafka.connect.errors.ConnectException: Failed to process a sink record at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:82) at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:93) ... 12 more Caused by: java.lang.NullPointerException at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.isFlattened(SinkRecordDescriptor.java:281) at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.build(SinkRecordDescriptor.java:273) at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:70) ... 13 more 2023-09-13 21:00:33,020 ERROR || WorkerSinkTask{id=vk_nau55_sink-0} Task threw an uncaught and unrecoverable exception. Task is being killed and will not recover until manually restarted [org.apache.kafka.connect.runtime.WorkerTask] org.apache.kafka.connect.errors.ConnectException: Exiting WorkerSinkTask due to unrecoverable exception. at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:618) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:336) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:237) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:206) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:204) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:259) at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:181) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: org.apache.kafka.connect.errors.ConnectException: JDBC sink connector failure at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:83) at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:587) ... 11 more Caused by: org.apache.kafka.connect.errors.ConnectException: Failed to process a sink record at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:82) at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:93) ... 12 more Caused by: java.lang.NullPointerException at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.isFlattened(SinkRecordDescriptor.java:281) at io.debezium.connector.jdbc.SinkRecordDescriptor$Builder.build(SinkRecordDescriptor.java:273) at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:70) ... 13 more 2023-09-13 21:00:33,020 INFO || Closing session. [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-13 21:00:33,020 INFO || Closing the session factory [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-13 21:00:33,022 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Revoke previously assigned partitions vk_nau55.DEBEZIUM.GBC_TBL_SERVICECALL_NC55-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 21:00:33,022 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Member connector-consumer-vk_nau55_sink-0-63d1c2d6-ab64-4c58-a83e-ab20c0125af1 sending LeaveGroup request to coordinator broker2:29092 (id: 2147483645 rack: null) due to the consumer is being closed [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 21:00:33,022 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Resetting generation and member id due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 21:00:33,023 INFO || [Consumer clientId=connector-consumer-vk_nau55_sink-0, groupId=connect-vk_nau55_sink] Request joining group due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-13 21:00:33,520 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 21:00:33,520 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-13 21:00:33,520 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-13 21:00:33,521 INFO || App info kafka.consumer for connector-consumer-vk_nau55_sink-0 unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-13 21:00:37,291 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,295 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,295 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,297 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink/status HTTP/1.1" 200 2228 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,297 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,300 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,300 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,301 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,302 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,303 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink/tasks/0/status HTTP/1.1" 200 2117 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,304 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,305 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:37,306 INFO || 10.0.2.5 - - [13/Sep/2023:18:00:37 +0000] "GET /connectors/vk_nau55_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-13 21:00:39,128 INFO || WorkerSourceTask{id=vk_nau55_src-0} Committing offsets for 3 acknowledged messages [org.apache.kafka.connect.runtime.WorkerSourceTask] 2023-09-13 21:00:50,762 INFO Oracle|vk_nau55|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource]