2023-09-21 23:54:04,534 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,539 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,539 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,541 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,541 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,544 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,544 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,546 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,546 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,548 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,548 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,550 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:04,550 INFO || 10.0.2.5 - - [21/Sep/2023:20:54:04 +0000] "GET /connectors/vk_nau58_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:54:30,455 INFO || [AdminClient clientId=naument--shared-admin] Node 2 disconnected. [org.apache.kafka.clients.NetworkClient] 2023-09-21 23:55:24,129 INFO Oracle|vk_nau58|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:55:45,784 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,788 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,788 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,791 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,791 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,793 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,794 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,796 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,796 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,798 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,799 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,801 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:45,801 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:45 +0000] "GET /connectors/vk_nau58_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:52,764 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:52 +0000] "GET /connectors/vk_nau58_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:52,764 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:52 +0000] "GET /connectors/vk_nau58_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:52,766 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:52 +0000] "GET /connectors/vk_nau58_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:52,766 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:52 +0000] "GET /connectors/vk_nau58_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,807 INFO || Successfully processed removal of connector 'vk_nau58_sink' [org.apache.kafka.connect.storage.KafkaConfigBackingStore] 2023-09-21 23:55:58,807 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau58_sink config removed [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,807 INFO || [Worker clientId=connect-1, groupId=naument] Handling connector-only config update by stopping connector vk_nau58_sink [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,807 INFO || Stopping connector vk_nau58_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:55:58,807 INFO || Scheduled shutdown for WorkerConnector{id=vk_nau58_sink} [org.apache.kafka.connect.runtime.WorkerConnector] 2023-09-21 23:55:58,808 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "DELETE /connectors/vk_nau58_sink HTTP/1.1" 204 0 "-" "ReactorNetty/1.1.6" 7 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,808 INFO || Completed shutdown for WorkerConnector{id=vk_nau58_sink} [org.apache.kafka.connect.runtime.WorkerConnector] 2023-09-21 23:55:58,809 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,809 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,810 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=171, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,813 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=171, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,814 INFO || Stopping connector vk_nau58_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:55:58,814 WARN || Ignoring stop request for unowned connector vk_nau58_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:55:58,814 INFO || Stopping task vk_nau58_sink-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:55:58,814 WARN || Ignoring await stop request for non-present connector vk_nau58_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:55:58,814 INFO || Closing session. [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-21 23:55:58,814 INFO || Closing the session factory [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-21 23:55:58,817 INFO || [Consumer clientId=connector-consumer-vk_nau58_sink-0, groupId=connect-vk_nau58_sink] Revoke previously assigned partitions vk_nau58.DEBEZIUM.GBC_TBL_SERVICECALL_NC58-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:55:58,817 INFO || [Consumer clientId=connector-consumer-vk_nau58_sink-0, groupId=connect-vk_nau58_sink] Member connector-consumer-vk_nau58_sink-0-81a26bd5-9803-4213-ae20-14c978d11c45 sending LeaveGroup request to coordinator broker3:29092 (id: 2147483644 rack: null) due to the consumer is being closed [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:55:58,817 INFO || [Consumer clientId=connector-consumer-vk_nau58_sink-0, groupId=connect-vk_nau58_sink] Resetting generation and member id due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:55:58,817 INFO || [Consumer clientId=connector-consumer-vk_nau58_sink-0, groupId=connect-vk_nau58_sink] Request joining group due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:55:58,829 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,864 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:55:58,864 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:55:58,864 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:55:58,866 INFO || App info kafka.consumer for connector-consumer-vk_nau58_sink-0 unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:55:58,867 INFO || [Worker clientId=connect-1, groupId=naument] Finished stopping tasks in preparation for rebalance [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,868 INFO || [Worker clientId=connect-1, groupId=naument] Finished flushing status backing store in preparation for rebalance [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,868 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 171 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3210, connectorIds=[vk_nau58_src], taskIds=[vk_nau58_src-0], revokedConnectorIds=[vk_nau58_sink], revokedTaskIds=[vk_nau58_sink-0], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,869 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3210 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,869 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,869 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,869 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,870 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=172, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,872 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=172, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:55:58,872 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 172 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3210, connectorIds=[vk_nau58_src], taskIds=[vk_nau58_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,872 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3210 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,872 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:55:58,873 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 43 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,876 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,879 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,881 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,884 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:55:58,886 INFO || 10.0.2.5 - - [21/Sep/2023:20:55:58 +0000] "GET /connectors/vk_nau58_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:01,925 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:01 +0000] "GET /connectors/vk_nau58_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:01,925 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:01 +0000] "GET /connectors/vk_nau58_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:01,928 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:01 +0000] "GET /connectors/vk_nau58_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:01,928 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:01 +0000] "GET /connectors/vk_nau58_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:05,662 INFO || Successfully processed removal of connector 'vk_nau58_src' [org.apache.kafka.connect.storage.KafkaConfigBackingStore] 2023-09-21 23:56:05,662 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau58_src config removed [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:05,663 INFO || [Worker clientId=connect-1, groupId=naument] Handling connector-only config update by stopping connector vk_nau58_src [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:05,663 INFO || Stopping connector vk_nau58_src [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:05,663 INFO || Scheduled shutdown for WorkerConnector{id=vk_nau58_src} [org.apache.kafka.connect.runtime.WorkerConnector] 2023-09-21 23:56:05,663 INFO || Completed shutdown for WorkerConnector{id=vk_nau58_src} [org.apache.kafka.connect.runtime.WorkerConnector] 2023-09-21 23:56:05,664 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:05,664 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:05,664 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:05 +0000] "DELETE /connectors/vk_nau58_src HTTP/1.1" 204 0 "-" "ReactorNetty/1.1.6" 6 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:05,665 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=173, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:05,667 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=173, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:05,667 INFO || Stopping connector vk_nau58_src [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:05,667 WARN || Ignoring stop request for unowned connector vk_nau58_src [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:05,667 INFO || Stopping task vk_nau58_src-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:05,667 WARN || Ignoring await stop request for non-present connector vk_nau58_src [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:05,683 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:05 +0000] "GET /connectors HTTP/1.1" 200 2 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:05,897 INFO Oracle|vk_nau58|snapshot Stopping down connector [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:07,301 INFO Oracle|vk_nau58|streaming startScn=290400852928, endScn=290400872928 [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:07,301 INFO Oracle|vk_nau58|streaming Streaming metrics dump: OracleStreamingChangeEventSourceMetrics{currentScn=290400875157, oldestScn=290400852929, committedScn=290400872926, offsetScn=290400852928, oldestScnChangeTime=2023-09-21T23:54:29Z, logMinerQueryCount=748, totalProcessedRows=1987500, totalCapturedDmlCount=18, totalDurationOfFetchingQuery=PT16M6.436918S, lastCapturedDmlCount=0, lastDurationOfFetchingQuery=PT7.386018S, maxCapturedDmlCount=3, maxDurationOfFetchingQuery=PT9.328084S, totalBatchProcessingDuration=PT26M31.403871S, lastBatchProcessingDuration=PT8.150656S, maxBatchProcessingThroughput=7, currentLogFileName=[/u04/oradata/naument1/redo_05b.log, /u02/oradata/naument1/redo_05a.log], minLogFilesMined=2, maxLogFilesMined=2, redoLogStatus=[/u02/oradata/naument1/redo_07a.log | ACTIVE, /u04/oradata/naument1/redo_07b.log | ACTIVE, /u02/oradata/naument1/redo_06a.log | ACTIVE, /u04/oradata/naument1/redo_06b.log | ACTIVE, /u02/oradata/naument1/redo_04a.log | ACTIVE, /u02/oradata/naument1/redo_01a.log | ACTIVE, /u04/oradata/naument1/redo_01b.log | ACTIVE, /u04/oradata/naument1/redo_04b.log | ACTIVE, /u02/oradata/naument1/redo_03a.log | ACTIVE, /u04/oradata/naument1/redo_03b.log | ACTIVE, /u02/oradata/naument1/redo_02a.log | ACTIVE, /u04/oradata/naument1/redo_02b.log | ACTIVE, /u04/oradata/naument1/redo_05b.log | CURRENT, /u02/oradata/naument1/redo_05a.log | CURRENT], switchCounter=0, batchSize=20000, millisecondToSleepBetweenMiningQuery=2800, keepTransactionsDuration=PT0S, networkConnectionProblemsCounter0, batchSizeDefault=20000, batchSizeMin=1000, batchSizeMax=100000, sleepTimeDefault=1000, sleepTimeMin=0, sleepTimeMax=3000, sleepTimeIncrement=200, totalParseTime=PT0.001381S, totalStartLogMiningSessionDuration=PT7M22.669627S, lastStartLogMiningSessionDuration=PT0.012812S, maxStartLogMiningSessionDuration=PT17.952802S, totalProcessTime=PT26M31.403871S, minBatchProcessTime=PT0.100254S, maxBatchProcessTime=PT21.880813S, totalResultSetNextTime=PT1M43.981729S, lagFromTheSource=DurationPT20.489185S, maxLagFromTheSourceDuration=PT34.832005S, minLagFromTheSourceDuration=PT0.230816S, lastCommitDuration=PT0.000001S, maxCommitDuration=PT0.020564S, activeTransactions=2, rolledBackTransactions=2407, oversizedTransactions=0, committedTransactions=232129, abandonedTransactionIds={}, rolledbackTransactionIds={090008001cd33500=090008001cd33500, 0c0004008b922c00=0c0004008b922c00, 04000400a4bd3000=04000400a4bd3000, 07001a002aae3300=07001a002aae3300, 0600040092642e00=0600040092642e00, 0e000a0019842a00=0e000a0019842a00, 0b00210085693300=0b00210085693300, 0b001a00ab653300=0b001a00ab653300, 0400140054b93000=0400140054b93000, 0600200074672e00=0600200074672e00}, registeredDmlCount=13, committedDmlCount=5, errorCount=0, warningCount=0, scnFreezeCount=0, unparsableDdlCount=0, miningSessionUserGlobalAreaMemory=24210056, miningSessionUserGlobalAreaMaxMemory=36654872, miningSessionProcessGlobalAreaMemory=91863880, miningSessionProcessGlobalAreaMaxMemory=118012744} [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:07,301 INFO Oracle|vk_nau58|streaming Offsets: OracleOffsetContext [scn=290400852928, commit_scn=["290400872926:1:030006002fa02e00"]] [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:07,301 INFO Oracle|vk_nau58|streaming Finished streaming [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:07,301 INFO Oracle|vk_nau58|streaming Connected metrics set to 'false' [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:07,302 INFO Oracle|vk_nau58|snapshot SignalProcessor stopped [io.debezium.pipeline.signal.SignalProcessor] 2023-09-21 23:56:07,323 INFO Oracle|vk_nau58|snapshot Connection gracefully closed [io.debezium.jdbc.JdbcConnection] 2023-09-21 23:56:07,324 INFO Oracle|vk_nau58|snapshot [Producer clientId=vk_nau58-schemahistory] Closing the Kafka producer with timeoutMillis = 30000 ms. [org.apache.kafka.clients.producer.KafkaProducer] 2023-09-21 23:56:07,325 INFO Oracle|vk_nau58|snapshot Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,325 INFO Oracle|vk_nau58|snapshot Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,325 INFO Oracle|vk_nau58|snapshot Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,325 INFO Oracle|vk_nau58|snapshot App info kafka.producer for vk_nau58-schemahistory unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:07,325 INFO Oracle|vk_nau58|snapshot [Producer clientId=connector-producer-vk_nau58_src-0] Closing the Kafka producer with timeoutMillis = 30000 ms. [org.apache.kafka.clients.producer.KafkaProducer] 2023-09-21 23:56:07,327 INFO Oracle|vk_nau58|snapshot Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,327 INFO Oracle|vk_nau58|snapshot Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,327 INFO Oracle|vk_nau58|snapshot Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,327 INFO Oracle|vk_nau58|snapshot App info kafka.producer for connector-producer-vk_nau58_src-0 unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:07,327 INFO || App info kafka.admin.client for connector-adminclient-vk_nau58_src-0 unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:07,328 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,328 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,328 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:07,330 INFO || [Worker clientId=connect-1, groupId=naument] Finished stopping tasks in preparation for rebalance [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,332 INFO || [Worker clientId=connect-1, groupId=naument] Finished flushing status backing store in preparation for rebalance [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,332 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 173 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3212, connectorIds=[], taskIds=[], revokedConnectorIds=[vk_nau58_src], revokedTaskIds=[vk_nau58_src-0], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,333 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3212 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,333 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,333 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:07,333 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:07,334 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=174, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:07,335 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=174, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:07,335 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 174 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3212, connectorIds=[], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,335 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3212 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:07,335 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,417 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors HTTP/1.1" 200 2 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,421 INFO || Loading the custom source info struct maker plugin: io.debezium.connector.oracle.OracleSourceInfoStructMaker [io.debezium.config.CommonConnectorConfig] 2023-09-21 23:56:13,483 INFO || Database Version: Oracle Database 12c Enterprise Edition Release 12.2.0.1.0 - 64bit Production [io.debezium.connector.oracle.OracleConnection] 2023-09-21 23:56:13,484 INFO || Connection gracefully closed [io.debezium.jdbc.JdbcConnection] 2023-09-21 23:56:13,485 INFO || AbstractConfig values: [org.apache.kafka.common.config.AbstractConfig] 2023-09-21 23:56:13,487 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau59_src config updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,488 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,488 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,489 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=175, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,490 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "POST /connectors HTTP/1.1" 201 1491 "-" "ReactorNetty/1.1.6" 72 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,491 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=175, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,491 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 175 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3213, connectorIds=[vk_nau59_src], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,491 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3213 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,491 INFO || [Worker clientId=connect-1, groupId=naument] Starting connector vk_nau59_src [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,491 INFO || Creating connector vk_nau59_src of type io.debezium.connector.oracle.OracleConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,492 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-21 23:56:13,492 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,492 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-21 23:56:13,492 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,492 INFO || Instantiated connector vk_nau59_src with version 2.4.0.Beta2 of type class io.debezium.connector.oracle.OracleConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,492 INFO || Finished creating connector vk_nau59_src [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,492 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,493 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-21 23:56:13,493 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,494 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-21 23:56:13,494 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1491 "-" "ReactorNetty/1.1.6" 4 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,494 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,496 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 404 73 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,501 INFO || [Worker clientId=connect-1, groupId=naument] Tasks [vk_nau59_src-0] configs updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,502 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,502 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,503 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=176, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,505 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=176, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:13,505 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 176 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3215, connectorIds=[vk_nau59_src], taskIds=[vk_nau59_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,505 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3215 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,505 INFO || [Worker clientId=connect-1, groupId=naument] Starting task vk_nau59_src-0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,505 INFO || Creating task vk_nau59_src-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,505 INFO || ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig] 2023-09-21 23:56:13,505 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,506 INFO || TaskConfig values: task.class = class io.debezium.connector.oracle.OracleConnectorTask [org.apache.kafka.connect.runtime.TaskConfig] 2023-09-21 23:56:13,506 INFO || Instantiated task vk_nau59_src-0 with version 2.4.0.Beta2 of type io.debezium.connector.oracle.OracleConnectorTask [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,506 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-21 23:56:13,507 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-21 23:56:13,508 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-21 23:56:13,508 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-21 23:56:13,508 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-21 23:56:13,509 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-21 23:56:13,509 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-21 23:56:13,509 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-21 23:56:13,509 INFO || Set up the key converter class io.confluent.connect.avro.AvroConverter for task vk_nau59_src-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,509 INFO || Set up the value converter class io.confluent.connect.avro.AvroConverter for task vk_nau59_src-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,509 INFO || Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task vk_nau59_src-0 using the worker config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,510 INFO || SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig] 2023-09-21 23:56:13,510 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,510 INFO || EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig] 2023-09-21 23:56:13,510 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.oracle.OracleConnector errors.log.enable = true errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none exactly.once.support = requested header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_src offsets.storage.topic = null predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 1 topic.creation.default.replication.factor = 1 topic.creation.groups = [] transaction.boundary = poll transaction.boundary.interval.ms = null transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:13,511 INFO || Initializing: org.apache.kafka.connect.runtime.TransformationChain{} [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:13,511 INFO || ProducerConfig values: acks = -1 auto.include.jmx.reporter = true batch.size = 16384 bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = connector-producer-vk_nau59_src-0 compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 9223372036854775807 max.in.flight.requests.per.connection = 1 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-21 23:56:13,513 INFO || These configurations '[metrics.context.connect.kafka.cluster.id, metrics.context.connect.group.id]' were supplied but are not used yet. [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-21 23:56:13,513 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,513 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,513 INFO || Kafka startTimeMs: 1695329773513 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,513 INFO || AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] client.dns.lookup = use_all_dns_ips client.id = connector-adminclient-vk_nau59_src-0 connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-21 23:56:13,515 INFO || These configurations '[group.id, max.partition.fetch.bytes, rest.advertised.port, task.shutdown.graceful.timeout.ms, plugin.path, metrics.context.connect.kafka.cluster.id, status.storage.replication.factor, offset.storage.topic, value.converter, key.converter, message.max.bytes, config.storage.topic, metrics.context.connect.group.id, rest.advertised.host.name, status.storage.topic, rest.host.name, offset.flush.timeout.ms, config.storage.replication.factor, offset.flush.interval.ms, rest.port, key.converter.schemas.enable, max.request.size, replica.fetch.max.bytes, value.converter.schemas.enable, offset.storage.replication.factor]' were supplied but are not used yet. [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-21 23:56:13,515 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,515 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,515 INFO || Kafka startTimeMs: 1695329773515 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,516 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:13,517 INFO || [Producer clientId=connector-producer-vk_nau59_src-0] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-21 23:56:13,517 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 6 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,519 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 8 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,521 INFO || Starting OracleConnectorTask with configuration: [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || connector.class = io.debezium.connector.oracle.OracleConnector [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || topic.creation.default.partitions = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || tasks.max = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 404 70 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,521 INFO || schema.history.internal.store.only.captured.tables.ddl = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || schema.history.internal.store.only.captured.databases.ddl = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || include.schema.changes = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || topic.prefix = vk_nau59 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || decimal.handling.mode = precise [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || schema.history.internal.kafka.topic = vk_nau59_src.schema-changes [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || topic.creation.default.include = vk_nau59\.* [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || topic.creation.default.replication.factor = 1 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || lob.enabled = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || value.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || errors.log.enable = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,521 INFO || key.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || snapshot.lock.timeout.ms = 5000 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.user = debezium [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.dbname = NAUMENT1 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || datatype.propagate.source.type = .* [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || topic.creation.default.compression.type = lz4 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.connection.adapter = logminer [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || schema.history.internal.kafka.bootstrap.servers = broker1:29092,broker3:29092,broker3:29092 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || topic.creation.default.retention.ms = 432000000 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.port = 1521 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || topic.creation.enable = true [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || value.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || log.mining.session.max.ms = 120000 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || task.class = io.debezium.connector.oracle.OracleConnectorTask [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.hostname = naumen-db-test.rgs.ru [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || database.password = ******** [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || name = vk_nau59_src [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || table.include.list = DEBEZIUM.GBC_TBL_SERVICECALL_NC59 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || key.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || snapshot.mode = always [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,522 INFO || Loading the custom source info struct maker plugin: io.debezium.connector.oracle.OracleSourceInfoStructMaker [io.debezium.config.CommonConnectorConfig] 2023-09-21 23:56:13,522 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:13 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 112 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:13,522 INFO || Loading the custom topic naming strategy plugin: io.debezium.schema.SchemaTopicNamingStrategy [io.debezium.config.CommonConnectorConfig] 2023-09-21 23:56:13,577 INFO || Database Version: Oracle Database 12c Enterprise Edition Release 12.2.0.1.0 - 64bit Production [io.debezium.connector.oracle.OracleConnection] 2023-09-21 23:56:13,587 INFO || KafkaSchemaHistory Consumer config: {key.deserializer=org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=org.apache.kafka.common.serialization.StringDeserializer, enable.auto.commit=false, group.id=vk_nau59-schemahistory, bootstrap.servers=broker1:29092,broker3:29092,broker3:29092, fetch.min.bytes=1, session.timeout.ms=10000, auto.offset.reset=earliest, client.id=vk_nau59-schemahistory} [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-21 23:56:13,587 INFO || KafkaSchemaHistory Producer config: {retries=1, value.serializer=org.apache.kafka.common.serialization.StringSerializer, acks=1, batch.size=32768, max.block.ms=10000, bootstrap.servers=broker1:29092,broker3:29092,broker3:29092, buffer.memory=1048576, key.serializer=org.apache.kafka.common.serialization.StringSerializer, client.id=vk_nau59-schemahistory, linger.ms=0} [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-21 23:56:13,587 INFO || Requested thread factory for connector OracleConnector, id = vk_nau59 named = db-history-config-check [io.debezium.util.Threads] 2023-09-21 23:56:13,587 INFO || Idempotence will be disabled because acks is set to 1, not set to 'all'. [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-21 23:56:13,587 INFO || ProducerConfig values: acks = 1 auto.include.jmx.reporter = true batch.size = 32768 bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] buffer.memory = 1048576 client.dns.lookup = use_all_dns_ips client.id = vk_nau59-schemahistory compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 10000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.adaptive.partitioning.enable = true partitioner.availability.timeout.ms = 0 partitioner.class = null partitioner.ignore.keys = false receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.StringSerializer [org.apache.kafka.clients.producer.ProducerConfig] 2023-09-21 23:56:13,589 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,589 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,589 INFO || Kafka startTimeMs: 1695329773589 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,591 INFO || [Producer clientId=vk_nau59-schemahistory] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-21 23:56:13,593 INFO || No previous offsets found [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:13,593 INFO || Connector started for the first time, database schema history recovery will not be executed [io.debezium.connector.oracle.OracleConnectorTask] 2023-09-21 23:56:13,593 INFO || ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = vk_nau59-schemahistory client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = vk_nau59-schemahistory group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 10000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-21 23:56:13,595 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,595 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,595 INFO || Kafka startTimeMs: 1695329773595 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,596 INFO || [Consumer clientId=vk_nau59-schemahistory, groupId=vk_nau59-schemahistory] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-21 23:56:13,598 INFO || [Consumer clientId=vk_nau59-schemahistory, groupId=vk_nau59-schemahistory] Resetting generation and member id due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:13,598 INFO || [Consumer clientId=vk_nau59-schemahistory, groupId=vk_nau59-schemahistory] Request joining group due to: consumer pro-actively leaving the group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:13,598 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,598 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,598 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,599 INFO || App info kafka.consumer for vk_nau59-schemahistory unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,599 INFO || AdminClientConfig values: auto.include.jmx.reporter = true bootstrap.servers = [broker1:29092, broker3:29092, broker3:29092] client.dns.lookup = use_all_dns_ips client.id = vk_nau59-schemahistory connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 1 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-21 23:56:13,600 INFO || These configurations '[value.serializer, acks, batch.size, max.block.ms, buffer.memory, key.serializer, linger.ms]' were supplied but are not used yet. [org.apache.kafka.clients.admin.AdminClientConfig] 2023-09-21 23:56:13,600 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,600 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,600 INFO || Kafka startTimeMs: 1695329773600 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,634 INFO || Database schema history topic '(name=vk_nau59_src.schema-changes, numPartitions=1, replicationFactor=default, replicasAssignments=null, configs={cleanup.policy=delete, retention.ms=9223372036854775807, retention.bytes=-1})' created [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2023-09-21 23:56:13,634 INFO || App info kafka.admin.client for vk_nau59-schemahistory unregistered [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:13,635 INFO || Metrics scheduler closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,635 INFO || Closing reporter org.apache.kafka.common.metrics.JmxReporter [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,635 INFO || Metrics reporters closed [org.apache.kafka.common.metrics.Metrics] 2023-09-21 23:56:13,636 INFO || Requested thread factory for connector OracleConnector, id = vk_nau59 named = SignalProcessor [io.debezium.util.Threads] 2023-09-21 23:56:13,636 INFO || Requested thread factory for connector OracleConnector, id = vk_nau59 named = change-event-source-coordinator [io.debezium.util.Threads] 2023-09-21 23:56:13,636 INFO || Requested thread factory for connector OracleConnector, id = vk_nau59 named = blocking-snapshot [io.debezium.util.Threads] 2023-09-21 23:56:13,636 INFO || Creating thread debezium-oracleconnector-vk_nau59-change-event-source-coordinator [io.debezium.util.Threads] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot Metrics registered [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot Context created [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot Snapshot mode is set to ALWAYS, not checking exiting offset. [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot According to the connector configuration both schema and data will be snapshot. [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-21 23:56:13,637 INFO || SignalProcessor started. Scheduling it every 5000ms [io.debezium.pipeline.signal.SignalProcessor] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot Snapshot step 1 - Preparing [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:13,637 INFO || Creating thread debezium-oracleconnector-vk_nau59-SignalProcessor [io.debezium.util.Threads] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot Snapshot step 2 - Determining captured tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:13,637 INFO Oracle|vk_nau59|snapshot WorkerSourceTask{id=vk_nau59_src-0} Source task finished initialization and start [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:14,847 INFO Oracle|vk_nau59|snapshot Adding table NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 to the list of capture schema tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:14,852 INFO Oracle|vk_nau59|snapshot Created connection pool with 1 threads [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:14,852 INFO Oracle|vk_nau59|snapshot Snapshot step 3 - Locking captured tables [NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59] [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:14,855 INFO Oracle|vk_nau59|snapshot Snapshot step 4 - Determining snapshot offset [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:14,974 INFO Oracle|vk_nau59|snapshot No in-progress transactions will be captured. [io.debezium.connector.oracle.logminer.LogMinerAdapter] 2023-09-21 23:56:14,976 INFO Oracle|vk_nau59|snapshot Connection gracefully closed [io.debezium.jdbc.JdbcConnection] 2023-09-21 23:56:14,976 INFO Oracle|vk_nau59|snapshot Snapshot step 5 - Reading structure of captured tables [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:14,976 INFO Oracle|vk_nau59|snapshot Only captured tables schema should be captured, capturing: [NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59] [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2023-09-21 23:56:15,629 INFO Oracle|vk_nau59|snapshot Snapshot step 6 - Persisting schema history [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:15,629 INFO Oracle|vk_nau59|snapshot Capturing structure of table NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:15,792 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:15 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:15,792 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:15 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:15,794 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:15 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:15,797 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:15 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:16,705 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:16 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:16,706 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:16 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:16,708 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:16 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:16,709 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:16 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:17,727 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:17 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:17,727 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:17 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:17,730 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:17 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:17,730 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:17 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:18,769 WARN Oracle|vk_nau59|snapshot The Kafka Connect schema name 'ATTR$ATTR' is not a valid Avro schema name, so replacing with 'ATTR_ATTR' [io.debezium.schema.SchemaNameAdjuster] 2023-09-21 23:56:18,774 INFO Oracle|vk_nau59|snapshot Already applied 1 database changes [io.debezium.relational.history.SchemaHistoryMetrics] 2023-09-21 23:56:18,776 INFO Oracle|vk_nau59|snapshot Snapshot step 7 - Snapshotting data [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:18,776 INFO Oracle|vk_nau59|snapshot Creating snapshot worker pool with 1 worker thread(s) [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:18,776 INFO Oracle|vk_nau59|snapshot For table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' using select statement: 'SELECT "ID", "CREATION_DATE", "CLAIM_TRANSFERDATE", "TITLE", "CLIENT_EMAIL", "CLAIM_SUMRETURN", "ATTR$ATTR" FROM "DEBEZIUM"."GBC_TBL_SERVICECALL_NC59" AS OF SCN 290400875236' [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:18,776 INFO Oracle|vk_nau59|snapshot Exporting data from table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' (1 of 1 tables) [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:18,790 INFO Oracle|vk_nau59|snapshot Finished exporting 3 records for table 'NAUMENT1.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' (1 of 1 tables); total duration '00:00:00.014' [io.debezium.relational.RelationalSnapshotChangeEventSource] 2023-09-21 23:56:18,791 INFO Oracle|vk_nau59|snapshot Snapshot - Final stage [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2023-09-21 23:56:18,791 INFO Oracle|vk_nau59|snapshot Snapshot completed [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2023-09-21 23:56:18,791 INFO Oracle|vk_nau59|snapshot Snapshot ended with SnapshotResult [status=COMPLETED, offset=OracleOffsetContext [scn=290400875236, commit_scn=[]]] [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:18,791 INFO Oracle|vk_nau59|streaming Connected metrics set to 'true' [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:18,791 INFO Oracle|vk_nau59|streaming Starting streaming [io.debezium.pipeline.ChangeEventSourceCoordinator] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Redo Log Group Sizes: [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #1: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #2: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #3: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #4: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #5: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #6: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,071 INFO Oracle|vk_nau59|streaming Group #7: 536870912 bytes [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-21 23:56:19,149 INFO Oracle|vk_nau59|snapshot 4 records sent during previous 00:00:05.642, last recorded offset of {server=vk_nau59} partition is {snapshot_scn=290400875236, snapshot=true, scn=290400875236, snapshot_completed=true} [io.debezium.connector.common.BaseSourceTask] 2023-09-21 23:56:19,169 INFO Oracle|vk_nau59|snapshot The task will send records to topic 'vk_nau59' for the first time. Checking whether topic exists [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,172 INFO Oracle|vk_nau59|snapshot Creating topic 'vk_nau59' [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,203 INFO Oracle|vk_nau59|snapshot Created topic (name=vk_nau59, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000}) on brokers at broker1:29092,broker2:29092,broker3:29092 [org.apache.kafka.connect.util.TopicAdmin] 2023-09-21 23:56:19,203 INFO Oracle|vk_nau59|snapshot Created topic '(name=vk_nau59, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000})' using creation group TopicCreationGroup{name='default', inclusionPattern=.*, exclusionPattern=, numPartitions=1, replicationFactor=1, otherConfigs={compression.type=lz4, retention.ms=432000000}} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,208 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Got error produce response with correlation id 5 on topic-partition vk_nau59-0, retrying (2147483646 attempts left). Error: UNKNOWN_TOPIC_OR_PARTITION [org.apache.kafka.clients.producer.internals.Sender] 2023-09-21 23:56:19,208 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Received unknown topic or partition error in produce request on partition vk_nau59-0. The topic-partition may not exist or the user may not have Describe access to it [org.apache.kafka.clients.producer.internals.Sender] 2023-09-21 23:56:19,230 INFO Oracle|vk_nau59|snapshot The task will send records to topic 'vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' for the first time. Checking whether topic exists [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,231 INFO Oracle|vk_nau59|snapshot Creating topic 'vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,261 INFO Oracle|vk_nau59|snapshot Created topic (name=vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000}) on brokers at broker1:29092,broker2:29092,broker3:29092 [org.apache.kafka.connect.util.TopicAdmin] 2023-09-21 23:56:19,261 INFO Oracle|vk_nau59|snapshot Created topic '(name=vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={compression.type=lz4, retention.ms=432000000})' using creation group TopicCreationGroup{name='default', inclusionPattern=.*, exclusionPattern=, numPartitions=1, replicationFactor=1, otherConfigs={compression.type=lz4, retention.ms=432000000}} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2023-09-21 23:56:19,262 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Error while fetching metadata with correlation id 6 : {vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-21 23:56:19,363 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Error while fetching metadata with correlation id 8 : {vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59=UNKNOWN_TOPIC_OR_PARTITION} [org.apache.kafka.clients.NetworkClient] 2023-09-21 23:56:19,466 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Got error produce response with correlation id 11 on topic-partition vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0, retrying (2147483646 attempts left). Error: UNKNOWN_TOPIC_OR_PARTITION [org.apache.kafka.clients.producer.internals.Sender] 2023-09-21 23:56:19,466 WARN || [Producer clientId=connector-producer-vk_nau59_src-0] Received unknown topic or partition error in produce request on partition vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0. The topic-partition may not exist or the user may not have Describe access to it [org.apache.kafka.clients.producer.internals.Sender] 2023-09-21 23:56:19,828 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,831 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,833 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,836 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,838 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,840 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:19,842 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:19 +0000] "GET /connectors/vk_nau59_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,436 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors HTTP/1.1" 200 16 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,440 INFO || AbstractConfig values: [org.apache.kafka.common.config.AbstractConfig] 2023-09-21 23:56:42,442 INFO || [Worker clientId=connect-1, groupId=naument] Connector vk_nau59_sink config updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,443 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,443 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,444 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "POST /connectors HTTP/1.1" 201 868 "-" "ReactorNetty/1.1.6" 7 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,444 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=177, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,446 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=177, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,446 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 177 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3216, connectorIds=[vk_nau59_sink, vk_nau59_src], taskIds=[vk_nau59_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,446 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3216 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,447 INFO || [Worker clientId=connect-1, groupId=naument] Starting connector vk_nau59_sink [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,447 INFO || Creating connector vk_nau59_sink of type io.debezium.connector.jdbc.JdbcSinkConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,447 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-21 23:56:42,447 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:42,448 INFO || Instantiated connector vk_nau59_sink with version 2.4.0.Beta2 of type class io.debezium.connector.jdbc.JdbcSinkConnector [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,448 INFO || Finished creating connector vk_nau59_sink [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,448 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,449 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink HTTP/1.1" 200 868 "-" "ReactorNetty/1.1.6" 4 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,449 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-21 23:56:42,449 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:42,452 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink/status HTTP/1.1" 200 111 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,455 INFO || [Worker clientId=connect-1, groupId=naument] Tasks [vk_nau59_sink-0] configs updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,456 INFO || [Worker clientId=connect-1, groupId=naument] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,456 INFO || [Worker clientId=connect-1, groupId=naument] (Re-)joining group [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,457 INFO || [Worker clientId=connect-1, groupId=naument] Successfully joined group with generation Generation{generationId=178, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,459 INFO || [Worker clientId=connect-1, groupId=naument] Successfully synced group in generation Generation{generationId=178, memberId='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', protocol='sessioned'} [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2023-09-21 23:56:42,459 INFO || [Worker clientId=connect-1, groupId=naument] Joined group at generation 178 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-c7f67616-75b4-453b-a9b1-fc066d178211', leaderUrl='http://172.18.0.6:8083/', offset=3218, connectorIds=[vk_nau59_sink, vk_nau59_src], taskIds=[vk_nau59_sink-0, vk_nau59_src-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,459 INFO || [Worker clientId=connect-1, groupId=naument] Starting connectors and tasks using config offset 3218 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,459 INFO || [Worker clientId=connect-1, groupId=naument] Starting task vk_nau59_sink-0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,459 INFO || Creating task vk_nau59_sink-0 [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,460 INFO || ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig] 2023-09-21 23:56:42,460 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:42,460 INFO || TaskConfig values: task.class = class io.debezium.connector.jdbc.JdbcSinkConnectorTask [org.apache.kafka.connect.runtime.TaskConfig] 2023-09-21 23:56:42,460 INFO || Instantiated task vk_nau59_sink-0 with version 2.4.0.Beta2 of type io.debezium.connector.jdbc.JdbcSinkConnectorTask [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,460 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-21 23:56:42,460 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-21 23:56:42,461 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-21 23:56:42,461 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-21 23:56:42,462 INFO || AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.connect.avro.AvroConverterConfig] 2023-09-21 23:56:42,462 INFO || KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroSerializerConfig] 2023-09-21 23:56:42,462 INFO || KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = URL basic.auth.user.info = [hidden] bearer.auth.cache.expiry.buffer.seconds = 300 bearer.auth.client.id = null bearer.auth.client.secret = null bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.custom.provider.class = null bearer.auth.identity.pool.id = null bearer.auth.issuer.endpoint.url = null bearer.auth.logical.cluster = null bearer.auth.scope = null bearer.auth.scope.claim.name = scope bearer.auth.sub.claim.name = sub bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy http.connect.timeout.ms = 60000 http.read.timeout.ms = 60000 id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.cache.size = 1000 latest.cache.ttl.sec = -1 latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 rule.actions = [] rule.executors = [] rule.service.loader.enable = true schema.format = null schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [http://naument-sr:8081] specific.avro.key.type = null specific.avro.reader = false specific.avro.value.type = null use.latest.version = false use.latest.with.metadata = null use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy [io.confluent.kafka.serializers.KafkaAvroDeserializerConfig] 2023-09-21 23:56:42,463 INFO || AvroDataConfig values: allow.optional.map.keys = false connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false ignore.default.for.nullables = false schemas.cache.config = 1000 scrub.invalid.names = false [io.confluent.connect.avro.AvroDataConfig] 2023-09-21 23:56:42,463 INFO || Set up the key converter class io.confluent.connect.avro.AvroConverter for task vk_nau59_sink-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,463 INFO || Set up the value converter class io.confluent.connect.avro.AvroConverter for task vk_nau59_sink-0 using the connector config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,463 INFO || Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task vk_nau59_sink-0 using the worker config [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,463 INFO || Initializing: org.apache.kafka.connect.runtime.TransformationChain{} [org.apache.kafka.connect.runtime.Worker] 2023-09-21 23:56:42,463 INFO || SinkConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.SinkConnectorConfig] 2023-09-21 23:56:42,463 INFO || EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.jdbc.JdbcSinkConnector errors.deadletterqueue.context.headers.enable = false errors.deadletterqueue.topic.name = errors.deadletterqueue.topic.replication.factor = 3 errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = class io.confluent.connect.avro.AvroConverter name = vk_nau59_sink predicates = [] tasks.max = 1 topics = [] topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 transforms = [] value.converter = class io.confluent.connect.avro.AvroConverter [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] 2023-09-21 23:56:42,463 INFO || ConsumerConfig values: allow.auto.create.topics = true auto.commit.interval.ms = 5000 auto.include.jmx.reporter = true auto.offset.reset = earliest bootstrap.servers = [broker1:29092, broker2:29092, broker3:29092] check.crcs = true client.dns.lookup = use_all_dns_ips client.id = connector-consumer-vk_nau59_sink-0 client.rack = connections.max.idle.ms = 540000 default.api.timeout.ms = 60000 enable.auto.commit = false exclude.internal.topics = true fetch.max.bytes = 52428800 fetch.max.wait.ms = 500 fetch.min.bytes = 1 group.id = connect-vk_nau59_sink group.instance.id = null heartbeat.interval.ms = 3000 interceptor.classes = [] internal.leave.group.on.close = true internal.throw.on.fetch.stable.offset.unsupported = false isolation.level = read_uncommitted key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer max.partition.fetch.bytes = 1048576 max.poll.interval.ms = 300000 max.poll.records = 500 metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 session.timeout.ms = 45000 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-21 23:56:42,466 INFO || These configurations '[metrics.context.connect.kafka.cluster.id, metrics.context.connect.group.id]' were supplied but are not used yet. [org.apache.kafka.clients.consumer.ConsumerConfig] 2023-09-21 23:56:42,466 INFO || Kafka version: 3.5.1 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:42,466 INFO || Kafka commitId: 2c6fb6c54472e90a [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:42,466 INFO || Kafka startTimeMs: 1695329802466 [org.apache.kafka.common.utils.AppInfoParser] 2023-09-21 23:56:42,467 INFO || [Worker clientId=connect-1, groupId=naument] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] 2023-09-21 23:56:42,467 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Subscribed to pattern: 'vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59' [org.apache.kafka.clients.consumer.KafkaConsumer] 2023-09-21 23:56:42,468 INFO || Starting JdbcSinkConnectorConfig with configuration: [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || connector.class = io.debezium.connector.jdbc.JdbcSinkConnector [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || table.name.format = vk_nau59_tbl_servicecall [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || connection.password = ******** [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || primary.key.mode = record_key [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || tasks.max = 1 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || connection.username = debeziumt [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || quote.identifiers = false [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || topics.regex = vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || value.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || delete.enabled = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || schema.evolution = basic [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || auto.evolve = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || task.class = io.debezium.connector.jdbc.JdbcSinkConnectorTask [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || name = vk_nau59_sink [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || auto.create = true [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || connection.url = jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || value.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || insert.mode = upsert [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || key.converter.schema.registry.url = http://naument-sr:8081 [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || key.converter = io.confluent.connect.avro.AvroConverter [io.debezium.connector.jdbc.JdbcSinkConnectorTask] 2023-09-21 23:56:42,468 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,469 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,472 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink/tasks/0/status HTTP/1.1" 404 71 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,474 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:42 +0000] "GET /connectors/vk_nau59_sink/status HTTP/1.1" 200 111 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:42,478 INFO || HHH000130: Instantiating explicit connection provider: org.hibernate.c3p0.internal.C3P0ConnectionProvider [org.hibernate.engine.jdbc.connections.internal.ConnectionProviderInitiator] 2023-09-21 23:56:42,478 INFO || HHH010002: C3P0 using driver: null at URL: jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1 [org.hibernate.orm.connections.pooling.c3p0] 2023-09-21 23:56:42,478 INFO || HHH10001001: Connection properties: {password=****, user=debeziumt} [org.hibernate.orm.connections.pooling.c3p0] 2023-09-21 23:56:42,478 INFO || HHH10001003: Autocommit mode: false [org.hibernate.orm.connections.pooling.c3p0] 2023-09-21 23:56:42,478 WARN || HHH10001006: No JDBC Driver class was specified by property hibernate.connection.driver_class [org.hibernate.orm.connections.pooling.c3p0] 2023-09-21 23:56:42,502 INFO || HHH10001007: JDBC isolation level: [org.hibernate.orm.connections.pooling.c3p0] 2023-09-21 23:56:42,506 INFO || Initializing c3p0 pool... com.mchange.v2.c3p0.PoolBackedDataSource@6cf2305a [ connectionPoolDataSource -> com.mchange.v2.c3p0.WrapperConnectionPoolDataSource@9ee9f477 [ acquireIncrement -> 32, acquireRetryAttempts -> 30, acquireRetryDelay -> 1000, autoCommitOnClose -> false, automaticTestTable -> null, breakAfterAcquireFailure -> false, checkoutTimeout -> 0, connectionCustomizerClassName -> null, connectionTesterClassName -> com.mchange.v2.c3p0.impl.DefaultConnectionTester, contextClassLoaderSource -> caller, debugUnreturnedConnectionStackTraces -> false, factoryClassLocation -> null, forceIgnoreUnresolvedTransactions -> false, forceSynchronousCheckins -> false, identityToken -> 2rvybeay1bzctscet1stb|623abd8c, idleConnectionTestPeriod -> 0, initialPoolSize -> 5, maxAdministrativeTaskTime -> 0, maxConnectionAge -> 0, maxIdleTime -> 0, maxIdleTimeExcessConnections -> 0, maxPoolSize -> 32, maxStatements -> 0, maxStatementsPerConnection -> 0, minPoolSize -> 5, nestedDataSource -> com.mchange.v2.c3p0.DriverManagerDataSource@d1803e8c [ description -> null, driverClass -> null, factoryClassLocation -> null, forceUseNamedDriverClass -> false, identityToken -> 2rvybeay1bzctscet1stb|371b1a52, jdbcUrl -> jdbc:postgresql://dwh-db-test.rgs.ru:5438/db_ods_test?currentSchema=naument1, properties -> {password=******, user=******} ], preferredTestQuery -> null, privilegeSpawnedThreads -> false, propertyCycle -> 0, statementCacheNumDeferredCloseThreads -> 0, testConnectionOnCheckin -> false, testConnectionOnCheckout -> false, unreturnedConnectionTimeout -> 0, usesTraditionalReflectiveProxies -> false; userOverrides: {} ], dataSourceName -> null, extensions -> {}, factoryClassLocation -> null, identityToken -> 2rvybeay1bzctscet1stb|424c17de, numHelperThreads -> 3 ] [com.mchange.v2.c3p0.impl.AbstractPoolBackedDataSource] 2023-09-21 23:56:42,528 INFO || HHH000400: Using dialect: org.hibernate.dialect.PostgreSQLDialect [SQL dialect] 2023-09-21 23:56:42,543 INFO || HHH000490: Using JtaPlatform implementation: [org.hibernate.engine.transaction.jta.platform.internal.NoJtaPlatform] [org.hibernate.engine.transaction.jta.platform.internal.JtaPlatformInitiator] 2023-09-21 23:56:42,543 INFO || Using dialect io.debezium.connector.jdbc.dialect.postgres.PostgresDatabaseDialect [io.debezium.connector.jdbc.dialect.DatabaseDialectResolver] 2023-09-21 23:56:42,547 INFO || Database TimeZone: Europe/Moscow [io.debezium.connector.jdbc.dialect.GeneralDatabaseDialect] 2023-09-21 23:56:42,547 INFO || Database version 13.2.0 [io.debezium.connector.jdbc.JdbcChangeEventSink] 2023-09-21 23:56:42,547 INFO || WorkerSinkTask{id=vk_nau59_sink-0} Sink task finished initialization and start [org.apache.kafka.connect.runtime.WorkerSinkTask] 2023-09-21 23:56:42,548 INFO || WorkerSinkTask{id=vk_nau59_sink-0} Executing sink task [org.apache.kafka.connect.runtime.WorkerSinkTask] 2023-09-21 23:56:42,551 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Cluster ID: gVJjK6cZTd-nXsXP2EIHEQ [org.apache.kafka.clients.Metadata] 2023-09-21 23:56:42,551 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Discovered group coordinator broker3:29092 (id: 2147483644 rack: null) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,552 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] (Re-)joining group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,554 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Request joining group due to: need to re-join with the given member-id: connector-consumer-vk_nau59_sink-0-87f31521-2194-47de-857d-83c57a84a251 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,555 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,555 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] (Re-)joining group [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,557 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Successfully joined group with generation Generation{generationId=1, memberId='connector-consumer-vk_nau59_sink-0-87f31521-2194-47de-857d-83c57a84a251', protocol='range'} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,557 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Finished assignment for group at generation 1: {connector-consumer-vk_nau59_sink-0-87f31521-2194-47de-857d-83c57a84a251=Assignment(partitions=[vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0])} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,560 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Successfully synced group in generation Generation{generationId=1, memberId='connector-consumer-vk_nau59_sink-0-87f31521-2194-47de-857d-83c57a84a251', protocol='range'} [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,561 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Notifying assignor about the new Assignment(partitions=[vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0]) [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,561 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Adding newly assigned partitions: vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,561 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Found no committed offset for partition vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0 [org.apache.kafka.clients.consumer.internals.ConsumerCoordinator] 2023-09-21 23:56:42,562 INFO || [Consumer clientId=connector-consumer-vk_nau59_sink-0, groupId=connect-vk_nau59_sink] Resetting offset for partition vk_nau59.DEBEZIUM.GBC_TBL_SERVICECALL_NC59-0 to position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[broker3:29092 (id: 3 rack: null)], epoch=0}}. [org.apache.kafka.clients.consumer.internals.SubscriptionState] Hibernate: CREATE TABLE vk_nau59_tbl_servicecall (ID decimal(19,0) NOT NULL, CREATION_DATE timestamp(6) NOT NULL, CLAIM_TRANSFERDATE timestamp(6) NULL, TITLE varchar(4000) NULL, CLIENT_EMAIL varchar(255) NULL, CLAIM_SUMRETURN double precision NULL, "ATTR$ATTR" varchar(30) NULL, PRIMARY KEY(ID)) Hibernate: INSERT INTO naument1.vk_nau59_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN,"ATTR$ATTR") VALUES (?,?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN,"ATTR$ATTR"=EXCLUDED."ATTR$ATTR" Hibernate: INSERT INTO naument1.vk_nau59_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN,"ATTR$ATTR") VALUES (?,?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN,"ATTR$ATTR"=EXCLUDED."ATTR$ATTR" Hibernate: INSERT INTO naument1.vk_nau59_tbl_servicecall (ID,CREATION_DATE,CLAIM_TRANSFERDATE,TITLE,CLIENT_EMAIL,CLAIM_SUMRETURN,"ATTR$ATTR") VALUES (?,?,?,?,?,?,?) ON CONFLICT (ID) DO UPDATE SET CREATION_DATE=EXCLUDED.CREATION_DATE,CLAIM_TRANSFERDATE=EXCLUDED.CLAIM_TRANSFERDATE,TITLE=EXCLUDED.TITLE,CLIENT_EMAIL=EXCLUDED.CLIENT_EMAIL,CLAIM_SUMRETURN=EXCLUDED.CLAIM_SUMRETURN,"ATTR$ATTR"=EXCLUDED."ATTR$ATTR" 2023-09-21 23:56:43,762 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,765 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,765 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,767 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,767 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,769 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,770 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,772 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,773 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,775 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,777 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,777 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:56:43,779 INFO || 10.0.2.5 - - [21/Sep/2023:20:56:43 +0000] "GET /connectors/vk_nau59_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-21 23:57:13,516 INFO || WorkerSourceTask{id=vk_nau59_src-0} Committing offsets for 4 acknowledged messages [org.apache.kafka.connect.runtime.WorkerSourceTask] 2023-09-21 23:58:24,553 INFO Oracle|vk_nau59|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-22 00:00:29,442 INFO Oracle|vk_nau59|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-22 00:01:13,567 INFO || [AdminClient clientId=connector-adminclient-vk_nau59_src-0] Node -3 disconnected. [org.apache.kafka.clients.NetworkClient] 2023-09-22 00:02:32,465 INFO Oracle|vk_nau59|streaming LogMiner session has exceeded maximum session time of 'Optional[PT2M]', forcing restart. [io.debezium.connector.oracle.logminer.LogMinerStreamingChangeEventSource] 2023-09-22 00:03:18,175 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors HTTP/1.1" 200 32 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,179 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink HTTP/1.1" 200 906 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,179 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src HTTP/1.1" 200 1528 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,181 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src/status HTTP/1.1" 200 168 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,182 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink/status HTTP/1.1" 200 167 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,184 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src/config HTTP/1.1" 200 1431 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,185 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink/config HTTP/1.1" 200 809 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,187 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink/tasks HTTP/1.1" 200 930 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,189 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src/tasks HTTP/1.1" 200 1551 "-" "ReactorNetty/1.1.6" 3 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,189 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,191 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src/tasks/0/status HTTP/1.1" 200 56 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,192 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_sink/topics HTTP/1.1" 200 75 "-" "ReactorNetty/1.1.6" 2 [org.apache.kafka.connect.runtime.rest.RestServer] 2023-09-22 00:03:18,193 INFO || 10.0.2.5 - - [21/Sep/2023:21:03:18 +0000] "GET /connectors/vk_nau59_src/topics HTTP/1.1" 200 85 "-" "ReactorNetty/1.1.6" 1 [org.apache.kafka.connect.runtime.rest.RestServer]