"name" : "ESETTABLE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "ESETCOLUMN", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PREDLHS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 300, "position" : 4, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "NOEQPREDS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOLTPREDS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOGTPREDS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOLTEQPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOGTEQPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NONEQPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 10, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOISNLPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 11, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOISNNLPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 12, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOBETPREDS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 13, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NONVLPREDS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 14, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "NOLIKEPRS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 15, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,714 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:06,714 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,714 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,714 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,714 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,714 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5515) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=7421]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,714 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=7421, buffer=java.nio.HeapByteBuffer[pos=0 lim=7421 cap=7421]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,714 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,715 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5515): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11034, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,715 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5515 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,715 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11034. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:06,715 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,715 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11034 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,715 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246635,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPSETSTATS,scn=5535835,ssn=0},ts_ms=1666014246713,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPSETSTATS" ("ESETOWNER" VARCHAR2(32) NOT NULL ENABLE, "ESETTABLE" VARCHAR2(32) NOT NULL ENABLE, "ESETCOLUMN" VARCHAR2(32) NOT NULL ENABLE, "PREDLHS" VARCHAR2(300) NOT NULL ENABLE, "NOEQPREDS" NUMBER, "NOLTPREDS" NUMBER, "NOGTPREDS" NUMBER, "NOLTEQPRS" NUMBER, "NOGTEQPRS" NUMBER, "NONEQPRS" NUMBER, "NOISNLPRS" NUMBER, "NOISNNLPRS" NUMBER, "NOBETPREDS" NUMBER, "NONVLPREDS" NUMBER, "NOLIKEPRS" NUMBER, CONSTRAINT "STATS_PKEY" PRIMARY KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN", "PREDLHS") ENABLE, CONSTRAINT "REF_STATS_KEY" FOREIGN KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN") REFERENCES "EXFSYS"."EXF$EXPRSET" ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPSETSTATS",table=Struct{primaryKeyColumnNames=[ESETOWNER, ESETTABLE, ESETCOLUMN, PREDLHS],columns=[Struct{name=ESETOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETTABLE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETCOLUMN,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PREDLHS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=NOEQPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=NONEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=NOBETPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=NONVLPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLIKEPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:06,715 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$IDXSECOBJ [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:06,715 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$IDXSECOBJ' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-2, correlationId=945): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=833240796, responses=[]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 833240796 with response=(), implied=(dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-0 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-2 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-4 at position FetchPosition{offset=1, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-1 at position FetchPosition{offset=3, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-3 at position FetchPosition{offset=3, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Built incremental fetch (sessionId=833240796, epoch=939) for node 1. Added (), altered (), removed (), replaced () out of (dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(), toForget=(), toReplace=(), implied=(dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Polling for fetches with timeout 2147024153 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:06,727 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-2, correlationId=946) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=833240796, sessionEpoch=939, topics=[], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-2 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-4 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-1 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-3 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,727 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Polling for fetches with timeout 2147024153 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:06,756 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,756 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,757 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,757 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$IDXSECOBJ','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,775 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,776 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#" NUMBER NOT NULL ENABLE, "IDXOWNER" VARCHAR2(32) NOT NULL ENABLE, "IDXNAME" VARCHAR2(25) NOT NULL ENABLE, "IDXATTRSET" VARCHAR2(32) NOT NULL ENABLE, "IDXESETTAB" VARCHAR2(32) NOT NULL ENABLE, "IDXESETCOL" VARCHAR2(32) NOT NULL ENABLE, "IDXPREDTAB" VARCHAR2(32), "IDXACCFUNC" VARCHAR2(32), "IDXSTATUS" VARCHAR2(11), "OPTFCCPUCT" NUMBER, "OPTFCIOCT" NUMBER, "OPTIXSELVT" NUMBER, "OPTIXCPUCT" NUMBER, "OPTIXIOCT" NUMBER, "OPTPTFSCCT" NUMBER, "IDXPTABSTG" VARCHAR2(1000), "IDXPQUERY" CLOB, CONSTRAINT "DUPLICATE_IDX_NAME" PRIMARY KEY ("IDXOWNER", "IDXNAME") ENABLE, CONSTRAINT "DUPLICATE_IDX_OBJNO" UNIQUE ("IDXOBJ#") USING INDEX ENABLE ) ORGANIZATION INDEX NOCOMPRESS INCLUDING "OPTPTFSCCT" OVERFLOW;, tables=[columns: { IDXOBJ# NUMBER(0) NOT NULL IDXOWNER VARCHAR2(32) NOT NULL IDXNAME VARCHAR2(25) NOT NULL IDXATTRSET VARCHAR2(32) NOT NULL IDXESETTAB VARCHAR2(32) NOT NULL IDXESETCOL VARCHAR2(32) NOT NULL IDXPREDTAB VARCHAR2(32) DEFAULT VALUE NULL IDXACCFUNC VARCHAR2(32) DEFAULT VALUE NULL IDXSTATUS VARCHAR2(11) DEFAULT VALUE NULL OPTFCCPUCT NUMBER(0) DEFAULT VALUE NULL OPTFCIOCT NUMBER(0) DEFAULT VALUE NULL OPTIXSELVT NUMBER(0) DEFAULT VALUE NULL OPTIXCPUCT NUMBER(0) DEFAULT VALUE NULL OPTIXIOCT NUMBER(0) DEFAULT VALUE NULL OPTPTFSCCT NUMBER(0) DEFAULT VALUE NULL IDXPTABSTG VARCHAR2(1000) DEFAULT VALUE NULL IDXPQUERY CLOB(4000) DEFAULT VALUE NULL } primary key: [IDXOWNER, IDXNAME] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014246776] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,776 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#" NUMBER NOT NULL ENABLE, "IDXOWNER" VARCHAR2(32) NOT NULL ENABLE, "IDXNAME" VARCHAR2(25) NOT NULL ENABLE, "IDXATTRSET" VARCHAR2(32) NOT NULL ENABLE, "IDXESETTAB" VARCHAR2(32) NOT NULL ENABLE, "IDXESETCOL" VARCHAR2(32) NOT NULL ENABLE, "IDXPREDTAB" VARCHAR2(32), "IDXACCFUNC" VARCHAR2(32), "IDXSTATUS" VARCHAR2(11), "OPTFCCPUCT" NUMBER, "OPTFCIOCT" NUMBER, "OPTIXSELVT" NUMBER, "OPTIXCPUCT" NUMBER, "OPTIXIOCT" NUMBER, "OPTPTFSCCT" NUMBER, "IDXPTABSTG" VARCHAR2(1000), "IDXPQUERY" CLOB, CONSTRAINT "DUPLICATE_IDX_NAME" PRIMARY KEY ("IDXOWNER", "IDXNAME") ENABLE, CONSTRAINT "DUPLICATE_IDX_OBJNO" UNIQUE ("IDXOBJ#") USING INDEX ENABLE ) ORGANIZATION INDEX NOCOMPRESS INCLUDING "OPTPTFSCCT" OVERFLOW; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,776 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246776, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$IDXSECOBJ\" \n (\t\"IDXOBJ#\" NUMBER NOT NULL ENABLE, \n\t\"IDXOWNER\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXNAME\" VARCHAR2(25) NOT NULL ENABLE, \n\t\"IDXATTRSET\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXESETTAB\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXESETCOL\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXPREDTAB\" VARCHAR2(32), \n\t\"IDXACCFUNC\" VARCHAR2(32), \n\t\"IDXSTATUS\" VARCHAR2(11), \n\t\"OPTFCCPUCT\" NUMBER, \n\t\"OPTFCIOCT\" NUMBER, \n\t\"OPTIXSELVT\" NUMBER, \n\t\"OPTIXCPUCT\" NUMBER, \n\t\"OPTIXIOCT\" NUMBER, \n\t\"OPTPTFSCCT\" NUMBER, \n\t\"IDXPTABSTG\" VARCHAR2(1000), \n\t\"IDXPQUERY\" CLOB, \n\t CONSTRAINT \"DUPLICATE_IDX_NAME\" PRIMARY KEY (\"IDXOWNER\", \"IDXNAME\") ENABLE, \n\t CONSTRAINT \"DUPLICATE_IDX_OBJNO\" UNIQUE (\"IDXOBJ#\")\n USING INDEX ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS INCLUDING \"OPTPTFSCCT\" OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$IDXSECOBJ\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "IDXOWNER", "IDXNAME" ], "columns" : [ { "name" : "IDXOBJ#", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXOWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXNAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 25, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXATTRSET", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXESETTAB", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 5, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXESETCOL", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXPREDTAB", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXACCFUNC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXSTATUS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 11, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTFCCPUCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 10, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTFCIOCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 11, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXSELVT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 12, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXCPUCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 13, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXIOCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 14, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTPTFSCCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 15, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXPTABSTG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 1000, "position" : 16, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXPQUERY", "jdbcType" : 2005, "typeName" : "CLOB", "typeExpression" : "CLOB", "charsetName" : null, "length" : 4000, "position" : 17, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,776 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246776, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$IDXSECOBJ\" \n (\t\"IDXOBJ#\" NUMBER NOT NULL ENABLE, \n\t\"IDXOWNER\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXNAME\" VARCHAR2(25) NOT NULL ENABLE, \n\t\"IDXATTRSET\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXESETTAB\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXESETCOL\" VARCHAR2(32) NOT NULL ENABLE, \n\t\"IDXPREDTAB\" VARCHAR2(32), \n\t\"IDXACCFUNC\" VARCHAR2(32), \n\t\"IDXSTATUS\" VARCHAR2(11), \n\t\"OPTFCCPUCT\" NUMBER, \n\t\"OPTFCIOCT\" NUMBER, \n\t\"OPTIXSELVT\" NUMBER, \n\t\"OPTIXCPUCT\" NUMBER, \n\t\"OPTIXIOCT\" NUMBER, \n\t\"OPTPTFSCCT\" NUMBER, \n\t\"IDXPTABSTG\" VARCHAR2(1000), \n\t\"IDXPQUERY\" CLOB, \n\t CONSTRAINT \"DUPLICATE_IDX_NAME\" PRIMARY KEY (\"IDXOWNER\", \"IDXNAME\") ENABLE, \n\t CONSTRAINT \"DUPLICATE_IDX_OBJNO\" UNIQUE (\"IDXOBJ#\")\n USING INDEX ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS INCLUDING \"OPTPTFSCCT\" OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$IDXSECOBJ\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "IDXOWNER", "IDXNAME" ], "columns" : [ { "name" : "IDXOBJ#", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXOWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXNAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 25, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXATTRSET", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXESETTAB", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 5, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXESETCOL", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "IDXPREDTAB", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXACCFUNC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXSTATUS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 11, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTFCCPUCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 10, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTFCIOCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 11, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXSELVT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 12, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXCPUCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 13, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIXIOCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 14, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTPTFSCCT", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 15, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXPTABSTG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 1000, "position" : 16, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IDXPQUERY", "jdbcType" : 2005, "typeName" : "CLOB", "typeExpression" : "CLOB", "charsetName" : null, "length" : 4000, "position" : 17, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,776 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:06,777 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,777 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,777 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,777 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,777 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5516) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=8183]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,777 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=8183, buffer=java.nio.HeapByteBuffer[pos=0 lim=8183 cap=8183]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,777 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,777 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5516): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11035, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,778 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5516 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,778 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11035. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:06,778 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,778 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11035 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,778 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246715,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$IDXSECOBJ,scn=5535835,ssn=0},ts_ms=1666014246776,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#" NUMBER NOT NULL ENABLE, "IDXOWNER" VARCHAR2(32) NOT NULL ENABLE, "IDXNAME" VARCHAR2(25) NOT NULL ENABLE, "IDXATTRSET" VARCHAR2(32) NOT NULL ENABLE, "IDXESETTAB" VARCHAR2(32) NOT NULL ENABLE, "IDXESETCOL" VARCHAR2(32) NOT NULL ENABLE, "IDXPREDTAB" VARCHAR2(32), "IDXACCFUNC" VARCHAR2(32), "IDXSTATUS" VARCHAR2(11), "OPTFCCPUCT" NUMBER, "OPTFCIOCT" NUMBER, "OPTIXSELVT" NUMBER, "OPTIXCPUCT" NUMBER, "OPTIXIOCT" NUMBER, "OPTPTFSCCT" NUMBER, "IDXPTABSTG" VARCHAR2(1000), "IDXPQUERY" CLOB, CONSTRAINT "DUPLICATE_IDX_NAME" PRIMARY KEY ("IDXOWNER", "IDXNAME") ENABLE, CONSTRAINT "DUPLICATE_IDX_OBJNO" UNIQUE ("IDXOBJ#") USING INDEX ENABLE ) ORGANIZATION INDEX NOCOMPRESS INCLUDING "OPTPTFSCCT" OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$IDXSECOBJ",table=Struct{primaryKeyColumnNames=[IDXOWNER, IDXNAME],columns=[Struct{name=IDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXATTRSET,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETCOL,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXPREDTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXACCFUNC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXSTATUS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=11,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXSELVT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTPTFSCCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPTABSTG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1000,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPQUERY,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=17,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:06,778 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$JAVAMSG [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:06,778 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$JAVAMSG' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,818 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,819 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,820 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,820 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$JAVAMSG','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,832 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,833 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE GLOBAL TEMPORARY TABLE "EXFSYS"."EXF$JAVAMSG" ("CODE" VARCHAR2(15), "MESSAGE" VARCHAR2(500) ) ON COMMIT PRESERVE ROWS ;, tables=[columns: { CODE VARCHAR2(15) DEFAULT VALUE NULL MESSAGE VARCHAR2(500) DEFAULT VALUE NULL } primary key: [] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014246833] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,833 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE GLOBAL TEMPORARY TABLE "EXFSYS"."EXF$JAVAMSG" ("CODE" VARCHAR2(15), "MESSAGE" VARCHAR2(500) ) ON COMMIT PRESERVE ROWS ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,833 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246833, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE GLOBAL TEMPORARY TABLE \"EXFSYS\".\"EXF$JAVAMSG\" \n (\t\"CODE\" VARCHAR2(15), \n\t\"MESSAGE\" VARCHAR2(500)\n ) ON COMMIT PRESERVE ROWS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$JAVAMSG\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "CODE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 15, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "MESSAGE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,833 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246833, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE GLOBAL TEMPORARY TABLE \"EXFSYS\".\"EXF$JAVAMSG\" \n (\t\"CODE\" VARCHAR2(15), \n\t\"MESSAGE\" VARCHAR2(500)\n ) ON COMMIT PRESERVE ROWS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$JAVAMSG\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "CODE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 15, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "MESSAGE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,833 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:06,833 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,833 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,833 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,833 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,833 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5517) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=1541]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,833 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=1541, buffer=java.nio.HeapByteBuffer[pos=0 lim=1541 cap=1541]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,833 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,834 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5517): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11036, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,834 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5517 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,834 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11036. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:06,834 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,834 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11036 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,834 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246778,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$JAVAMSG,scn=5535835,ssn=0},ts_ms=1666014246833,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE GLOBAL TEMPORARY TABLE "EXFSYS"."EXF$JAVAMSG" ("CODE" VARCHAR2(15), "MESSAGE" VARCHAR2(500) ) ON COMMIT PRESERVE ROWS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$JAVAMSG",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=MESSAGE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=2,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:06,834 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$PARAMETER [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:06,834 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$PARAMETER' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=946): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=1518104247, responses=[]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 1518104247 with response=(), implied=(dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-7 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-13 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-11 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-1 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-5 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-3 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-23 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-17 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-15 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-21 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-19 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-10 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-8 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-14 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-12 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-2 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,846 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-0 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-6 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-4 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-24 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-18 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-16 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-22 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-20 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-9 at position FetchPosition{offset=12, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Built incremental fetch (sessionId=1518104247, epoch=940) for node 1. Added (), altered (), removed (), replaced () out of (dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(), toForget=(), toReplace=(), implied=(dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Polling for fetches with timeout 2147432944 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:06,847 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=947) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=1518104247, sessionEpoch=940, topics=[], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-7 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-13 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-11 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-1 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-5 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-3 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-23 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-17 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-15 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-21 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-19 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-10 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-8 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-14 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-12 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-2 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-6 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-4 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-24 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-18 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-16 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-22 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-20 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-9 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:06,847 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Polling for fetches with timeout 2147432943 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:06,851 TRACE || [Heartbeat groupID=dbz] Sending heartbeat request with 3000ms remaining on timer [org.apache.kafka.clients.consumer.internals.Heartbeat] 2022-10-17 13:44:06,851 DEBUG || [Worker clientId=connect-1, groupId=dbz] Sending Heartbeat request with generation 5 and member id connect-1-74f44e69-79bb-4a51-9f4c-bd42f4a3419f to coordinator kafka:29092 (id: 2147483646 rack: null) [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2022-10-17 13:44:06,852 DEBUG || [Worker clientId=connect-1, groupId=dbz] Sending HEARTBEAT request with header RequestHeader(apiKey=HEARTBEAT, apiVersion=4, clientId=connect-1, correlationId=165) and timeout 40000 to node 2147483646: HeartbeatRequestData(groupId='dbz', generationId=5, memberId='connect-1-74f44e69-79bb-4a51-9f4c-bd42f4a3419f', groupInstanceId=null) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,852 DEBUG || [Worker clientId=connect-1, groupId=dbz] Received HEARTBEAT response from node 2147483646 for request with header RequestHeader(apiKey=HEARTBEAT, apiVersion=4, clientId=connect-1, correlationId=165): HeartbeatResponseData(throttleTimeMs=0, errorCode=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,853 DEBUG || [Worker clientId=connect-1, groupId=dbz] Received successful Heartbeat response [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] 2022-10-17 13:44:06,885 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,886 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,886 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,887 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$PARAMETER','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,898 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,899 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$PARAMETER" ("NUM" NUMBER, "NAME" VARCHAR2(64), "VALTYPE" NUMBER, "VALUE" VARCHAR2(512), CONSTRAINT "DUP_PARAMETER" PRIMARY KEY ("NUM") USING INDEX ENABLE ) ;, tables=[columns: { NUM NUMBER(0) NOT NULL NAME VARCHAR2(64) DEFAULT VALUE NULL VALTYPE NUMBER(0) DEFAULT VALUE NULL VALUE VARCHAR2(512) DEFAULT VALUE NULL } primary key: [NUM] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014246899] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,899 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$PARAMETER" ("NUM" NUMBER, "NAME" VARCHAR2(64), "VALTYPE" NUMBER, "VALUE" VARCHAR2(512), CONSTRAINT "DUP_PARAMETER" PRIMARY KEY ("NUM") USING INDEX ENABLE ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,899 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246899, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PARAMETER\" \n (\t\"NUM\" NUMBER, \n\t\"NAME\" VARCHAR2(64), \n\t\"VALTYPE\" NUMBER, \n\t\"VALUE\" VARCHAR2(512), \n\t CONSTRAINT \"DUP_PARAMETER\" PRIMARY KEY (\"NUM\")\n USING INDEX ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PARAMETER\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "NUM" ], "columns" : [ { "name" : "NUM", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 64, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "VALTYPE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "VALUE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 512, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,899 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246899, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PARAMETER\" \n (\t\"NUM\" NUMBER, \n\t\"NAME\" VARCHAR2(64), \n\t\"VALTYPE\" NUMBER, \n\t\"VALUE\" VARCHAR2(512), \n\t CONSTRAINT \"DUP_PARAMETER\" PRIMARY KEY (\"NUM\")\n USING INDEX ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PARAMETER\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "NUM" ], "columns" : [ { "name" : "NUM", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 64, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "VALTYPE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "VALUE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 512, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,899 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:06,899 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,899 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,899 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,899 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,900 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5518) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=2398]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,900 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=2398, buffer=java.nio.HeapByteBuffer[pos=0 lim=2398 cap=2398]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,900 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,900 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5518): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11037, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,901 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5518 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,901 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11037. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:06,901 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,901 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11037 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,901 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246834,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PARAMETER,scn=5535835,ssn=0},ts_ms=1666014246899,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PARAMETER" ("NUM" NUMBER, "NAME" VARCHAR2(64), "VALTYPE" NUMBER, "VALUE" VARCHAR2(512), CONSTRAINT "DUP_PARAMETER" PRIMARY KEY ("NUM") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PARAMETER",table=Struct{primaryKeyColumnNames=[NUM],columns=[Struct{name=NUM,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=64,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=VALTYPE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=VALUE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=512,position=4,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:06,901 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$PLAN_TABLE [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:06,901 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$PLAN_TABLE' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,941 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,942 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,942 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,943 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$PLAN_TABLE','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,961 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:06,962 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$PLAN_TABLE" ("STATEMENT_ID" VARCHAR2(30), "PLAN_ID" NUMBER, "TIMESTAMP" DATE, "REMARKS" VARCHAR2(4000), "OPERATION" VARCHAR2(30), "OPTIONS" VARCHAR2(255), "OBJECT_NODE" VARCHAR2(128), "OBJECT_OWNER" VARCHAR2(30), "OBJECT_NAME" VARCHAR2(30), "OBJECT_ALIAS" VARCHAR2(65), "OBJECT_INSTANCE" NUMBER(*,0), "OBJECT_TYPE" VARCHAR2(30), "OPTIMIZER" VARCHAR2(255), "SEARCH_COLUMNS" NUMBER, "ID" NUMBER(*,0), "PARENT_ID" NUMBER(*,0), "DEPTH" NUMBER(*,0), "POSITION" NUMBER(*,0), "COST" NUMBER(*,0), "CARDINALITY" NUMBER(*,0), "BYTES" NUMBER(*,0), "OTHER_TAG" VARCHAR2(255), "PARTITION_START" VARCHAR2(255), "PARTITION_STOP" VARCHAR2(255), "PARTITION_ID" NUMBER(*,0), "OTHER" LONG, "DISTRIBUTION" VARCHAR2(30), "CPU_COST" NUMBER(*,0), "IO_COST" NUMBER(*,0), "TEMP_SPACE" NUMBER(*,0), "ACCESS_PREDICATES" VARCHAR2(4000), "FILTER_PREDICATES" VARCHAR2(4000), "PROJECTION" VARCHAR2(4000), "TIME" NUMBER(*,0), "QBLOCK_NAME" VARCHAR2(30), "OTHER_XML" CLOB, CONSTRAINT "PLAN_STMT_ID" PRIMARY KEY ("STATEMENT_ID", "ID") USING INDEX ENABLE ) ;, tables=[columns: { STATEMENT_ID VARCHAR2(30) NOT NULL PLAN_ID NUMBER(0) DEFAULT VALUE NULL TIMESTAMP DATE DEFAULT VALUE NULL REMARKS VARCHAR2(4000) DEFAULT VALUE NULL OPERATION VARCHAR2(30) DEFAULT VALUE NULL OPTIONS VARCHAR2(255) DEFAULT VALUE NULL OBJECT_NODE VARCHAR2(128) DEFAULT VALUE NULL OBJECT_OWNER VARCHAR2(30) DEFAULT VALUE NULL OBJECT_NAME VARCHAR2(30) DEFAULT VALUE NULL OBJECT_ALIAS VARCHAR2(65) DEFAULT VALUE NULL OBJECT_INSTANCE NUMBER(38, 0) DEFAULT VALUE NULL OBJECT_TYPE VARCHAR2(30) DEFAULT VALUE NULL OPTIMIZER VARCHAR2(255) DEFAULT VALUE NULL SEARCH_COLUMNS NUMBER(0) DEFAULT VALUE NULL ID NUMBER(38, 0) NOT NULL PARENT_ID NUMBER(38, 0) DEFAULT VALUE NULL DEPTH NUMBER(38, 0) DEFAULT VALUE NULL POSITION NUMBER(38, 0) DEFAULT VALUE NULL COST NUMBER(38, 0) DEFAULT VALUE NULL CARDINALITY NUMBER(38, 0) DEFAULT VALUE NULL BYTES NUMBER(38, 0) DEFAULT VALUE NULL OTHER_TAG VARCHAR2(255) DEFAULT VALUE NULL PARTITION_START VARCHAR2(255) DEFAULT VALUE NULL PARTITION_STOP VARCHAR2(255) DEFAULT VALUE NULL PARTITION_ID NUMBER(38, 0) DEFAULT VALUE NULL OTHER LONG(0) DEFAULT VALUE NULL DISTRIBUTION VARCHAR2(30) DEFAULT VALUE NULL CPU_COST NUMBER(38, 0) DEFAULT VALUE NULL IO_COST NUMBER(38, 0) DEFAULT VALUE NULL TEMP_SPACE NUMBER(38, 0) DEFAULT VALUE NULL ACCESS_PREDICATES VARCHAR2(4000) DEFAULT VALUE NULL FILTER_PREDICATES VARCHAR2(4000) DEFAULT VALUE NULL PROJECTION VARCHAR2(4000) DEFAULT VALUE NULL TIME NUMBER(38, 0) DEFAULT VALUE NULL QBLOCK_NAME VARCHAR2(30) DEFAULT VALUE NULL OTHER_XML CLOB(4000) DEFAULT VALUE NULL } primary key: [STATEMENT_ID, ID] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014246962] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,962 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$PLAN_TABLE" ("STATEMENT_ID" VARCHAR2(30), "PLAN_ID" NUMBER, "TIMESTAMP" DATE, "REMARKS" VARCHAR2(4000), "OPERATION" VARCHAR2(30), "OPTIONS" VARCHAR2(255), "OBJECT_NODE" VARCHAR2(128), "OBJECT_OWNER" VARCHAR2(30), "OBJECT_NAME" VARCHAR2(30), "OBJECT_ALIAS" VARCHAR2(65), "OBJECT_INSTANCE" NUMBER(*,0), "OBJECT_TYPE" VARCHAR2(30), "OPTIMIZER" VARCHAR2(255), "SEARCH_COLUMNS" NUMBER, "ID" NUMBER(*,0), "PARENT_ID" NUMBER(*,0), "DEPTH" NUMBER(*,0), "POSITION" NUMBER(*,0), "COST" NUMBER(*,0), "CARDINALITY" NUMBER(*,0), "BYTES" NUMBER(*,0), "OTHER_TAG" VARCHAR2(255), "PARTITION_START" VARCHAR2(255), "PARTITION_STOP" VARCHAR2(255), "PARTITION_ID" NUMBER(*,0), "OTHER" LONG, "DISTRIBUTION" VARCHAR2(30), "CPU_COST" NUMBER(*,0), "IO_COST" NUMBER(*,0), "TEMP_SPACE" NUMBER(*,0), "ACCESS_PREDICATES" VARCHAR2(4000), "FILTER_PREDICATES" VARCHAR2(4000), "PROJECTION" VARCHAR2(4000), "TIME" NUMBER(*,0), "QBLOCK_NAME" VARCHAR2(30), "OTHER_XML" CLOB, CONSTRAINT "PLAN_STMT_ID" PRIMARY KEY ("STATEMENT_ID", "ID") USING INDEX ENABLE ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:06,962 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246962, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PLAN_TABLE\" \n (\t\"STATEMENT_ID\" VARCHAR2(30), \n\t\"PLAN_ID\" NUMBER, \n\t\"TIMESTAMP\" DATE, \n\t\"REMARKS\" VARCHAR2(4000), \n\t\"OPERATION\" VARCHAR2(30), \n\t\"OPTIONS\" VARCHAR2(255), \n\t\"OBJECT_NODE\" VARCHAR2(128), \n\t\"OBJECT_OWNER\" VARCHAR2(30), \n\t\"OBJECT_NAME\" VARCHAR2(30), \n\t\"OBJECT_ALIAS\" VARCHAR2(65), \n\t\"OBJECT_INSTANCE\" NUMBER(*,0), \n\t\"OBJECT_TYPE\" VARCHAR2(30), \n\t\"OPTIMIZER\" VARCHAR2(255), \n\t\"SEARCH_COLUMNS\" NUMBER, \n\t\"ID\" NUMBER(*,0), \n\t\"PARENT_ID\" NUMBER(*,0), \n\t\"DEPTH\" NUMBER(*,0), \n\t\"POSITION\" NUMBER(*,0), \n\t\"COST\" NUMBER(*,0), \n\t\"CARDINALITY\" NUMBER(*,0), \n\t\"BYTES\" NUMBER(*,0), \n\t\"OTHER_TAG\" VARCHAR2(255), \n\t\"PARTITION_START\" VARCHAR2(255), \n\t\"PARTITION_STOP\" VARCHAR2(255), \n\t\"PARTITION_ID\" NUMBER(*,0), \n\t\"OTHER\" LONG, \n\t\"DISTRIBUTION\" VARCHAR2(30), \n\t\"CPU_COST\" NUMBER(*,0), \n\t\"IO_COST\" NUMBER(*,0), \n\t\"TEMP_SPACE\" NUMBER(*,0), \n\t\"ACCESS_PREDICATES\" VARCHAR2(4000), \n\t\"FILTER_PREDICATES\" VARCHAR2(4000), \n\t\"PROJECTION\" VARCHAR2(4000), \n\t\"TIME\" NUMBER(*,0), \n\t\"QBLOCK_NAME\" VARCHAR2(30), \n\t\"OTHER_XML\" CLOB, \n\t CONSTRAINT \"PLAN_STMT_ID\" PRIMARY KEY (\"STATEMENT_ID\", \"ID\")\n USING INDEX ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PLAN_TABLE\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "STATEMENT_ID", "ID" ], "columns" : [ { "name" : "STATEMENT_ID", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PLAN_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TIMESTAMP", "jdbcType" : 93, "typeName" : "DATE", "typeExpression" : "DATE", "charsetName" : null, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "REMARKS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPERATION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIONS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_NODE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 128, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_ALIAS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 65, "position" : 10, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_INSTANCE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 11, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_TYPE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 12, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIMIZER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 13, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "SEARCH_COLUMNS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 14, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 15, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PARENT_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 16, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DEPTH", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 17, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "POSITION", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 18, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 19, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "CARDINALITY", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 20, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "BYTES", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 21, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER_TAG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 22, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_START", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 23, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_STOP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 24, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 25, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER", "jdbcType" : -1, "typeName" : "LONG", "typeExpression" : "LONG", "charsetName" : null, "length" : 0, "position" : 26, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DISTRIBUTION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 27, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "CPU_COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 28, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IO_COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 29, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TEMP_SPACE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 30, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ACCESS_PREDICATES", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 31, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "FILTER_PREDICATES", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 32, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PROJECTION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 33, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TIME", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 34, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "QBLOCK_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 35, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER_XML", "jdbcType" : 2005, "typeName" : "CLOB", "typeExpression" : "CLOB", "charsetName" : null, "length" : 4000, "position" : 36, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,963 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014246962, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PLAN_TABLE\" \n (\t\"STATEMENT_ID\" VARCHAR2(30), \n\t\"PLAN_ID\" NUMBER, \n\t\"TIMESTAMP\" DATE, \n\t\"REMARKS\" VARCHAR2(4000), \n\t\"OPERATION\" VARCHAR2(30), \n\t\"OPTIONS\" VARCHAR2(255), \n\t\"OBJECT_NODE\" VARCHAR2(128), \n\t\"OBJECT_OWNER\" VARCHAR2(30), \n\t\"OBJECT_NAME\" VARCHAR2(30), \n\t\"OBJECT_ALIAS\" VARCHAR2(65), \n\t\"OBJECT_INSTANCE\" NUMBER(*,0), \n\t\"OBJECT_TYPE\" VARCHAR2(30), \n\t\"OPTIMIZER\" VARCHAR2(255), \n\t\"SEARCH_COLUMNS\" NUMBER, \n\t\"ID\" NUMBER(*,0), \n\t\"PARENT_ID\" NUMBER(*,0), \n\t\"DEPTH\" NUMBER(*,0), \n\t\"POSITION\" NUMBER(*,0), \n\t\"COST\" NUMBER(*,0), \n\t\"CARDINALITY\" NUMBER(*,0), \n\t\"BYTES\" NUMBER(*,0), \n\t\"OTHER_TAG\" VARCHAR2(255), \n\t\"PARTITION_START\" VARCHAR2(255), \n\t\"PARTITION_STOP\" VARCHAR2(255), \n\t\"PARTITION_ID\" NUMBER(*,0), \n\t\"OTHER\" LONG, \n\t\"DISTRIBUTION\" VARCHAR2(30), \n\t\"CPU_COST\" NUMBER(*,0), \n\t\"IO_COST\" NUMBER(*,0), \n\t\"TEMP_SPACE\" NUMBER(*,0), \n\t\"ACCESS_PREDICATES\" VARCHAR2(4000), \n\t\"FILTER_PREDICATES\" VARCHAR2(4000), \n\t\"PROJECTION\" VARCHAR2(4000), \n\t\"TIME\" NUMBER(*,0), \n\t\"QBLOCK_NAME\" VARCHAR2(30), \n\t\"OTHER_XML\" CLOB, \n\t CONSTRAINT \"PLAN_STMT_ID\" PRIMARY KEY (\"STATEMENT_ID\", \"ID\")\n USING INDEX ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PLAN_TABLE\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "STATEMENT_ID", "ID" ], "columns" : [ { "name" : "STATEMENT_ID", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PLAN_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TIMESTAMP", "jdbcType" : 93, "typeName" : "DATE", "typeExpression" : "DATE", "charsetName" : null, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "REMARKS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPERATION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIONS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_NODE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 128, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_ALIAS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 65, "position" : 10, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_INSTANCE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 11, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OBJECT_TYPE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 12, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OPTIMIZER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 13, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "SEARCH_COLUMNS", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 14, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 15, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PARENT_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 16, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DEPTH", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 17, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "POSITION", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 18, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 19, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "CARDINALITY", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 20, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "BYTES", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 21, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER_TAG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 22, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_START", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 23, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_STOP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 255, "position" : 24, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PARTITION_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 25, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER", "jdbcType" : -1, "typeName" : "LONG", "typeExpression" : "LONG", "charsetName" : null, "length" : 0, "position" : 26, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DISTRIBUTION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 27, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "CPU_COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 28, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "IO_COST", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 29, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TEMP_SPACE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 30, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ACCESS_PREDICATES", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 31, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "FILTER_PREDICATES", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 32, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PROJECTION", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 4000, "position" : 33, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "TIME", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 38, "scale" : 0, "position" : 34, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "QBLOCK_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 30, "position" : 35, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "OTHER_XML", "jdbcType" : 2005, "typeName" : "CLOB", "typeExpression" : "CLOB", "charsetName" : null, "length" : 4000, "position" : 36, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,963 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:06,963 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,963 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:06,964 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,964 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,964 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5519) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=16301]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,964 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=16301, buffer=java.nio.HeapByteBuffer[pos=0 lim=16301 cap=16301]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,964 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,964 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5519): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11038, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:06,964 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5519 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:06,964 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11038. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:06,965 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:06,965 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11038 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:06,965 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246901,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PLAN_TABLE,scn=5535835,ssn=0},ts_ms=1666014246962,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PLAN_TABLE" ("STATEMENT_ID" VARCHAR2(30), "PLAN_ID" NUMBER, "TIMESTAMP" DATE, "REMARKS" VARCHAR2(4000), "OPERATION" VARCHAR2(30), "OPTIONS" VARCHAR2(255), "OBJECT_NODE" VARCHAR2(128), "OBJECT_OWNER" VARCHAR2(30), "OBJECT_NAME" VARCHAR2(30), "OBJECT_ALIAS" VARCHAR2(65), "OBJECT_INSTANCE" NUMBER(*,0), "OBJECT_TYPE" VARCHAR2(30), "OPTIMIZER" VARCHAR2(255), "SEARCH_COLUMNS" NUMBER, "ID" NUMBER(*,0), "PARENT_ID" NUMBER(*,0), "DEPTH" NUMBER(*,0), "POSITION" NUMBER(*,0), "COST" NUMBER(*,0), "CARDINALITY" NUMBER(*,0), "BYTES" NUMBER(*,0), "OTHER_TAG" VARCHAR2(255), "PARTITION_START" VARCHAR2(255), "PARTITION_STOP" VARCHAR2(255), "PARTITION_ID" NUMBER(*,0), "OTHER" LONG, "DISTRIBUTION" VARCHAR2(30), "CPU_COST" NUMBER(*,0), "IO_COST" NUMBER(*,0), "TEMP_SPACE" NUMBER(*,0), "ACCESS_PREDICATES" VARCHAR2(4000), "FILTER_PREDICATES" VARCHAR2(4000), "PROJECTION" VARCHAR2(4000), "TIME" NUMBER(*,0), "QBLOCK_NAME" VARCHAR2(30), "OTHER_XML" CLOB, CONSTRAINT "PLAN_STMT_ID" PRIMARY KEY ("STATEMENT_ID", "ID") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PLAN_TABLE",table=Struct{primaryKeyColumnNames=[STATEMENT_ID, ID],columns=[Struct{name=STATEMENT_ID,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PLAN_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=TIMESTAMP,jdbcType=93,typeName=DATE,typeExpression=DATE,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=REMARKS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=OPERATION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIONS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=128,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_ALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_INSTANCE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_TYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIMIZER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=SEARCH_COLUMNS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=15,optional=false,autoIncremented=false,generated=false}, Struct{name=PARENT_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=DEPTH,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=17,optional=true,autoIncremented=false,generated=false}, Struct{name=POSITION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=18,optional=true,autoIncremented=false,generated=false}, Struct{name=COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=19,optional=true,autoIncremented=false,generated=false}, Struct{name=CARDINALITY,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=20,optional=true,autoIncremented=false,generated=false}, Struct{name=BYTES,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=21,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_TAG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=22,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_START,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=23,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_STOP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=24,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=25,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER,jdbcType=-1,typeName=LONG,typeExpression=LONG,length=0,position=26,optional=true,autoIncremented=false,generated=false}, Struct{name=DISTRIBUTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=27,optional=true,autoIncremented=false,generated=false}, Struct{name=CPU_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=28,optional=true,autoIncremented=false,generated=false}, Struct{name=IO_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=29,optional=true,autoIncremented=false,generated=false}, Struct{name=TEMP_SPACE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=30,optional=true,autoIncremented=false,generated=false}, Struct{name=ACCESS_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=31,optional=true,autoIncremented=false,generated=false}, Struct{name=FILTER_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=32,optional=true,autoIncremented=false,generated=false}, Struct{name=PROJECTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=33,optional=true,autoIncremented=false,generated=false}, Struct{name=TIME,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=34,optional=true,autoIncremented=false,generated=false}, Struct{name=QBLOCK_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=35,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_XML,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=36,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:06,965 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$PREDATTRMAP [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:06,965 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$PREDATTRMAP' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,008 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,009 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,009 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,010 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$PREDATTRMAP','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,026 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,027 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$PREDATTRMAP" ("PTIDXOBJ#" NUMBER NOT NULL ENABLE, "PTATTRSEXP" VARCHAR2(300) NOT NULL ENABLE, "PTATTRID" NUMBER, "PTATTRALIAS" VARCHAR2(25), "PTATTROPER" "EXFSYS"."EXF$INDEXOPER" , "PTATTRTYPE" VARCHAR2(65), "PTATTRPROP" NUMBER, "XMLTATTR" VARCHAR2(32) DEFAULT null, "XMLNSELP" NUMBER DEFAULT null, CONSTRAINT "DUP_ATTR_EXP" UNIQUE ("PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR") USING INDEX ENABLE, CONSTRAINT "REF_IDX_ATTRS" FOREIGN KEY ("PTIDXOBJ#") REFERENCES "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#") ON DELETE CASCADE ENABLE ) ;, tables=[columns: { PTIDXOBJ# NUMBER(0) NOT NULL PTATTRSEXP VARCHAR2(300) NOT NULL PTATTRID NUMBER(0) DEFAULT VALUE NULL PTATTRALIAS VARCHAR2(25) DEFAULT VALUE NULL PTATTROPER EXF$INDEXOPER(337) DEFAULT VALUE NULL PTATTRTYPE VARCHAR2(65) DEFAULT VALUE NULL PTATTRPROP NUMBER(0) DEFAULT VALUE NULL XMLTATTR VARCHAR2(32) DEFAULT VALUE null XMLNSELP NUMBER(0) DEFAULT VALUE null } primary key: [PTIDXOBJ#, PTATTRSEXP, PTATTRALIAS, XMLTATTR] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247027] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,027 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$PREDATTRMAP" ("PTIDXOBJ#" NUMBER NOT NULL ENABLE, "PTATTRSEXP" VARCHAR2(300) NOT NULL ENABLE, "PTATTRID" NUMBER, "PTATTRALIAS" VARCHAR2(25), "PTATTROPER" "EXFSYS"."EXF$INDEXOPER" , "PTATTRTYPE" VARCHAR2(65), "PTATTRPROP" NUMBER, "XMLTATTR" VARCHAR2(32) DEFAULT null, "XMLNSELP" NUMBER DEFAULT null, CONSTRAINT "DUP_ATTR_EXP" UNIQUE ("PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR") USING INDEX ENABLE, CONSTRAINT "REF_IDX_ATTRS" FOREIGN KEY ("PTIDXOBJ#") REFERENCES "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#") ON DELETE CASCADE ENABLE ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,027 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247027, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PREDATTRMAP\" \n (\t\"PTIDXOBJ#\" NUMBER NOT NULL ENABLE, \n\t\"PTATTRSEXP\" VARCHAR2(300) NOT NULL ENABLE, \n\t\"PTATTRID\" NUMBER, \n\t\"PTATTRALIAS\" VARCHAR2(25), \n\t\"PTATTROPER\" \"EXFSYS\".\"EXF$INDEXOPER\" , \n\t\"PTATTRTYPE\" VARCHAR2(65), \n\t\"PTATTRPROP\" NUMBER, \n\t\"XMLTATTR\" VARCHAR2(32) DEFAULT null, \n\t\"XMLNSELP\" NUMBER DEFAULT null, \n\t CONSTRAINT \"DUP_ATTR_EXP\" UNIQUE (\"PTIDXOBJ#\", \"PTATTRSEXP\", \"PTATTRALIAS\", \"XMLTATTR\")\n USING INDEX ENABLE, \n\t CONSTRAINT \"REF_IDX_ATTRS\" FOREIGN KEY (\"PTIDXOBJ#\")\n\t REFERENCES \"EXFSYS\".\"EXF$IDXSECOBJ\" (\"IDXOBJ#\") ON DELETE CASCADE ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PREDATTRMAP\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR" ], "columns" : [ { "name" : "PTIDXOBJ#", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PTATTRSEXP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 300, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PTATTRID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRALIAS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 25, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTROPER", "jdbcType" : 2003, "typeName" : "EXF$INDEXOPER", "typeExpression" : "EXF$INDEXOPER", "charsetName" : null, "length" : 337, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRTYPE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 65, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRPROP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "XMLTATTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "defaultValueExpression" : "null", "enumValues" : [ ] }, { "name" : "XMLNSELP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "defaultValueExpression" : "null", "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,027 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247027, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$PREDATTRMAP\" \n (\t\"PTIDXOBJ#\" NUMBER NOT NULL ENABLE, \n\t\"PTATTRSEXP\" VARCHAR2(300) NOT NULL ENABLE, \n\t\"PTATTRID\" NUMBER, \n\t\"PTATTRALIAS\" VARCHAR2(25), \n\t\"PTATTROPER\" \"EXFSYS\".\"EXF$INDEXOPER\" , \n\t\"PTATTRTYPE\" VARCHAR2(65), \n\t\"PTATTRPROP\" NUMBER, \n\t\"XMLTATTR\" VARCHAR2(32) DEFAULT null, \n\t\"XMLNSELP\" NUMBER DEFAULT null, \n\t CONSTRAINT \"DUP_ATTR_EXP\" UNIQUE (\"PTIDXOBJ#\", \"PTATTRSEXP\", \"PTATTRALIAS\", \"XMLTATTR\")\n USING INDEX ENABLE, \n\t CONSTRAINT \"REF_IDX_ATTRS\" FOREIGN KEY (\"PTIDXOBJ#\")\n\t REFERENCES \"EXFSYS\".\"EXF$IDXSECOBJ\" (\"IDXOBJ#\") ON DELETE CASCADE ENABLE\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$PREDATTRMAP\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR" ], "columns" : [ { "name" : "PTIDXOBJ#", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PTATTRSEXP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 300, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PTATTRID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRALIAS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 25, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTROPER", "jdbcType" : 2003, "typeName" : "EXF$INDEXOPER", "typeExpression" : "EXF$INDEXOPER", "charsetName" : null, "length" : 337, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRTYPE", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 65, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PTATTRPROP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "XMLTATTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "defaultValueExpression" : "null", "enumValues" : [ ] }, { "name" : "XMLNSELP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 9, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "defaultValueExpression" : "null", "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,027 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,027 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,028 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,028 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,028 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,028 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5520) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=4960]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,028 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=4960, buffer=java.nio.HeapByteBuffer[pos=0 lim=4960 cap=4960]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,028 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,029 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5520): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11039, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,029 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5520 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,029 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11039. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,029 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,029 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11039 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,030 INFO Oracle|oracle|snapshot Already applied 11040 database changes [io.debezium.relational.history.SchemaHistoryMetrics] 2022-10-17 13:44:07,030 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246965,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PREDATTRMAP,scn=5535835,ssn=0},ts_ms=1666014247027,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PREDATTRMAP" ("PTIDXOBJ#" NUMBER NOT NULL ENABLE, "PTATTRSEXP" VARCHAR2(300) NOT NULL ENABLE, "PTATTRID" NUMBER, "PTATTRALIAS" VARCHAR2(25), "PTATTROPER" "EXFSYS"."EXF$INDEXOPER" , "PTATTRTYPE" VARCHAR2(65), "PTATTRPROP" NUMBER, "XMLTATTR" VARCHAR2(32) DEFAULT null, "XMLNSELP" NUMBER DEFAULT null, CONSTRAINT "DUP_ATTR_EXP" UNIQUE ("PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR") USING INDEX ENABLE, CONSTRAINT "REF_IDX_ATTRS" FOREIGN KEY ("PTIDXOBJ#") REFERENCES "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#") ON DELETE CASCADE ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PREDATTRMAP",table=Struct{primaryKeyColumnNames=[PTIDXOBJ#, PTATTRSEXP, PTATTRALIAS, XMLTATTR],columns=[Struct{name=PTIDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRSEXP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTROPER,jdbcType=2003,typeName=EXF$INDEXOPER,typeExpression=EXF$INDEXOPER,length=337,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRTYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRPROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=XMLTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}, Struct{name=XMLNSELP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,030 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$VALIDIOPER [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,030 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$VALIDIOPER' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,064 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,065 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,065 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} About to send 8 records to Kafka [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,065 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246503,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPRSET,scn=5535835,ssn=0},ts_ms=1666014246574,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPRSET" ("EXSOWNER" VARCHAR2(32) NOT NULL ENABLE, "EXSTABNM" VARCHAR2(32) NOT NULL ENABLE, "EXSCOLNM" VARCHAR2(32) NOT NULL ENABLE, "EXSATSNM" VARCHAR2(32) NOT NULL ENABLE, "EXSTABOBJ" NUMBER, "EXSPRVTRIG" VARCHAR2(32), "EXSETLANL" DATE, "EXSETNEXP" NUMBER, "EXSETSPRP" NUMBER, "AVGPRPEXP" NUMBER, CONSTRAINT "DUPL_EXPRSET" PRIMARY KEY ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ENABLE, CONSTRAINT "REF_EXPRSET_ATTRSET" FOREIGN KEY ("EXSOWNER", "EXSATSNM") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPRSET",table=Struct{primaryKeyColumnNames=[EXSOWNER, EXSTABNM, EXSCOLNM],columns=[Struct{name=EXSOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSTABNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSCOLNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSATSNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSTABOBJ,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSPRVTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETLANL,jdbcType=93,typeName=DATE,typeExpression=DATE,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETNEXP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETSPRP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=AVGPRPEXP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,065 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246503,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPRSET,scn=5535835,ssn=0},ts_ms=1666014246574,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPRSET" ("EXSOWNER" VARCHAR2(32) NOT NULL ENABLE, "EXSTABNM" VARCHAR2(32) NOT NULL ENABLE, "EXSCOLNM" VARCHAR2(32) NOT NULL ENABLE, "EXSATSNM" VARCHAR2(32) NOT NULL ENABLE, "EXSTABOBJ" NUMBER, "EXSPRVTRIG" VARCHAR2(32), "EXSETLANL" DATE, "EXSETNEXP" NUMBER, "EXSETSPRP" NUMBER, "AVGPRPEXP" NUMBER, CONSTRAINT "DUPL_EXPRSET" PRIMARY KEY ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ENABLE, CONSTRAINT "REF_EXPRSET_ATTRSET" FOREIGN KEY ("EXSOWNER", "EXSATSNM") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPRSET",table=Struct{primaryKeyColumnNames=[EXSOWNER, EXSTABNM, EXSCOLNM],columns=[Struct{name=EXSOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSTABNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSCOLNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSATSNM,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=EXSTABOBJ,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSPRVTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETLANL,jdbcType=93,typeName=DATE,typeExpression=DATE,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETNEXP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=EXSETSPRP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=AVGPRPEXP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,065 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@1c5f5816, value=[B@307d5428, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@68e1acdb to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,065 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246576,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPSETPRIVS,scn=5535835,ssn=0},ts_ms=1666014246633,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPSETPRIVS" ("ESOWNER" VARCHAR2(32), "ESEXPTAB" VARCHAR2(32), "ESEXPCOL" VARCHAR2(32), "ESGRANTEE" VARCHAR2(32), "ESCRTPRIV" VARCHAR2(1), "ESUPDPRIV" VARCHAR2(1), CONSTRAINT "ESPRIVS_PKEY" PRIMARY KEY ("ESOWNER", "ESEXPTAB", "ESEXPCOL", "ESGRANTEE") ENABLE, CONSTRAINT "REF_PRIV_EXPR_SET" FOREIGN KEY ("ESOWNER", "ESEXPTAB", "ESEXPCOL") REFERENCES "EXFSYS"."EXF$EXPRSET" ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPSETPRIVS",table=Struct{primaryKeyColumnNames=[ESOWNER, ESEXPTAB, ESEXPCOL, ESGRANTEE],columns=[Struct{name=ESOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ESEXPTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=ESEXPCOL,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=ESGRANTEE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=ESCRTPRIV,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=ESUPDPRIV,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1,position=6,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,065 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246576,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPSETPRIVS,scn=5535835,ssn=0},ts_ms=1666014246633,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPSETPRIVS" ("ESOWNER" VARCHAR2(32), "ESEXPTAB" VARCHAR2(32), "ESEXPCOL" VARCHAR2(32), "ESGRANTEE" VARCHAR2(32), "ESCRTPRIV" VARCHAR2(1), "ESUPDPRIV" VARCHAR2(1), CONSTRAINT "ESPRIVS_PKEY" PRIMARY KEY ("ESOWNER", "ESEXPTAB", "ESEXPCOL", "ESGRANTEE") ENABLE, CONSTRAINT "REF_PRIV_EXPR_SET" FOREIGN KEY ("ESOWNER", "ESEXPTAB", "ESEXPCOL") REFERENCES "EXFSYS"."EXF$EXPRSET" ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPSETPRIVS",table=Struct{primaryKeyColumnNames=[ESOWNER, ESEXPTAB, ESEXPCOL, ESGRANTEE],columns=[Struct{name=ESOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ESEXPTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=ESEXPCOL,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=ESGRANTEE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=ESCRTPRIV,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=ESUPDPRIV,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1,position=6,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,065 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@1eac55aa, value=[B@442a8e43, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@21c86569 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,065 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,065 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3677) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=7414]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,066 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246635,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPSETSTATS,scn=5535835,ssn=0},ts_ms=1666014246713,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPSETSTATS" ("ESETOWNER" VARCHAR2(32) NOT NULL ENABLE, "ESETTABLE" VARCHAR2(32) NOT NULL ENABLE, "ESETCOLUMN" VARCHAR2(32) NOT NULL ENABLE, "PREDLHS" VARCHAR2(300) NOT NULL ENABLE, "NOEQPREDS" NUMBER, "NOLTPREDS" NUMBER, "NOGTPREDS" NUMBER, "NOLTEQPRS" NUMBER, "NOGTEQPRS" NUMBER, "NONEQPRS" NUMBER, "NOISNLPRS" NUMBER, "NOISNNLPRS" NUMBER, "NOBETPREDS" NUMBER, "NONVLPREDS" NUMBER, "NOLIKEPRS" NUMBER, CONSTRAINT "STATS_PKEY" PRIMARY KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN", "PREDLHS") ENABLE, CONSTRAINT "REF_STATS_KEY" FOREIGN KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN") REFERENCES "EXFSYS"."EXF$EXPRSET" ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPSETSTATS",table=Struct{primaryKeyColumnNames=[ESETOWNER, ESETTABLE, ESETCOLUMN, PREDLHS],columns=[Struct{name=ESETOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETTABLE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETCOLUMN,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PREDLHS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=NOEQPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=NONEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=NOBETPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=NONVLPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLIKEPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=7414, buffer=java.nio.HeapByteBuffer[pos=0 lim=7414 cap=7414]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,066 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246635,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$EXPSETSTATS,scn=5535835,ssn=0},ts_ms=1666014246713,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$EXPSETSTATS" ("ESETOWNER" VARCHAR2(32) NOT NULL ENABLE, "ESETTABLE" VARCHAR2(32) NOT NULL ENABLE, "ESETCOLUMN" VARCHAR2(32) NOT NULL ENABLE, "PREDLHS" VARCHAR2(300) NOT NULL ENABLE, "NOEQPREDS" NUMBER, "NOLTPREDS" NUMBER, "NOGTPREDS" NUMBER, "NOLTEQPRS" NUMBER, "NOGTEQPRS" NUMBER, "NONEQPRS" NUMBER, "NOISNLPRS" NUMBER, "NOISNNLPRS" NUMBER, "NOBETPREDS" NUMBER, "NONVLPREDS" NUMBER, "NOLIKEPRS" NUMBER, CONSTRAINT "STATS_PKEY" PRIMARY KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN", "PREDLHS") ENABLE, CONSTRAINT "REF_STATS_KEY" FOREIGN KEY ("ESETOWNER", "ESETTABLE", "ESETCOLUMN") REFERENCES "EXFSYS"."EXF$EXPRSET" ("EXSOWNER", "EXSTABNM", "EXSCOLNM") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$EXPSETSTATS",table=Struct{primaryKeyColumnNames=[ESETOWNER, ESETTABLE, ESETCOLUMN, PREDLHS],columns=[Struct{name=ESETOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETTABLE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=ESETCOLUMN,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PREDLHS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=NOEQPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=NOGTEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=NONEQPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=NOISNNLPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=NOBETPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=NONVLPREDS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=NOLIKEPRS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,066 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@c5b99f8, value=[B@1c9346a6, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@7e928b70 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,066 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246715,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$IDXSECOBJ,scn=5535835,ssn=0},ts_ms=1666014246776,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#" NUMBER NOT NULL ENABLE, "IDXOWNER" VARCHAR2(32) NOT NULL ENABLE, "IDXNAME" VARCHAR2(25) NOT NULL ENABLE, "IDXATTRSET" VARCHAR2(32) NOT NULL ENABLE, "IDXESETTAB" VARCHAR2(32) NOT NULL ENABLE, "IDXESETCOL" VARCHAR2(32) NOT NULL ENABLE, "IDXPREDTAB" VARCHAR2(32), "IDXACCFUNC" VARCHAR2(32), "IDXSTATUS" VARCHAR2(11), "OPTFCCPUCT" NUMBER, "OPTFCIOCT" NUMBER, "OPTIXSELVT" NUMBER, "OPTIXCPUCT" NUMBER, "OPTIXIOCT" NUMBER, "OPTPTFSCCT" NUMBER, "IDXPTABSTG" VARCHAR2(1000), "IDXPQUERY" CLOB, CONSTRAINT "DUPLICATE_IDX_NAME" PRIMARY KEY ("IDXOWNER", "IDXNAME") ENABLE, CONSTRAINT "DUPLICATE_IDX_OBJNO" UNIQUE ("IDXOBJ#") USING INDEX ENABLE ) ORGANIZATION INDEX NOCOMPRESS INCLUDING "OPTPTFSCCT" OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$IDXSECOBJ",table=Struct{primaryKeyColumnNames=[IDXOWNER, IDXNAME],columns=[Struct{name=IDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXATTRSET,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETCOL,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXPREDTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXACCFUNC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXSTATUS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=11,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXSELVT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTPTFSCCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPTABSTG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1000,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPQUERY,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=17,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,066 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246715,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$IDXSECOBJ,scn=5535835,ssn=0},ts_ms=1666014246776,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#" NUMBER NOT NULL ENABLE, "IDXOWNER" VARCHAR2(32) NOT NULL ENABLE, "IDXNAME" VARCHAR2(25) NOT NULL ENABLE, "IDXATTRSET" VARCHAR2(32) NOT NULL ENABLE, "IDXESETTAB" VARCHAR2(32) NOT NULL ENABLE, "IDXESETCOL" VARCHAR2(32) NOT NULL ENABLE, "IDXPREDTAB" VARCHAR2(32), "IDXACCFUNC" VARCHAR2(32), "IDXSTATUS" VARCHAR2(11), "OPTFCCPUCT" NUMBER, "OPTFCIOCT" NUMBER, "OPTIXSELVT" NUMBER, "OPTIXCPUCT" NUMBER, "OPTIXIOCT" NUMBER, "OPTPTFSCCT" NUMBER, "IDXPTABSTG" VARCHAR2(1000), "IDXPQUERY" CLOB, CONSTRAINT "DUPLICATE_IDX_NAME" PRIMARY KEY ("IDXOWNER", "IDXNAME") ENABLE, CONSTRAINT "DUPLICATE_IDX_OBJNO" UNIQUE ("IDXOBJ#") USING INDEX ENABLE ) ORGANIZATION INDEX NOCOMPRESS INCLUDING "OPTPTFSCCT" OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$IDXSECOBJ",table=Struct{primaryKeyColumnNames=[IDXOWNER, IDXNAME],columns=[Struct{name=IDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXOWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXATTRSET,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXESETCOL,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=false,autoIncremented=false,generated=false}, Struct{name=IDXPREDTAB,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXACCFUNC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXSTATUS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=11,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTFCIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXSELVT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXCPUCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIXIOCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTPTFSCCT,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=15,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPTABSTG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=1000,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=IDXPQUERY,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=17,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,066 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@3aea236a, value=[B@6a2ef996, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@6c0e3403 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,066 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,066 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246778,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$JAVAMSG,scn=5535835,ssn=0},ts_ms=1666014246833,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE GLOBAL TEMPORARY TABLE "EXFSYS"."EXF$JAVAMSG" ("CODE" VARCHAR2(15), "MESSAGE" VARCHAR2(500) ) ON COMMIT PRESERVE ROWS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$JAVAMSG",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=MESSAGE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=2,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,066 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3677): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11031, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3677 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,067 TRACE || Successfully produced messages to oracle-0 with base offset 11031. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,067 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11031 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,066 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246778,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$JAVAMSG,scn=5535835,ssn=0},ts_ms=1666014246833,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE GLOBAL TEMPORARY TABLE "EXFSYS"."EXF$JAVAMSG" ("CODE" VARCHAR2(15), "MESSAGE" VARCHAR2(500) ) ON COMMIT PRESERVE ROWS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$JAVAMSG",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=MESSAGE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=2,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,067 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@6db77935, value=[B@70c8cf5c, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@6d7b436a to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,067 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246834,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PARAMETER,scn=5535835,ssn=0},ts_ms=1666014246899,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PARAMETER" ("NUM" NUMBER, "NAME" VARCHAR2(64), "VALTYPE" NUMBER, "VALUE" VARCHAR2(512), CONSTRAINT "DUP_PARAMETER" PRIMARY KEY ("NUM") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PARAMETER",table=Struct{primaryKeyColumnNames=[NUM],columns=[Struct{name=NUM,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=64,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=VALTYPE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=VALUE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=512,position=4,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,067 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3678) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=15245]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=15245, buffer=java.nio.HeapByteBuffer[pos=0 lim=15245 cap=15245]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,067 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246834,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PARAMETER,scn=5535835,ssn=0},ts_ms=1666014246899,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PARAMETER" ("NUM" NUMBER, "NAME" VARCHAR2(64), "VALTYPE" NUMBER, "VALUE" VARCHAR2(512), CONSTRAINT "DUP_PARAMETER" PRIMARY KEY ("NUM") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PARAMETER",table=Struct{primaryKeyColumnNames=[NUM],columns=[Struct{name=NUM,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=64,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=VALTYPE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=VALUE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=512,position=4,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,067 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@4600f0f, value=[B@5248a067, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@64cc375d to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,067 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,067 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246901,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PLAN_TABLE,scn=5535835,ssn=0},ts_ms=1666014246962,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PLAN_TABLE" ("STATEMENT_ID" VARCHAR2(30), "PLAN_ID" NUMBER, "TIMESTAMP" DATE, "REMARKS" VARCHAR2(4000), "OPERATION" VARCHAR2(30), "OPTIONS" VARCHAR2(255), "OBJECT_NODE" VARCHAR2(128), "OBJECT_OWNER" VARCHAR2(30), "OBJECT_NAME" VARCHAR2(30), "OBJECT_ALIAS" VARCHAR2(65), "OBJECT_INSTANCE" NUMBER(*,0), "OBJECT_TYPE" VARCHAR2(30), "OPTIMIZER" VARCHAR2(255), "SEARCH_COLUMNS" NUMBER, "ID" NUMBER(*,0), "PARENT_ID" NUMBER(*,0), "DEPTH" NUMBER(*,0), "POSITION" NUMBER(*,0), "COST" NUMBER(*,0), "CARDINALITY" NUMBER(*,0), "BYTES" NUMBER(*,0), "OTHER_TAG" VARCHAR2(255), "PARTITION_START" VARCHAR2(255), "PARTITION_STOP" VARCHAR2(255), "PARTITION_ID" NUMBER(*,0), "OTHER" LONG, "DISTRIBUTION" VARCHAR2(30), "CPU_COST" NUMBER(*,0), "IO_COST" NUMBER(*,0), "TEMP_SPACE" NUMBER(*,0), "ACCESS_PREDICATES" VARCHAR2(4000), "FILTER_PREDICATES" VARCHAR2(4000), "PROJECTION" VARCHAR2(4000), "TIME" NUMBER(*,0), "QBLOCK_NAME" VARCHAR2(30), "OTHER_XML" CLOB, CONSTRAINT "PLAN_STMT_ID" PRIMARY KEY ("STATEMENT_ID", "ID") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PLAN_TABLE",table=Struct{primaryKeyColumnNames=[STATEMENT_ID, ID],columns=[Struct{name=STATEMENT_ID,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PLAN_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=TIMESTAMP,jdbcType=93,typeName=DATE,typeExpression=DATE,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=REMARKS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=OPERATION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIONS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=128,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_ALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_INSTANCE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_TYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIMIZER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=SEARCH_COLUMNS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=15,optional=false,autoIncremented=false,generated=false}, Struct{name=PARENT_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=DEPTH,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=17,optional=true,autoIncremented=false,generated=false}, Struct{name=POSITION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=18,optional=true,autoIncremented=false,generated=false}, Struct{name=COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=19,optional=true,autoIncremented=false,generated=false}, Struct{name=CARDINALITY,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=20,optional=true,autoIncremented=false,generated=false}, Struct{name=BYTES,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=21,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_TAG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=22,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_START,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=23,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_STOP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=24,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=25,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER,jdbcType=-1,typeName=LONG,typeExpression=LONG,length=0,position=26,optional=true,autoIncremented=false,generated=false}, Struct{name=DISTRIBUTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=27,optional=true,autoIncremented=false,generated=false}, Struct{name=CPU_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=28,optional=true,autoIncremented=false,generated=false}, Struct{name=IO_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=29,optional=true,autoIncremented=false,generated=false}, Struct{name=TEMP_SPACE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=30,optional=true,autoIncremented=false,generated=false}, Struct{name=ACCESS_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=31,optional=true,autoIncremented=false,generated=false}, Struct{name=FILTER_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=32,optional=true,autoIncremented=false,generated=false}, Struct{name=PROJECTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=33,optional=true,autoIncremented=false,generated=false}, Struct{name=TIME,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=34,optional=true,autoIncremented=false,generated=false}, Struct{name=QBLOCK_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=35,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_XML,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=36,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,068 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246901,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PLAN_TABLE,scn=5535835,ssn=0},ts_ms=1666014246962,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PLAN_TABLE" ("STATEMENT_ID" VARCHAR2(30), "PLAN_ID" NUMBER, "TIMESTAMP" DATE, "REMARKS" VARCHAR2(4000), "OPERATION" VARCHAR2(30), "OPTIONS" VARCHAR2(255), "OBJECT_NODE" VARCHAR2(128), "OBJECT_OWNER" VARCHAR2(30), "OBJECT_NAME" VARCHAR2(30), "OBJECT_ALIAS" VARCHAR2(65), "OBJECT_INSTANCE" NUMBER(*,0), "OBJECT_TYPE" VARCHAR2(30), "OPTIMIZER" VARCHAR2(255), "SEARCH_COLUMNS" NUMBER, "ID" NUMBER(*,0), "PARENT_ID" NUMBER(*,0), "DEPTH" NUMBER(*,0), "POSITION" NUMBER(*,0), "COST" NUMBER(*,0), "CARDINALITY" NUMBER(*,0), "BYTES" NUMBER(*,0), "OTHER_TAG" VARCHAR2(255), "PARTITION_START" VARCHAR2(255), "PARTITION_STOP" VARCHAR2(255), "PARTITION_ID" NUMBER(*,0), "OTHER" LONG, "DISTRIBUTION" VARCHAR2(30), "CPU_COST" NUMBER(*,0), "IO_COST" NUMBER(*,0), "TEMP_SPACE" NUMBER(*,0), "ACCESS_PREDICATES" VARCHAR2(4000), "FILTER_PREDICATES" VARCHAR2(4000), "PROJECTION" VARCHAR2(4000), "TIME" NUMBER(*,0), "QBLOCK_NAME" VARCHAR2(30), "OTHER_XML" CLOB, CONSTRAINT "PLAN_STMT_ID" PRIMARY KEY ("STATEMENT_ID", "ID") USING INDEX ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PLAN_TABLE",table=Struct{primaryKeyColumnNames=[STATEMENT_ID, ID],columns=[Struct{name=STATEMENT_ID,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PLAN_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=TIMESTAMP,jdbcType=93,typeName=DATE,typeExpression=DATE,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=REMARKS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=OPERATION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIONS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NODE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=128,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=8,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=9,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_ALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=10,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_INSTANCE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=11,optional=true,autoIncremented=false,generated=false}, Struct{name=OBJECT_TYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=12,optional=true,autoIncremented=false,generated=false}, Struct{name=OPTIMIZER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=13,optional=true,autoIncremented=false,generated=false}, Struct{name=SEARCH_COLUMNS,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=14,optional=true,autoIncremented=false,generated=false}, Struct{name=ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=15,optional=false,autoIncremented=false,generated=false}, Struct{name=PARENT_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=16,optional=true,autoIncremented=false,generated=false}, Struct{name=DEPTH,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=17,optional=true,autoIncremented=false,generated=false}, Struct{name=POSITION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=18,optional=true,autoIncremented=false,generated=false}, Struct{name=COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=19,optional=true,autoIncremented=false,generated=false}, Struct{name=CARDINALITY,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=20,optional=true,autoIncremented=false,generated=false}, Struct{name=BYTES,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=21,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_TAG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=22,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_START,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=23,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_STOP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=255,position=24,optional=true,autoIncremented=false,generated=false}, Struct{name=PARTITION_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=25,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER,jdbcType=-1,typeName=LONG,typeExpression=LONG,length=0,position=26,optional=true,autoIncremented=false,generated=false}, Struct{name=DISTRIBUTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=27,optional=true,autoIncremented=false,generated=false}, Struct{name=CPU_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=28,optional=true,autoIncremented=false,generated=false}, Struct{name=IO_COST,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=29,optional=true,autoIncremented=false,generated=false}, Struct{name=TEMP_SPACE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=30,optional=true,autoIncremented=false,generated=false}, Struct{name=ACCESS_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=31,optional=true,autoIncremented=false,generated=false}, Struct{name=FILTER_PREDICATES,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=32,optional=true,autoIncremented=false,generated=false}, Struct{name=PROJECTION,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=4000,position=33,optional=true,autoIncremented=false,generated=false}, Struct{name=TIME,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=38,scale=0,position=34,optional=true,autoIncremented=false,generated=false}, Struct{name=QBLOCK_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=30,position=35,optional=true,autoIncremented=false,generated=false}, Struct{name=OTHER_XML,jdbcType=2005,typeName=CLOB,typeExpression=CLOB,length=4000,position=36,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,068 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,068 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3678): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11032, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,068 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@57073da6, value=[B@405bf4b9, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@45284530 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,068 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,068 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,068 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3678 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,068 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246965,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PREDATTRMAP,scn=5535835,ssn=0},ts_ms=1666014247027,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PREDATTRMAP" ("PTIDXOBJ#" NUMBER NOT NULL ENABLE, "PTATTRSEXP" VARCHAR2(300) NOT NULL ENABLE, "PTATTRID" NUMBER, "PTATTRALIAS" VARCHAR2(25), "PTATTROPER" "EXFSYS"."EXF$INDEXOPER" , "PTATTRTYPE" VARCHAR2(65), "PTATTRPROP" NUMBER, "XMLTATTR" VARCHAR2(32) DEFAULT null, "XMLNSELP" NUMBER DEFAULT null, CONSTRAINT "DUP_ATTR_EXP" UNIQUE ("PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR") USING INDEX ENABLE, CONSTRAINT "REF_IDX_ATTRS" FOREIGN KEY ("PTIDXOBJ#") REFERENCES "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#") ON DELETE CASCADE ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PREDATTRMAP",table=Struct{primaryKeyColumnNames=[PTIDXOBJ#, PTATTRSEXP, PTATTRALIAS, XMLTATTR],columns=[Struct{name=PTIDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRSEXP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTROPER,jdbcType=2003,typeName=EXF$INDEXOPER,typeExpression=EXF$INDEXOPER,length=337,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRTYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRPROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=XMLTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}, Struct{name=XMLNSELP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,068 TRACE || Successfully produced messages to oracle-0 with base offset 11032. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,068 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11032 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,068 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11033 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,068 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014246965,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$PREDATTRMAP,scn=5535835,ssn=0},ts_ms=1666014247027,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$PREDATTRMAP" ("PTIDXOBJ#" NUMBER NOT NULL ENABLE, "PTATTRSEXP" VARCHAR2(300) NOT NULL ENABLE, "PTATTRID" NUMBER, "PTATTRALIAS" VARCHAR2(25), "PTATTROPER" "EXFSYS"."EXF$INDEXOPER" , "PTATTRTYPE" VARCHAR2(65), "PTATTRPROP" NUMBER, "XMLTATTR" VARCHAR2(32) DEFAULT null, "XMLNSELP" NUMBER DEFAULT null, CONSTRAINT "DUP_ATTR_EXP" UNIQUE ("PTIDXOBJ#", "PTATTRSEXP", "PTATTRALIAS", "XMLTATTR") USING INDEX ENABLE, CONSTRAINT "REF_IDX_ATTRS" FOREIGN KEY ("PTIDXOBJ#") REFERENCES "EXFSYS"."EXF$IDXSECOBJ" ("IDXOBJ#") ON DELETE CASCADE ENABLE ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$PREDATTRMAP",table=Struct{primaryKeyColumnNames=[PTIDXOBJ#, PTATTRSEXP, PTATTRALIAS, XMLTATTR],columns=[Struct{name=PTIDXOBJ#,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRSEXP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=300,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PTATTRID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRALIAS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=25,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTROPER,jdbcType=2003,typeName=EXF$INDEXOPER,typeExpression=EXF$INDEXOPER,length=337,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRTYPE,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=65,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=PTATTRPROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=XMLTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=8,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}, Struct{name=XMLNSELP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=9,optional=true,autoIncremented=false,generated=false,defaultValueExpression=null}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,069 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,068 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@227b5dac, value=[B@3d9a9b66, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@3a1c5820 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,069 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3679) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=14137]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,069 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Nothing to send to Kafka. Polling source for additional records [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=14137, buffer=java.nio.HeapByteBuffer[pos=0 lim=14137 cap=14137]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,069 DEBUG Oracle|oracle|oracle-connector-task polling records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,069 DEBUG Oracle|oracle|oracle-connector-task no records available or batch size not reached yet, sleeping a bit... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,069 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,070 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3679): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11034, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3679 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,070 TRACE || Successfully produced messages to oracle-0 with base offset 11034. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,070 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11034 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,070 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11035 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,070 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3680) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=5253]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=5253, buffer=java.nio.HeapByteBuffer[pos=0 lim=5253 cap=5253]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,070 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3680): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11036, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,070 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3680 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 TRACE || Successfully produced messages to oracle-0 with base offset 11036. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,071 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11036 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3681) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=15284]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=15284, buffer=java.nio.HeapByteBuffer[pos=0 lim=15284 cap=15284]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,071 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3681): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11037, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3681 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 TRACE || Successfully produced messages to oracle-0 with base offset 11037. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,071 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11037 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3682) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=7171]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=7171, buffer=java.nio.HeapByteBuffer[pos=0 lim=7171 cap=7171]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,071 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,072 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3682): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11038, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,072 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3682 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,072 TRACE || Successfully produced messages to oracle-0 with base offset 11038. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,072 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11038 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,072 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,073 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,073 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,074 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,075 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$VALIDIOPER','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,085 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,086 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$VALIDIOPER" ("OPERSTR" VARCHAR2(15) ) ;, tables=[columns: { OPERSTR VARCHAR2(15) DEFAULT VALUE NULL } primary key: [] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247086] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,086 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$VALIDIOPER" ("OPERSTR" VARCHAR2(15) ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,086 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247086, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VALIDIOPER\" \n (\t\"OPERSTR\" VARCHAR2(15)\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VALIDIOPER\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "OPERSTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 15, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,086 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247086, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VALIDIOPER\" \n (\t\"OPERSTR\" VARCHAR2(15)\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VALIDIOPER\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "OPERSTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 15, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,086 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,086 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,086 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,086 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,086 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,086 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5521) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=1089]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,086 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=1089, buffer=java.nio.HeapByteBuffer[pos=0 lim=1089 cap=1089]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,086 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,087 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5521): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11040, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,087 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5521 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,087 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11040. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,087 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,087 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11040 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,087 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247030,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDIOPER,scn=5535835,ssn=0},ts_ms=1666014247086,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDIOPER" ("OPERSTR" VARCHAR2(15) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDIOPER",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=OPERSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,087 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$VALIDPRIVS [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,087 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$VALIDPRIVS' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,139 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,139 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,140 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,140 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$VALIDPRIVS','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,151 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$VALIDPRIVS" ("CODE" NUMBER, "PRIVSTR" VARCHAR2(20) ) ;, tables=[columns: { CODE NUMBER(0) DEFAULT VALUE NULL PRIVSTR VARCHAR2(20) DEFAULT VALUE NULL } primary key: [] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247151] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,151 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$VALIDPRIVS" ("CODE" NUMBER, "PRIVSTR" VARCHAR2(20) ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247151, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VALIDPRIVS\" \n (\t\"CODE\" NUMBER, \n\t\"PRIVSTR\" VARCHAR2(20)\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VALIDPRIVS\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "CODE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PRIVSTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 20, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247151, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VALIDPRIVS\" \n (\t\"CODE\" NUMBER, \n\t\"PRIVSTR\" VARCHAR2(20)\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VALIDPRIVS\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "CODE", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "PRIVSTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 20, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,151 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,151 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,152 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,152 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5522) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=1492]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,152 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=1492, buffer=java.nio.HeapByteBuffer[pos=0 lim=1492 cap=1492]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,152 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,152 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5522): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11041, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,152 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5522 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,152 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11041. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,152 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,152 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11041 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,152 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247087,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDPRIVS,scn=5535835,ssn=0},ts_ms=1666014247151,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDPRIVS" ("CODE" NUMBER, "PRIVSTR" VARCHAR2(20) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDPRIVS",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=PRIVSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=20,position=2,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,152 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.EXF$VERSION [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,152 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='EXF$VERSION' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,158 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-3, correlationId=945): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=1626396239, responses=[]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,158 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 1626396239 with response=(), implied=(dbz_configs-0) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,158 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_configs-0 at position FetchPosition{offset=4, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,158 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Built incremental fetch (sessionId=1626396239, epoch=939) for node 1. Added (), altered (), removed (), replaced () out of (dbz_configs-0) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,158 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(), toForget=(), toReplace=(), implied=(dbz_configs-0), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,159 TRACE || [Consumer clientId=consumer-dbz-3, groupId=dbz] Polling for fetches with timeout 2147012453 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,159 DEBUG || [Consumer clientId=consumer-dbz-3, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-3, correlationId=946) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=1626396239, sessionEpoch=939, topics=[], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,159 TRACE || [Consumer clientId=consumer-dbz-3, groupId=dbz] Skipping fetch for partition dbz_configs-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,159 TRACE || [Consumer clientId=consumer-dbz-3, groupId=dbz] Polling for fetches with timeout 2147012452 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,192 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,193 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,193 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,194 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','EXF$VERSION','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,204 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,205 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."EXF$VERSION" ("EXFVERSION" NUMBER ) ;, tables=[columns: { EXFVERSION NUMBER(0) DEFAULT VALUE NULL } primary key: [] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247204] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,205 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."EXF$VERSION" ("EXFVERSION" NUMBER ) ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,205 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247204, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VERSION\" \n (\t\"EXFVERSION\" NUMBER\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VERSION\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "EXFVERSION", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,205 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247204, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"EXF$VERSION\" \n (\t\"EXFVERSION\" NUMBER\n ) ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"EXF$VERSION\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ ], "columns" : [ { "name" : "EXFVERSION", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,205 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,205 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,205 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,205 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,205 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,205 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5523) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=1077]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,205 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=1077, buffer=java.nio.HeapByteBuffer[pos=0 lim=1077 cap=1077]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,205 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,206 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5523): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11042, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,206 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5523 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,206 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11042. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,206 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,206 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11042 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,206 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247152,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VERSION,scn=5535835,ssn=0},ts_ms=1666014247204,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VERSION" ("EXFVERSION" NUMBER ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VERSION",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=EXFVERSION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,206 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$COLLGRPBYSPEC [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,206 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='RLM$COLLGRPBYSPEC' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-2, correlationId=946): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=833240796, responses=[]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 833240796 with response=(), implied=(dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-0 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-2 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-4 at position FetchPosition{offset=1, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-1 at position FetchPosition{offset=3, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_statuses-3 at position FetchPosition{offset=3, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Built incremental fetch (sessionId=833240796, epoch=940) for node 1. Added (), altered (), removed (), replaced () out of (dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(), toForget=(), toReplace=(), implied=(dbz_statuses-0, dbz_statuses-1, dbz_statuses-2, dbz_statuses-3, dbz_statuses-4), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Polling for fetches with timeout 2147023652 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,228 DEBUG || [Consumer clientId=consumer-dbz-2, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-2, correlationId=947) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=833240796, sessionEpoch=940, topics=[], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-2 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-4 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-1 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Skipping fetch for partition dbz_statuses-3 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,228 TRACE || [Consumer clientId=consumer-dbz-2, groupId=dbz] Polling for fetches with timeout 2147023652 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,247 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,247 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,248 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,248 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','RLM$COLLGRPBYSPEC','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,263 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,264 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."RLM$COLLGRPBYSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "PRIMEVTTP" VARCHAR2(32), "GRPATTIDX" NUMBER, "ATTRNAME" VARCHAR2(100), "EVTATTR" VARCHAR2(32), CONSTRAINT "RLM$GRPBYSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME") ENABLE, CONSTRAINT "RLM$GRPBYSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;, tables=[columns: { RSET_OWNER VARCHAR2(32) NOT NULL RSET_NAME VARCHAR2(32) NOT NULL PRIMEVTTP VARCHAR2(32) NOT NULL GRPATTIDX NUMBER(0) DEFAULT VALUE NULL ATTRNAME VARCHAR2(100) NOT NULL EVTATTR VARCHAR2(32) DEFAULT VALUE NULL } primary key: [RSET_OWNER, RSET_NAME, PRIMEVTTP, ATTRNAME] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247264] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,264 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."RLM$COLLGRPBYSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "PRIMEVTTP" VARCHAR2(32), "GRPATTIDX" NUMBER, "ATTRNAME" VARCHAR2(100), "EVTATTR" VARCHAR2(32), CONSTRAINT "RLM$GRPBYSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME") ENABLE, CONSTRAINT "RLM$GRPBYSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,264 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247264, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$COLLGRPBYSPEC\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"PRIMEVTTP\" VARCHAR2(32), \n\t\"GRPATTIDX\" NUMBER, \n\t\"ATTRNAME\" VARCHAR2(100), \n\t\"EVTATTR\" VARCHAR2(32), \n\t CONSTRAINT \"RLM$GRPBYSPECPK\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"PRIMEVTTP\", \"ATTRNAME\") ENABLE, \n\t CONSTRAINT \"RLM$GRPBYSPECFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$COLLGRPBYSPEC\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PRIMEVTTP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "GRPATTIDX", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ATTRNAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 100, "position" : 5, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVTATTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,264 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247264, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$COLLGRPBYSPEC\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"PRIMEVTTP\" VARCHAR2(32), \n\t\"GRPATTIDX\" NUMBER, \n\t\"ATTRNAME\" VARCHAR2(100), \n\t\"EVTATTR\" VARCHAR2(32), \n\t CONSTRAINT \"RLM$GRPBYSPECPK\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"PRIMEVTTP\", \"ATTRNAME\") ENABLE, \n\t CONSTRAINT \"RLM$GRPBYSPECFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$COLLGRPBYSPEC\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PRIMEVTTP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "GRPATTIDX", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ATTRNAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 100, "position" : 5, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVTATTR", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,264 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,264 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,264 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,264 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,264 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,264 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5524) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=3636]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,264 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=3636, buffer=java.nio.HeapByteBuffer[pos=0 lim=3636 cap=3636]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,264 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,265 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5524): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11043, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,265 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5524 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,265 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11043. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,265 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,265 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11043 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,265 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247206,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$COLLGRPBYSPEC,scn=5535835,ssn=0},ts_ms=1666014247264,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$COLLGRPBYSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "PRIMEVTTP" VARCHAR2(32), "GRPATTIDX" NUMBER, "ATTRNAME" VARCHAR2(100), "EVTATTR" VARCHAR2(32), CONSTRAINT "RLM$GRPBYSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME") ENABLE, CONSTRAINT "RLM$GRPBYSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$COLLGRPBYSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, PRIMEVTTP, ATTRNAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PRIMEVTTP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=GRPATTIDX,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=ATTRNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=EVTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,265 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$DMLEVTTRIGS [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,265 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='RLM$DMLEVTTRIGS' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,305 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,305 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,306 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,306 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','RLM$DMLEVTTRIGS','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,321 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,321 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."RLM$DMLEVTTRIGS" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "TATAB_NAME" VARCHAR2(32), "PRESTTRIG" VARCHAR2(32), "DMLROWTRIG" VARCHAR2(32), "POSTSTTRIG" VARCHAR2(32), "DBCNFREGID" NUMBER, "DBCNFCBKPRC" VARCHAR2(75), CONSTRAINT "RLM$DMLEVTPKEY" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "TATAB_NAME") ENABLE, CONSTRAINT "RLM$DMLEVTREFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;, tables=[columns: { RSET_OWNER VARCHAR2(32) NOT NULL RSET_NAME VARCHAR2(32) NOT NULL TATAB_NAME VARCHAR2(32) NOT NULL PRESTTRIG VARCHAR2(32) DEFAULT VALUE NULL DMLROWTRIG VARCHAR2(32) DEFAULT VALUE NULL POSTSTTRIG VARCHAR2(32) DEFAULT VALUE NULL DBCNFREGID NUMBER(0) DEFAULT VALUE NULL DBCNFCBKPRC VARCHAR2(75) DEFAULT VALUE NULL } primary key: [RSET_OWNER, RSET_NAME, TATAB_NAME] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247321] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,321 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."RLM$DMLEVTTRIGS" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "TATAB_NAME" VARCHAR2(32), "PRESTTRIG" VARCHAR2(32), "DMLROWTRIG" VARCHAR2(32), "POSTSTTRIG" VARCHAR2(32), "DBCNFREGID" NUMBER, "DBCNFCBKPRC" VARCHAR2(75), CONSTRAINT "RLM$DMLEVTPKEY" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "TATAB_NAME") ENABLE, CONSTRAINT "RLM$DMLEVTREFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,322 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247321, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$DMLEVTTRIGS\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"TATAB_NAME\" VARCHAR2(32), \n\t\"PRESTTRIG\" VARCHAR2(32), \n\t\"DMLROWTRIG\" VARCHAR2(32), \n\t\"POSTSTTRIG\" VARCHAR2(32), \n\t\"DBCNFREGID\" NUMBER, \n\t\"DBCNFCBKPRC\" VARCHAR2(75), \n\t CONSTRAINT \"RLM$DMLEVTPKEY\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"TATAB_NAME\") ENABLE, \n\t CONSTRAINT \"RLM$DMLEVTREFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$DMLEVTTRIGS\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "TATAB_NAME" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "TATAB_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PRESTTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DMLROWTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "POSTSTTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DBCNFREGID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DBCNFCBKPRC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 75, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,322 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247321, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$DMLEVTTRIGS\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"TATAB_NAME\" VARCHAR2(32), \n\t\"PRESTTRIG\" VARCHAR2(32), \n\t\"DMLROWTRIG\" VARCHAR2(32), \n\t\"POSTSTTRIG\" VARCHAR2(32), \n\t\"DBCNFREGID\" NUMBER, \n\t\"DBCNFCBKPRC\" VARCHAR2(75), \n\t CONSTRAINT \"RLM$DMLEVTPKEY\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"TATAB_NAME\") ENABLE, \n\t CONSTRAINT \"RLM$DMLEVTREFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$DMLEVTTRIGS\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "TATAB_NAME" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "TATAB_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "PRESTTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DMLROWTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "POSTSTTRIG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DBCNFREGID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 7, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "DBCNFCBKPRC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 75, "position" : 8, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,322 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,322 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,322 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,322 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,322 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,322 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5525) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=4431]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,322 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=4431, buffer=java.nio.HeapByteBuffer[pos=0 lim=4431 cap=4431]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,322 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,323 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5525): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11044, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,323 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5525 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,323 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11044. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,323 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,323 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11044 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,323 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247265,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$DMLEVTTRIGS,scn=5535835,ssn=0},ts_ms=1666014247321,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$DMLEVTTRIGS" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "TATAB_NAME" VARCHAR2(32), "PRESTTRIG" VARCHAR2(32), "DMLROWTRIG" VARCHAR2(32), "POSTSTTRIG" VARCHAR2(32), "DBCNFREGID" NUMBER, "DBCNFCBKPRC" VARCHAR2(75), CONSTRAINT "RLM$DMLEVTPKEY" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "TATAB_NAME") ENABLE, CONSTRAINT "RLM$DMLEVTREFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$DMLEVTTRIGS",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, TATAB_NAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=TATAB_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PRESTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=DMLROWTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=POSTSTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFREGID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFCBKPRC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=8,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,323 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$EQUALSPEC [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,323 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='RLM$EQUALSPEC' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=947): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=1518104247, responses=[]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 1518104247 with response=(), implied=(dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-7 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-13 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-11 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-1 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-5 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-3 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-23 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-17 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-15 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-21 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-19 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-10 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-8 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-14 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-12 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-2 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-0 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-6 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-4 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-24 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-18 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-16 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-22 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-20 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-9 at position FetchPosition{offset=12, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Built incremental fetch (sessionId=1518104247, epoch=941) for node 1. Added (), altered (), removed (), replaced () out of (dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(), toForget=(), toReplace=(), implied=(dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,348 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Polling for fetches with timeout 2147432442 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,348 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=948) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=1518104247, sessionEpoch=941, topics=[], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-7 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-13 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-11 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-1 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-5 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-3 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-23 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-17 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-15 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-21 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-19 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-10 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-8 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-14 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-12 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-2 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-6 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-4 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-24 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-18 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-16 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-22 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-20 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-9 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,349 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Polling for fetches with timeout 2147432441 [org.apache.kafka.clients.consumer.KafkaConsumer] 2022-10-17 13:44:07,367 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,368 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,369 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,369 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','RLM$EQUALSPEC','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,403 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,404 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."RLM$EQUALSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "OPCODE_ID" NUMBER, "EQLKEYTYP" VARCHAR2(100), "EQSPECFLG" NUMBER, "EQUALATTRS" "EXFSYS"."RLM$EQUALATTR" , CONSTRAINT "RLM$EQSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "OPCODE_ID") ENABLE, CONSTRAINT "RLM$EQUALSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;, tables=[columns: { RSET_OWNER VARCHAR2(32) NOT NULL RSET_NAME VARCHAR2(32) NOT NULL OPCODE_ID NUMBER(0) NOT NULL EQLKEYTYP VARCHAR2(100) DEFAULT VALUE NULL EQSPECFLG NUMBER(0) DEFAULT VALUE NULL EQUALATTRS RLM$EQUALATTR(1073) DEFAULT VALUE NULL } primary key: [RSET_OWNER, RSET_NAME, OPCODE_ID] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247404] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,404 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."RLM$EQUALSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "OPCODE_ID" NUMBER, "EQLKEYTYP" VARCHAR2(100), "EQSPECFLG" NUMBER, "EQUALATTRS" "EXFSYS"."RLM$EQUALATTR" , CONSTRAINT "RLM$EQSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "OPCODE_ID") ENABLE, CONSTRAINT "RLM$EQUALSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,404 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247404, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$EQUALSPEC\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"OPCODE_ID\" NUMBER, \n\t\"EQLKEYTYP\" VARCHAR2(100), \n\t\"EQSPECFLG\" NUMBER, \n\t\"EQUALATTRS\" \"EXFSYS\".\"RLM$EQUALATTR\" , \n\t CONSTRAINT \"RLM$EQSPECPK\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"OPCODE_ID\") ENABLE, \n\t CONSTRAINT \"RLM$EQUALSPECFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$EQUALSPEC\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "OPCODE_ID" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "OPCODE_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EQLKEYTYP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 100, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EQSPECFLG", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EQUALATTRS", "jdbcType" : 2003, "typeName" : "RLM$EQUALATTR", "typeExpression" : "RLM$EQUALATTR", "charsetName" : null, "length" : 1073, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,404 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247404, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$EQUALSPEC\" \n (\t\"RSET_OWNER\" VARCHAR2(32), \n\t\"RSET_NAME\" VARCHAR2(32), \n\t\"OPCODE_ID\" NUMBER, \n\t\"EQLKEYTYP\" VARCHAR2(100), \n\t\"EQSPECFLG\" NUMBER, \n\t\"EQUALATTRS\" \"EXFSYS\".\"RLM$EQUALATTR\" , \n\t CONSTRAINT \"RLM$EQSPECPK\" PRIMARY KEY (\"RSET_OWNER\", \"RSET_NAME\", \"OPCODE_ID\") ENABLE, \n\t CONSTRAINT \"RLM$EQUALSPECFK\" FOREIGN KEY (\"RSET_OWNER\", \"RSET_NAME\")\n\t REFERENCES \"EXFSYS\".\"RLM$RULESET\" (\"RSET_OWNER\", \"RSET_NAME\") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$EQUALSPEC\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "RSET_OWNER", "RSET_NAME", "OPCODE_ID" ], "columns" : [ { "name" : "RSET_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "RSET_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "OPCODE_ID", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EQLKEYTYP", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 100, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EQSPECFLG", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EQUALATTRS", "jdbcType" : 2003, "typeName" : "RLM$EQUALATTR", "typeExpression" : "RLM$EQUALATTR", "charsetName" : null, "length" : 1073, "position" : 6, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,404 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,404 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,404 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,404 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,405 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,405 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5526) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=3624]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,405 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=3624, buffer=java.nio.HeapByteBuffer[pos=0 lim=3624 cap=3624]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,405 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,405 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5526): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11045, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,405 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5526 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,405 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11045. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,405 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,405 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11045 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,406 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247323,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EQUALSPEC,scn=5535835,ssn=0},ts_ms=1666014247404,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EQUALSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "OPCODE_ID" NUMBER, "EQLKEYTYP" VARCHAR2(100), "EQSPECFLG" NUMBER, "EQUALATTRS" "EXFSYS"."RLM$EQUALATTR" , CONSTRAINT "RLM$EQSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "OPCODE_ID") ENABLE, CONSTRAINT "RLM$EQUALSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EQUALSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, OPCODE_ID],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=OPCODE_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=EQLKEYTYP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EQSPECFLG,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=EQUALATTRS,jdbcType=2003,typeName=RLM$EQUALATTR,typeExpression=RLM$EQUALATTR,length=1073,position=6,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,406 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$ERRCODE [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,406 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='RLM$ERRCODE' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,450 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,451 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,451 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,452 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','RLM$ERRCODE','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,467 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,467 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."RLM$ERRCODE" ("ERRNO" NUMBER, "ERRMSG" VARCHAR2(200), "ERRDSC" VARCHAR2(500), "ERRACT" VARCHAR2(500), CONSTRAINT "RLM$ERRCODEPK" PRIMARY KEY ("ERRNO") ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;, tables=[columns: { ERRNO NUMBER(0) NOT NULL ERRMSG VARCHAR2(200) DEFAULT VALUE NULL ERRDSC VARCHAR2(500) DEFAULT VALUE NULL ERRACT VARCHAR2(500) DEFAULT VALUE NULL } primary key: [ERRNO] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247467] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,467 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."RLM$ERRCODE" ("ERRNO" NUMBER, "ERRMSG" VARCHAR2(200), "ERRDSC" VARCHAR2(500), "ERRACT" VARCHAR2(500), CONSTRAINT "RLM$ERRCODEPK" PRIMARY KEY ("ERRNO") ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,467 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247467, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$ERRCODE\" \n (\t\"ERRNO\" NUMBER, \n\t\"ERRMSG\" VARCHAR2(200), \n\t\"ERRDSC\" VARCHAR2(500), \n\t\"ERRACT\" VARCHAR2(500), \n\t CONSTRAINT \"RLM$ERRCODEPK\" PRIMARY KEY (\"ERRNO\") ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$ERRCODE\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "ERRNO" ], "columns" : [ { "name" : "ERRNO", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "ERRMSG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 200, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ERRDSC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ERRACT", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,468 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247467, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$ERRCODE\" \n (\t\"ERRNO\" NUMBER, \n\t\"ERRMSG\" VARCHAR2(200), \n\t\"ERRDSC\" VARCHAR2(500), \n\t\"ERRACT\" VARCHAR2(500), \n\t CONSTRAINT \"RLM$ERRCODEPK\" PRIMARY KEY (\"ERRNO\") ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$ERRCODE\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "ERRNO" ], "columns" : [ { "name" : "ERRNO", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "ERRMSG", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 200, "position" : 2, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ERRDSC", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "ERRACT", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 500, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,468 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,468 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,468 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,468 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,468 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,468 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5527) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=2445]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,468 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=2445, buffer=java.nio.HeapByteBuffer[pos=0 lim=2445 cap=2445]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,468 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,469 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5527): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11046, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,469 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5527 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,469 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11046. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,469 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,469 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11046 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,469 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247406,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$ERRCODE,scn=5535835,ssn=0},ts_ms=1666014247467,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$ERRCODE" ("ERRNO" NUMBER, "ERRMSG" VARCHAR2(200), "ERRDSC" VARCHAR2(500), "ERRACT" VARCHAR2(500), CONSTRAINT "RLM$ERRCODEPK" PRIMARY KEY ("ERRNO") ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$ERRCODE",table=Struct{primaryKeyColumnNames=[ERRNO],columns=[Struct{name=ERRNO,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ERRMSG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=200,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRDSC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRACT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=4,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,469 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$EVENTSTRUCT [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,469 TRACE Oracle|oracle|snapshot running 'SELECT COUNT(1) FROM ALL_ALL_TABLES WHERE OWNER='EXFSYS' AND TABLE_NAME='RLM$EVENTSTRUCT' AND TABLE_TYPE IS NULL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,513 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'STORAGE', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,513 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SEGMENT_ATTRIBUTES', false); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,514 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'SQLTERMINATOR', true); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,514 TRACE Oracle|oracle|snapshot running 'SELECT dbms_metadata.get_ddl('TABLE','RLM$EVENTSTRUCT','EXFSYS') FROM DUAL' [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,529 TRACE Oracle|oracle|snapshot Executing statement begin dbms_metadata.set_transform_param(DBMS_METADATA.SESSION_TRANSFORM, 'DEFAULT'); end; [io.debezium.jdbc.JdbcConnection] 2022-10-17 13:44:07,530 DEBUG Oracle|oracle|snapshot Applying schema change event SchemaChangeEvent [database=MIS, schema=EXFSYS, ddl= CREATE TABLE "EXFSYS"."RLM$EVENTSTRUCT" ("EVST_OWNER" VARCHAR2(32), "EVST_NAME" VARCHAR2(32), "EVST_PROP" NUMBER, "EVST_PRCT" VARCHAR2(32), "EVST_PRCTTLS" VARCHAR2(75), CONSTRAINT "RLM$EVST_PRIMKEY" PRIMARY KEY ("EVST_OWNER", "EVST_NAME") ENABLE, CONSTRAINT "RLM$EVST_REFKEY" FOREIGN KEY ("EVST_OWNER", "EVST_NAME") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;, tables=[columns: { EVST_OWNER VARCHAR2(32) NOT NULL EVST_NAME VARCHAR2(32) NOT NULL EVST_PROP NUMBER(0) DEFAULT VALUE NULL EVST_PRCT VARCHAR2(32) DEFAULT VALUE NULL EVST_PRCTTLS VARCHAR2(75) DEFAULT VALUE NULL } primary key: [EVST_OWNER, EVST_NAME] default charset: null comment: null attributes: { } ], type=CREATE, ts_ms=1666014247530] [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,530 DEBUG Oracle|oracle|snapshot Recorded DDL statements for database 'MIS': CREATE TABLE "EXFSYS"."RLM$EVENTSTRUCT" ("EVST_OWNER" VARCHAR2(32), "EVST_NAME" VARCHAR2(32), "EVST_PROP" NUMBER, "EVST_PRCT" VARCHAR2(32), "EVST_PRCTTLS" VARCHAR2(75), CONSTRAINT "RLM$EVST_PRIMKEY" PRIMARY KEY ("EVST_OWNER", "EVST_NAME") ENABLE, CONSTRAINT "RLM$EVST_REFKEY" FOREIGN KEY ("EVST_OWNER", "EVST_NAME") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ; [io.debezium.connector.oracle.OracleDatabaseSchema] 2022-10-17 13:44:07,530 TRACE Oracle|oracle|snapshot Storing record into database schema history: { "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247530, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$EVENTSTRUCT\" \n (\t\"EVST_OWNER\" VARCHAR2(32), \n\t\"EVST_NAME\" VARCHAR2(32), \n\t\"EVST_PROP\" NUMBER, \n\t\"EVST_PRCT\" VARCHAR2(32), \n\t\"EVST_PRCTTLS\" VARCHAR2(75), \n\t CONSTRAINT \"RLM$EVST_PRIMKEY\" PRIMARY KEY (\"EVST_OWNER\", \"EVST_NAME\") ENABLE, \n\t CONSTRAINT \"RLM$EVST_REFKEY\" FOREIGN KEY (\"EVST_OWNER\", \"EVST_NAME\")\n\t REFERENCES \"EXFSYS\".\"EXF$ATTRSET\" (\"ATSOWNER\", \"ATSNAME\") ON DELETE CASCADE ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$EVENTSTRUCT\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "EVST_OWNER", "EVST_NAME" ], "columns" : [ { "name" : "EVST_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVST_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVST_PROP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EVST_PRCT", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EVST_PRCTTLS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 75, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] } [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,530 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Attempting to append record ProducerRecord(topic=mis-oracle-refbook.schema_changes, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=null, value={ "source" : { "server" : "oracle" }, "position" : { "snapshot_scn" : "5535835", "snapshot" : true, "scn" : "5535835", "snapshot_completed" : false }, "ts_ms" : 1666014247530, "databaseName" : "MIS", "schemaName" : "EXFSYS", "ddl" : "\n CREATE TABLE \"EXFSYS\".\"RLM$EVENTSTRUCT\" \n (\t\"EVST_OWNER\" VARCHAR2(32), \n\t\"EVST_NAME\" VARCHAR2(32), \n\t\"EVST_PROP\" NUMBER, \n\t\"EVST_PRCT\" VARCHAR2(32), \n\t\"EVST_PRCTTLS\" VARCHAR2(75), \n\t CONSTRAINT \"RLM$EVST_PRIMKEY\" PRIMARY KEY (\"EVST_OWNER\", \"EVST_NAME\") ENABLE, \n\t CONSTRAINT \"RLM$EVST_REFKEY\" FOREIGN KEY (\"EVST_OWNER\", \"EVST_NAME\")\n\t REFERENCES \"EXFSYS\".\"EXF$ATTRSET\" (\"ATSOWNER\", \"ATSNAME\") ON DELETE CASCADE ENABLE\n ) ORGANIZATION INDEX NOCOMPRESS ;", "tableChanges" : [ { "type" : "CREATE", "id" : "\"MIS\".\"EXFSYS\".\"RLM$EVENTSTRUCT\"", "table" : { "defaultCharsetName" : null, "primaryKeyColumnNames" : [ "EVST_OWNER", "EVST_NAME" ], "columns" : [ { "name" : "EVST_OWNER", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 1, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVST_NAME", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 2, "optional" : false, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : false, "enumValues" : [ ] }, { "name" : "EVST_PROP", "jdbcType" : 2, "typeName" : "NUMBER", "typeExpression" : "NUMBER", "charsetName" : null, "length" : 0, "position" : 3, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EVST_PRCT", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 32, "position" : 4, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] }, { "name" : "EVST_PRCTTLS", "jdbcType" : 12, "typeName" : "VARCHAR2", "typeExpression" : "VARCHAR2", "charsetName" : null, "length" : 75, "position" : 5, "optional" : true, "autoIncremented" : false, "generated" : false, "comment" : null, "hasDefaultValue" : true, "enumValues" : [ ] } ], "attributes" : [ ] }, "comment" : null } ] }, timestamp=null) with callback null to topic mis-oracle-refbook.schema_changes partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,530 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Allocating a new 32768 byte message buffer for topic mis-oracle-refbook.schema_changes partition 0 with remaining timeout 10000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,530 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Waking up the sender since topic mis-oracle-refbook.schema_changes partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,530 TRACE Oracle|oracle|snapshot [Producer clientId=oracle-schemahistory] Flushing accumulated records in producer. [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,530 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,530 TRACE || [Producer clientId=oracle-schemahistory] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,530 DEBUG || [Producer clientId=oracle-schemahistory] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5528) and timeout 30000 to node 1: {acks=1,timeout=30000,partitionSizes=[mis-oracle-refbook.schema_changes-0=3117]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,530 TRACE || [Producer clientId=oracle-schemahistory] Sent produce request to 1: (type=ProduceRequest, acks=1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=3117, buffer=java.nio.HeapByteBuffer[pos=0 lim=3117 cap=3117]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,530 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,531 DEBUG || [Producer clientId=oracle-schemahistory] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=oracle-schemahistory, correlationId=5528): ProduceResponseData(responses=[TopicProduceResponse(name='mis-oracle-refbook.schema_changes', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11047, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,531 TRACE || [Producer clientId=oracle-schemahistory] Received produce response from node 1 with correlation id 5528 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,531 TRACE || Successfully produced messages to mis-oracle-refbook.schema_changes-0 with base offset 11047. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,531 TRACE || [Producer clientId=oracle-schemahistory] The number of partitions is too small: available=1, all=1, not using adaptive for topic mis-oracle-refbook.schema_changes [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,531 DEBUG Oracle|oracle|snapshot Stored record in topic 'mis-oracle-refbook.schema_changes' partition 0 at offset 11047 [io.debezium.storage.kafka.history.KafkaSchemaHistory] 2022-10-17 13:44:07,531 TRACE Oracle|oracle|snapshot Enqueuing source record 'DataChangeEvent [record=SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247469,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EVENTSTRUCT,scn=5535835,ssn=0},ts_ms=1666014247530,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EVENTSTRUCT" ("EVST_OWNER" VARCHAR2(32), "EVST_NAME" VARCHAR2(32), "EVST_PROP" NUMBER, "EVST_PRCT" VARCHAR2(32), "EVST_PRCTTLS" VARCHAR2(75), CONSTRAINT "RLM$EVST_PRIMKEY" PRIMARY KEY ("EVST_OWNER", "EVST_NAME") ENABLE, CONSTRAINT "RLM$EVST_REFKEY" FOREIGN KEY ("EVST_OWNER", "EVST_NAME") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EVENTSTRUCT",table=Struct{primaryKeyColumnNames=[EVST_OWNER, EVST_NAME],columns=[Struct{name=EVST_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_PROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCTTLS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=5,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)}]' [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,531 INFO Oracle|oracle|snapshot Capturing structure of table MIS.EXFSYS.RLM$INCRRRSCHACT [io.debezium.connector.oracle.OracleSnapshotChangeEventSource] 2022-10-17 13:44:07,532 INFO Oracle|oracle|snapshot Snapshot - Final stage [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2022-10-17 13:44:07,532 WARN Oracle|oracle|snapshot Snapshot was not completed successfully, it will be re-executed upon connector restart [io.debezium.pipeline.source.AbstractSnapshotChangeEventSource] 2022-10-17 13:44:07,532 ERROR Oracle|oracle|snapshot Producer failure [io.debezium.pipeline.ErrorHandler] io.debezium.DebeziumException: java.lang.NullPointerException at io.debezium.pipeline.source.AbstractSnapshotChangeEventSource.execute(AbstractSnapshotChangeEventSource.java:85) at io.debezium.pipeline.ChangeEventSourceCoordinator.doSnapshot(ChangeEventSourceCoordinator.java:155) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:137) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:109) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.NullPointerException at io.debezium.connector.oracle.OracleSnapshotChangeEventSource.createSchemaChangeEventsForTables(OracleSnapshotChangeEventSource.java:215) at io.debezium.relational.RelationalSnapshotChangeEventSource.doExecute(RelationalSnapshotChangeEventSource.java:122) at io.debezium.pipeline.source.AbstractSnapshotChangeEventSource.execute(AbstractSnapshotChangeEventSource.java:76) ... 8 more 2022-10-17 13:44:07,533 INFO Oracle|oracle|snapshot Connected metrics set to 'false' [io.debezium.pipeline.ChangeEventSourceCoordinator] 2022-10-17 13:44:07,570 DEBUG Oracle|oracle|oracle-connector-task checking for more records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,570 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} About to send 8 records to Kafka [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,570 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247030,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDIOPER,scn=5535835,ssn=0},ts_ms=1666014247086,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDIOPER" ("OPERSTR" VARCHAR2(15) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDIOPER",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=OPERSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,570 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247030,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDIOPER,scn=5535835,ssn=0},ts_ms=1666014247086,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDIOPER" ("OPERSTR" VARCHAR2(15) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDIOPER",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=OPERSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=15,position=1,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,570 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,570 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@732075f0, value=[B@1a889159, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@420d0770 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,570 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,570 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,570 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247087,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDPRIVS,scn=5535835,ssn=0},ts_ms=1666014247151,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDPRIVS" ("CODE" NUMBER, "PRIVSTR" VARCHAR2(20) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDPRIVS",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=PRIVSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=20,position=2,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,570 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,571 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247087,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VALIDPRIVS,scn=5535835,ssn=0},ts_ms=1666014247151,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VALIDPRIVS" ("CODE" NUMBER, "PRIVSTR" VARCHAR2(20) ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VALIDPRIVS",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=CODE,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}, Struct{name=PRIVSTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=20,position=2,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,571 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,571 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,571 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3683) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=4283]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,571 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@4f39db4b, value=[B@11e589ac, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@3b792a67 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,571 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,571 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,571 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=4283, buffer=java.nio.HeapByteBuffer[pos=0 lim=4283 cap=4283]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,571 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247152,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VERSION,scn=5535835,ssn=0},ts_ms=1666014247204,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VERSION" ("EXFVERSION" NUMBER ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VERSION",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=EXFVERSION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,572 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,572 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247152,snapshot=true,db=MIS,schema=EXFSYS,table=EXF$VERSION,scn=5535835,ssn=0},ts_ms=1666014247204,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."EXF$VERSION" ("EXFVERSION" NUMBER ) ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."EXF$VERSION",table=Struct{primaryKeyColumnNames=[],columns=[Struct{name=EXFVERSION,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,572 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,572 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@30e2ae50, value=[B@df7f909, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@7f2fc0c2 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,572 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247206,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$COLLGRPBYSPEC,scn=5535835,ssn=0},ts_ms=1666014247264,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$COLLGRPBYSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "PRIMEVTTP" VARCHAR2(32), "GRPATTIDX" NUMBER, "ATTRNAME" VARCHAR2(100), "EVTATTR" VARCHAR2(32), CONSTRAINT "RLM$GRPBYSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME") ENABLE, CONSTRAINT "RLM$GRPBYSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$COLLGRPBYSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, PRIMEVTTP, ATTRNAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PRIMEVTTP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=GRPATTIDX,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=ATTRNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=EVTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,573 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247206,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$COLLGRPBYSPEC,scn=5535835,ssn=0},ts_ms=1666014247264,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$COLLGRPBYSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "PRIMEVTTP" VARCHAR2(32), "GRPATTIDX" NUMBER, "ATTRNAME" VARCHAR2(100), "EVTATTR" VARCHAR2(32), CONSTRAINT "RLM$GRPBYSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "PRIMEVTTP", "ATTRNAME") ENABLE, CONSTRAINT "RLM$GRPBYSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$COLLGRPBYSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, PRIMEVTTP, ATTRNAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=PRIMEVTTP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=GRPATTIDX,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=ATTRNAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=5,optional=false,autoIncremented=false,generated=false}, Struct{name=EVTATTR,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,573 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,573 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3683): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11039, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,574 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3683 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,574 TRACE || Successfully produced messages to oracle-0 with base offset 11039. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,574 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11039 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,573 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@48dd9207, value=[B@6de83562, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@5f65b667 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,574 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,575 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247265,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$DMLEVTTRIGS,scn=5535835,ssn=0},ts_ms=1666014247321,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$DMLEVTTRIGS" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "TATAB_NAME" VARCHAR2(32), "PRESTTRIG" VARCHAR2(32), "DMLROWTRIG" VARCHAR2(32), "POSTSTTRIG" VARCHAR2(32), "DBCNFREGID" NUMBER, "DBCNFCBKPRC" VARCHAR2(75), CONSTRAINT "RLM$DMLEVTPKEY" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "TATAB_NAME") ENABLE, CONSTRAINT "RLM$DMLEVTREFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$DMLEVTTRIGS",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, TATAB_NAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=TATAB_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PRESTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=DMLROWTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=POSTSTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFREGID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFCBKPRC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=8,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,575 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,575 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3684) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=14985]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,575 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247265,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$DMLEVTTRIGS,scn=5535835,ssn=0},ts_ms=1666014247321,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$DMLEVTTRIGS" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "TATAB_NAME" VARCHAR2(32), "PRESTTRIG" VARCHAR2(32), "DMLROWTRIG" VARCHAR2(32), "POSTSTTRIG" VARCHAR2(32), "DBCNFREGID" NUMBER, "DBCNFCBKPRC" VARCHAR2(75), CONSTRAINT "RLM$DMLEVTPKEY" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "TATAB_NAME") ENABLE, CONSTRAINT "RLM$DMLEVTREFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$DMLEVTTRIGS",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, TATAB_NAME],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=TATAB_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=PRESTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=DMLROWTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=POSTSTTRIG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=6,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFREGID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=7,optional=true,autoIncremented=false,generated=false}, Struct{name=DBCNFCBKPRC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=8,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,575 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,575 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@5ecd5499, value=[B@7aa4b8ed, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@56cbd1fe to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,575 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,575 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=14985, buffer=java.nio.HeapByteBuffer[pos=0 lim=14985 cap=14985]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,575 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,576 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247323,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EQUALSPEC,scn=5535835,ssn=0},ts_ms=1666014247404,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EQUALSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "OPCODE_ID" NUMBER, "EQLKEYTYP" VARCHAR2(100), "EQSPECFLG" NUMBER, "EQUALATTRS" "EXFSYS"."RLM$EQUALATTR" , CONSTRAINT "RLM$EQSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "OPCODE_ID") ENABLE, CONSTRAINT "RLM$EQUALSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EQUALSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, OPCODE_ID],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=OPCODE_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=EQLKEYTYP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EQSPECFLG,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=EQUALATTRS,jdbcType=2003,typeName=RLM$EQUALATTR,typeExpression=RLM$EQUALATTR,length=1073,position=6,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,576 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247323,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EQUALSPEC,scn=5535835,ssn=0},ts_ms=1666014247404,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EQUALSPEC" ("RSET_OWNER" VARCHAR2(32), "RSET_NAME" VARCHAR2(32), "OPCODE_ID" NUMBER, "EQLKEYTYP" VARCHAR2(100), "EQSPECFLG" NUMBER, "EQUALATTRS" "EXFSYS"."RLM$EQUALATTR" , CONSTRAINT "RLM$EQSPECPK" PRIMARY KEY ("RSET_OWNER", "RSET_NAME", "OPCODE_ID") ENABLE, CONSTRAINT "RLM$EQUALSPECFK" FOREIGN KEY ("RSET_OWNER", "RSET_NAME") REFERENCES "EXFSYS"."RLM$RULESET" ("RSET_OWNER", "RSET_NAME") ON DELETE CASCADE DEFERRABLE INITIALLY DEFERRED ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EQUALSPEC",table=Struct{primaryKeyColumnNames=[RSET_OWNER, RSET_NAME, OPCODE_ID],columns=[Struct{name=RSET_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=RSET_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=OPCODE_ID,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=false,autoIncremented=false,generated=false}, Struct{name=EQLKEYTYP,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=100,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EQSPECFLG,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=5,optional=true,autoIncremented=false,generated=false}, Struct{name=EQUALATTRS,jdbcType=2003,typeName=RLM$EQUALATTR,typeExpression=RLM$EQUALATTR,length=1073,position=6,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@10b26256, value=[B@4f8ec37d, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@47c2bf53 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,576 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247406,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$ERRCODE,scn=5535835,ssn=0},ts_ms=1666014247467,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$ERRCODE" ("ERRNO" NUMBER, "ERRMSG" VARCHAR2(200), "ERRDSC" VARCHAR2(500), "ERRACT" VARCHAR2(500), CONSTRAINT "RLM$ERRCODEPK" PRIMARY KEY ("ERRNO") ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$ERRCODE",table=Struct{primaryKeyColumnNames=[ERRNO],columns=[Struct{name=ERRNO,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ERRMSG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=200,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRDSC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRACT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=4,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,576 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247406,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$ERRCODE,scn=5535835,ssn=0},ts_ms=1666014247467,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$ERRCODE" ("ERRNO" NUMBER, "ERRMSG" VARCHAR2(200), "ERRDSC" VARCHAR2(500), "ERRACT" VARCHAR2(500), CONSTRAINT "RLM$ERRCODEPK" PRIMARY KEY ("ERRNO") ENABLE ) ORGANIZATION INDEX NOCOMPRESS OVERFLOW;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$ERRCODE",table=Struct{primaryKeyColumnNames=[ERRNO],columns=[Struct{name=ERRNO,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=ERRMSG,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=200,position=2,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRDSC,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=ERRACT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=500,position=4,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@1feba596, value=[B@1c48d4ef, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@25646930 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Allocating a new 16384 byte message buffer for topic oracle partition 0 with remaining timeout 9223372036854775807ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,576 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3684): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11040, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,576 TRACE || Applying transformation io.debezium.transforms.ByLogicalTableRouter to SourceRecord{sourcePartition={server=oracle}, sourceOffset={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} ConnectRecord{topic='oracle', kafkaPartition=0, key=Struct{databaseName=MIS}, keySchema=Schema{io.debezium.connector.oracle.SchemaChangeKey:STRUCT}, value=Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247469,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EVENTSTRUCT,scn=5535835,ssn=0},ts_ms=1666014247530,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EVENTSTRUCT" ("EVST_OWNER" VARCHAR2(32), "EVST_NAME" VARCHAR2(32), "EVST_PROP" NUMBER, "EVST_PRCT" VARCHAR2(32), "EVST_PRCTTLS" VARCHAR2(75), CONSTRAINT "RLM$EVST_PRIMKEY" PRIMARY KEY ("EVST_OWNER", "EVST_NAME") ENABLE, CONSTRAINT "RLM$EVST_REFKEY" FOREIGN KEY ("EVST_OWNER", "EVST_NAME") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EVENTSTRUCT",table=Struct{primaryKeyColumnNames=[EVST_OWNER, EVST_NAME],columns=[Struct{name=EVST_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_PROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCTTLS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=5,optional=true,autoIncremented=false,generated=false}]}}]}, valueSchema=Schema{io.debezium.connector.oracle.SchemaChangeValue:STRUCT}, timestamp=null, headers=ConnectHeaders(headers=)} [org.apache.kafka.connect.runtime.TransformationChain] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3684 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,576 TRACE || Successfully produced messages to oracle-0 with base offset 11040. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,576 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11040 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Appending record to the topic oracle with key Struct{databaseName=MIS}, value Struct{source=Struct{version=2.0.0.Final,connector=oracle,name=oracle,ts_ms=1666014247469,snapshot=true,db=MIS,schema=EXFSYS,table=RLM$EVENTSTRUCT,scn=5535835,ssn=0},ts_ms=1666014247530,databaseName=MIS,schemaName=EXFSYS,ddl= CREATE TABLE "EXFSYS"."RLM$EVENTSTRUCT" ("EVST_OWNER" VARCHAR2(32), "EVST_NAME" VARCHAR2(32), "EVST_PROP" NUMBER, "EVST_PRCT" VARCHAR2(32), "EVST_PRCTTLS" VARCHAR2(75), CONSTRAINT "RLM$EVST_PRIMKEY" PRIMARY KEY ("EVST_OWNER", "EVST_NAME") ENABLE, CONSTRAINT "RLM$EVST_REFKEY" FOREIGN KEY ("EVST_OWNER", "EVST_NAME") REFERENCES "EXFSYS"."EXF$ATTRSET" ("ATSOWNER", "ATSNAME") ON DELETE CASCADE ENABLE ) ORGANIZATION INDEX NOCOMPRESS ;,tableChanges=[Struct{type=CREATE,id="MIS"."EXFSYS"."RLM$EVENTSTRUCT",table=Struct{primaryKeyColumnNames=[EVST_OWNER, EVST_NAME],columns=[Struct{name=EVST_OWNER,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=1,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_NAME,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=2,optional=false,autoIncremented=false,generated=false}, Struct{name=EVST_PROP,jdbcType=2,typeName=NUMBER,typeExpression=NUMBER,length=0,position=3,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCT,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=32,position=4,optional=true,autoIncremented=false,generated=false}, Struct{name=EVST_PRCTTLS,jdbcType=12,typeName=VARCHAR2,typeExpression=VARCHAR2,length=75,position=5,optional=true,autoIncremented=false,generated=false}]}}]} [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || Topic creation by the connector is disabled or the topic oracle was previously created.If auto.create.topics.enable is enabled on the broker, the topic will be created with default settings [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,576 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Attempting to append record ProducerRecord(topic=oracle, partition=0, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@350bbe9a, value=[B@6fbefb78, timestamp=null) with callback org.apache.kafka.connect.runtime.AbstractWorkerSourceTask$$Lambda$1184/0x00000008408af440@69153256 to topic oracle partition 0 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,577 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Waking up the sender since topic oracle partition 0 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,576 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11041 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,577 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11042 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,577 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Nothing to send to Kafka. Polling source for additional records [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,577 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,577 DEBUG Oracle|oracle|oracle-connector-task polling records... [io.debezium.connector.base.ChangeEventQueue] 2022-10-17 13:44:07,577 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,577 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3685) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=13026]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,577 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=13026, buffer=java.nio.HeapByteBuffer[pos=0 lim=13026 cap=13026]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,577 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,578 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3685): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11043, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3685 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,578 TRACE || Successfully produced messages to oracle-0 with base offset 11043. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,578 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11043 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,578 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11044 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,578 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3686) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[oracle-0=11097]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=0, records=MemoryRecords(size=11097, buffer=java.nio.HeapByteBuffer[pos=0 lim=11097 cap=11097]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,578 DEBUG || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=connector-producer-mis-oracle-refbook-0, correlationId=3686): ProduceResponseData(responses=[TopicProduceResponse(name='oracle', partitionResponses=[PartitionProduceResponse(index=0, errorCode=0, baseOffset=11045, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,578 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] Received produce response from node 1 with correlation id 3686 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,578 TRACE || Successfully produced messages to oracle-0 with base offset 11045. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,579 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11045 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,579 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Wrote record successfully: topic oracle partition 0 offset 11046 [org.apache.kafka.connect.runtime.AbstractWorkerSourceTask] 2022-10-17 13:44:07,579 TRACE || [Producer clientId=connector-producer-mis-oracle-refbook-0] The number of partitions is too small: available=1, all=1, not using adaptive for topic oracle [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,579 DEBUG || WorkerSourceTask{id=mis-oracle-refbook-0} Committing offsets [org.apache.kafka.connect.runtime.WorkerSourceTask] 2022-10-17 13:44:07,579 DEBUG || WorkerSourceTask{id=mis-oracle-refbook-0} Either no records were produced by the task since the last offset commit, or every record has been filtered out by a transformation or dropped due to transformation or conversion errors. [org.apache.kafka.connect.runtime.WorkerSourceTask] 2022-10-17 13:44:07,579 DEBUG || Submitting 1 entries to backing store. The offsets are: {{server=oracle}={snapshot_scn=5535835, snapshot=true, scn=5535835, snapshot_completed=false}} [org.apache.kafka.connect.storage.OffsetStorageWriter] 2022-10-17 13:44:07,579 TRACE || [Producer clientId=producer-1] Attempting to append record ProducerRecord(topic=dbz_offsets, partition=null, headers=RecordHeaders(headers = [], isReadOnly = false), key=[B@6d8894be, value=[B@51f18fb, timestamp=null) with callback org.apache.kafka.connect.storage.KafkaOffsetBackingStore$SetCallbackFuture@1de2561b to topic dbz_offsets partition 9 [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,579 TRACE || [Producer clientId=producer-1] Allocating a new 16384 byte message buffer for topic dbz_offsets partition 9 with remaining timeout 60000ms [org.apache.kafka.clients.producer.internals.RecordAccumulator] 2022-10-17 13:44:07,579 TRACE || [Producer clientId=producer-1] Waking up the sender since topic dbz_offsets partition 9 is either full or getting a new batch [org.apache.kafka.clients.producer.KafkaProducer] 2022-10-17 13:44:07,579 TRACE || [Producer clientId=producer-1] No load stats for topic dbz_offsets, not using adaptive [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,580 TRACE || [Producer clientId=producer-1] Nodes with data ready to send: [kafka:29092 (id: 1 rack: null)] [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,580 DEBUG || [Producer clientId=producer-1] Sending PRODUCE request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=producer-1, correlationId=11) and timeout 30000 to node 1: {acks=-1,timeout=30000,partitionSizes=[dbz_offsets-9=197]} [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,580 TRACE || [Producer clientId=producer-1] Sent produce request to 1: (type=ProduceRequest, acks=-1, timeout=30000, partitionRecords=([PartitionProduceData(index=9, records=MemoryRecords(size=197, buffer=java.nio.HeapByteBuffer[pos=0 lim=197 cap=197]))]), transactionalId='' [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,580 TRACE || [Producer clientId=producer-1] No load stats for topic dbz_offsets, not using adaptive [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,580 DEBUG || [Producer clientId=producer-1] Received PRODUCE response from node 1 for request with header RequestHeader(apiKey=PRODUCE, apiVersion=9, clientId=producer-1, correlationId=11): ProduceResponseData(responses=[TopicProduceResponse(name='dbz_offsets', partitionResponses=[PartitionProduceResponse(index=9, errorCode=0, baseOffset=12, logAppendTimeMs=-1, logStartOffset=0, recordErrors=[], errorMessage=null)])], throttleTimeMs=0) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,580 TRACE || [Producer clientId=producer-1] Received produce response from node 1 with correlation id 11 [org.apache.kafka.clients.producer.internals.Sender] 2022-10-17 13:44:07,580 TRACE || Successfully produced messages to dbz_offsets-9 with base offset 12. [org.apache.kafka.clients.producer.internals.ProducerBatch] 2022-10-17 13:44:07,580 TRACE || WorkerSourceTask{id=mis-oracle-refbook-0} Finished flushing offsets to storage [org.apache.kafka.connect.runtime.WorkerSourceTask] 2022-10-17 13:44:07,580 TRACE || [Producer clientId=producer-1] No load stats for topic dbz_offsets, not using adaptive [org.apache.kafka.clients.producer.internals.BuiltInPartitioner] 2022-10-17 13:44:07,580 DEBUG || WorkerSourceTask{id=mis-oracle-refbook-0} Finished commitOffsets successfully in 1 ms [org.apache.kafka.connect.runtime.WorkerSourceTask] 2022-10-17 13:44:07,580 ERROR || WorkerSourceTask{id=mis-oracle-refbook-0} Task threw an uncaught and unrecoverable exception. Task is being killed and will not recover until manually restarted [org.apache.kafka.connect.runtime.WorkerTask] org.apache.kafka.connect.errors.ConnectException: An exception occurred in the change event producer. This connector will be stopped. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:53) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:116) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: io.debezium.DebeziumException: java.lang.NullPointerException at io.debezium.pipeline.source.AbstractSnapshotChangeEventSource.execute(AbstractSnapshotChangeEventSource.java:85) at io.debezium.pipeline.ChangeEventSourceCoordinator.doSnapshot(ChangeEventSourceCoordinator.java:155) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:137) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:109) ... 5 more Caused by: java.lang.NullPointerException at io.debezium.connector.oracle.OracleSnapshotChangeEventSource.createSchemaChangeEventsForTables(OracleSnapshotChangeEventSource.java:215) at io.debezium.relational.RelationalSnapshotChangeEventSource.doExecute(RelationalSnapshotChangeEventSource.java:122) at io.debezium.pipeline.source.AbstractSnapshotChangeEventSource.execute(AbstractSnapshotChangeEventSource.java:76) ... 8 more 2022-10-17 13:44:07,581 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Received FETCH response from node 1 for request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=948): FetchResponseData(throttleTimeMs=0, errorCode=0, sessionId=1518104247, responses=[FetchableTopicResponse(topic='', topicId=86YDyE3vTMuaOLiRRRs2oQ, partitions=[PartitionData(partitionIndex=9, errorCode=0, highWatermark=13, lastStableOffset=13, logStartOffset=0, divergingEpoch=EpochEndOffset(epoch=-1, endOffset=-1), currentLeader=LeaderIdAndEpoch(leaderId=-1, leaderEpoch=-1), snapshotId=SnapshotId(endOffset=-1, epoch=-1), abortedTransactions=null, preferredReadReplica=-1, records=MemoryRecords(size=197, buffer=java.nio.HeapByteBuffer[pos=0 lim=197 cap=200]))])]) [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,581 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Node 1 sent an incremental fetch response with throttleTimeMs = 0 for session 1518104247 with response=(dbz_offsets-9), implied=(dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,581 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Fetch READ_UNCOMMITTED at offset 12 for partition dbz_offsets-9 returned fetch data PartitionData(partitionIndex=9, errorCode=0, highWatermark=13, lastStableOffset=13, logStartOffset=0, divergingEpoch=EpochEndOffset(epoch=-1, endOffset=-1), currentLeader=LeaderIdAndEpoch(leaderId=-1, leaderEpoch=-1), snapshotId=SnapshotId(endOffset=-1, epoch=-1), abortedTransactions=null, preferredReadReplica=-1, records=MemoryRecords(size=197, buffer=java.nio.HeapByteBuffer[pos=0 lim=197 cap=200])) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,581 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Preparing to read 197 bytes of data for partition dbz_offsets-9 with offset FetchPosition{offset=12, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,581 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Updating high watermark for partition dbz_offsets-9 to 13 [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,581 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Updating log start offset for partition dbz_offsets-9 to 0 [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,581 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Updating last stable offset for partition dbz_offsets-9 to 13 [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,581 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Returning 1 fetched records at offset FetchPosition{offset=12, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} for assigned partition dbz_offsets-9 [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Updating fetch position from FetchPosition{offset=12, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to FetchPosition{offset=13, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} for partition dbz_offsets-9 and returning 1 records from `poll()` [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 INFO || Stopping down connector [io.debezium.connector.common.BaseSourceTask] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-7 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-13 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-11 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-1 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-5 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-3 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-23 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-17 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-15 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-21 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-19 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-10 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-8 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-14 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-12 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-2 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-0 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-6 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-4 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-24 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-18 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-16 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-22 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-20 at position FetchPosition{offset=0, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Added READ_UNCOMMITTED fetch request for partition dbz_offsets-9 at position FetchPosition{offset=13, offsetEpoch=Optional[0], currentLeader=LeaderAndEpoch{leader=Optional[kafka:29092 (id: 1 rack: null)], epoch=0}} to node kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Built incremental fetch (sessionId=1518104247, epoch=942) for node 1. Added (), altered (86YDyE3vTMuaOLiRRRs2oQ:dbz_offsets-9), removed (), replaced () out of (dbz_offsets-9, dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20) [org.apache.kafka.clients.FetchSessionHandler] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending READ_UNCOMMITTED IncrementalFetchRequest(toSend=(dbz_offsets-9), toForget=(), toReplace=(), implied=(dbz_offsets-7, dbz_offsets-13, dbz_offsets-11, dbz_offsets-1, dbz_offsets-5, dbz_offsets-3, dbz_offsets-23, dbz_offsets-17, dbz_offsets-15, dbz_offsets-21, dbz_offsets-19, dbz_offsets-10, dbz_offsets-8, dbz_offsets-14, dbz_offsets-12, dbz_offsets-2, dbz_offsets-0, dbz_offsets-6, dbz_offsets-4, dbz_offsets-24, dbz_offsets-18, dbz_offsets-16, dbz_offsets-22, dbz_offsets-20), canUseTopicIds=True) to broker kafka:29092 (id: 1 rack: null) [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,582 DEBUG || [Consumer clientId=consumer-dbz-1, groupId=dbz] Sending FETCH request with header RequestHeader(apiKey=FETCH, apiVersion=13, clientId=consumer-dbz-1, correlationId=949) and timeout 30000 to node 1: FetchRequestData(clusterId=null, replicaId=-1, maxWaitMs=500, minBytes=1, maxBytes=52428800, isolationLevel=0, sessionId=1518104247, sessionEpoch=942, topics=[FetchTopic(topic='dbz_offsets', topicId=86YDyE3vTMuaOLiRRRs2oQ, partitions=[FetchPartition(partition=9, currentLeaderEpoch=0, fetchOffset=13, lastFetchedEpoch=-1, logStartOffset=-1, partitionMaxBytes=1048576)])], forgottenTopicsData=[], rackId='') [org.apache.kafka.clients.NetworkClient] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-7 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-13 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-11 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-1 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-5 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-3 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-23 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-17 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-15 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-21 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-19 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-10 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-8 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-14 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-12 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-2 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-0 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-6 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-4 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-24 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-18 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-16 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-22 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-20 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Skipping fetch for partition dbz_offsets-9 because previous request to kafka:29092 (id: 1 rack: null) has not been processed [org.apache.kafka.clients.consumer.internals.Fetcher] 2022-10-17 13:44:07,583 TRACE || [Consumer clientId=consumer-dbz-1, groupId=dbz] Polling for fetches with timeout 2147483647 [org.apache.kafka.clients.consumer.KafkaConsumer]