-
Bug
-
Resolution: Done
-
Major
-
2.3.0.Final
-
None
Bug report
When the LOG_MINING_TRANSACTION_SNAPSHOT_BOUNDARY_MODE is configured as non-skip, and there is a transaction in V$TRANSACTION, the corresponding start_scn is 0, log mining from the oldest scn occurs when the oracle connector is started for the first time, and a fairly large span of SCN number range queries appears. Data that results in no log mining being distributed for a long time
What Debezium connector do you use and what version?
debezium-connector-oracle-2.3.0.Final
What is the connector configuration?
connector.class = io.debezium.connector.oracle.OracleConnector
snapshot.locking.mode = none
log.mining.buffer.drop.on.stop = false
max.queue.size = 81920
schema.include.list = TEST_CDC
internal.log.mining.read.only = false
topic.heartbeat.prefix = __debezium_heartbeat
log.mining.strategy = online_catalog
include.schema.changes = true
schema.history.internal.store.only.captured.tables.ddl = true
schema.history.internal.file.filename = /data/debezium/cdc/td1/history/schema.dat
tombstones.on.delete = false
unavailable.value.placeholder = __debezium_value
topic.prefix = td1
offset.storage.file.filename = /data/debezium/cdc/td1/offsets.dat
poll.interval.ms = 100
lob.enabled = true
errors.retry.delay.initial.ms = 300
log.mining.archive.log.only.mode = false
value.converter = org.apache.kafka.connect.json.JsonConverter
key.converter = org.apache.kafka.connect.json.JsonConverter
database.user = C##TEST
database.dbname = ORCL
custom.retriable.exception = .(ORA-00600|ORA-01289|ORA-01291|ORA-31603|ORA-26824|ORA-26876|Invalid value: null used for required field: "typeName")(.|\s)
offset.storage = org.apache.kafka.connect.storage.FileOffsetBackingStore
database.pdb.name = PDB1
database.connection.adapter = logminer
log.mining.buffer.type = memory
internal.log.mining.transaction.snapshot.boundary.mode = all
offset.flush.timeout.ms = 5000
errors.retry.delay.max.ms = 10000
event.processing.failure.handling.mode = fail
schema.history.internal.skip.unparseable.ddl = true
log.mining.restart.connection = true
database.port = 1521
offset.flush.interval.ms = 5000
schema.history.internal = io.debezium.storage.file.history.FileSchemaHistory
log.mining.session.max.ms = 1800000
errors.max.retries = -1
database.hostname = 10.10.93.27
database.password = ********
name = td1
log.mining.batch.size.default = 100000
max.batch.size = 20480
skipped.operations = none
table.include.list = TEST_CDC.AB01
What is the captured database version and mode of depoyment?
ORACLE 19C
What behaviour do you expect?
Log mining starts from the database current sc when the connector is first started, or from the smallest start_scn in the active transaction at first startup
What behaviour do you see?
When an illegal value occurs in the start_scn in the obtained active transaction, log mining starts from the oldest scn
Do you see the same behaviour using the latest relesead Debezium version?
(Ideally, also verify with latest Alpha/Beta/CR version)
<Your answer>
Do you have the connector logs, ideally from start till finish?
(You might be asked later to provide DEBUG/TRACE level log)
<Your answer>
How to reproduce the issue using our tutorial deployment?
<Your answer>
Feature request or enhancement
For feature requests or enhancements, provide this information, please:
Which use case/requirement will be addressed by the proposed feature?
<Your answer>
Implementation ideas (optional)
<Your answer>
- links to
-
RHEA-2024:129636 Red Hat build of Debezium 2.5.4 release