Uploaded image for project: 'Debezium'
  1. Debezium
  2. DBZ-9550

Support for `errors.tolerance` in Debezium CDC Sink

XMLWordPrintable

    • Icon: Feature Request Feature Request
    • Resolution: Unresolved
    • Icon: Major Major
    • Backlog
    • 3.1.3.Final
    • debezium-core
    • None
    • False
    • Hide

      None

      Show
      None
    • False
    • Important

      In order to make your issue reports as actionable as possible, please provide the following information, depending on the issue type.

      Bug report

      For bug reports, provide this information, please:

      What Debezium connector do you use and what version?

      3.1.3.Final

      What is the connector configuration?

       

      {
        "config": {
          "collection.name.format": "${source.table}",
          "connection.password": "${secret:***}",
          "connection.url": "jdbc:postgresql://***:5432/***",
          "connection.user": "${secret:***}",
          "connection.username": "${secret:***}",
          "connector.class": "io.debezium.connector.jdbc.JdbcSinkConnector",
          "connector.name": "postgrescdcsink-os-project-snapshots-cdc-sink",
          "consumer.override.bootstrap.servers": "SASL_SSL://***:9092",
          "consumer.override.group.id": "connect-debezium-sink-osps-deps-decomp",
          "consumer.override.sasl.jaas.config": "org.apache.kafka.common.security.plain.PlainLoginModule required username='${secret:***}' password='${secret:***}';",
          "consumer.override.sasl.mechanism": "PLAIN",
          "consumer.override.security.protocol": "SASL_SSL",
          "delete.enabled": "true",
          "dialect.name": "PostgreSqlDatabaseDialect",
          "errors.log.enable": "true",
          "errors.tolerance": "all",
          "hibernate.c3p0.idle_test_period": "300",
          "hibernate.c3p0.testConnectionOnCheckin": "true",
          "insert.mode": "upsert",
          "key.converter": "io.confluent.connect.avro.AvroConverter",
          "key.converter.basic.auth.credentials.source": "USER_INFO",
          "key.converter.basic.auth.user.info": "${secret:***}",
          "key.converter.schema.registry.url": "https://***.eu-central-1.aws.confluent.cloud",
          "name": "postgrescdcsink-os-project-snapshots-cdc-sink",
          "primary.key.mode": "record_key",
          "quote.identifiers": "true",
          "schema.evolution": "basic",
          "tasks.max": "3",
          "topics": "registry_monitor_deps_decomp.public.monitor_dependencies,registry_monitor_deps_decomp.public.pkg_info,registry_monitor_deps_decomp.public.pkg_version_info",
          "transforms": "Load",
          "transforms.Load.connector.name": "postgrescdcsink-os-project-snapshots-cdc-sink",
          "transforms.Load.included.columns": "public.monitor_dependencies.licenses,public.monitor_dependencies.dependencies_with_issues,public.pkg_info.deprecated_versions",
          "transforms.Load.service.bucket.name": "event-bus-connect-offload-bucket-082765316871",
          "transforms.Load.service.class": "io.snyk.kafka.connect.transform.big_column_offloader.service.S3Service",
          "transforms.Load.type": "io.snyk.kafka.connect.transform.big_column_offloader.LoadBigColumns",
          "use.time.zone": "UTC",
          "value.converter": "io.confluent.connect.avro.AvroConverter",
          "value.converter.basic.auth.credentials.source": "USER_INFO",
          "value.converter.basic.auth.user.info": "${secret:***}",
          "value.converter.schema.registry.url": "https://***.eu-central-1.aws.confluent.cloud",
          "value.converter.value.subject.name.strategy": "io.confluent.kafka.serializers.subject.TopicNameStrategy"
        },

       

      What is the captured database version and mode of deployment?

      PostgreSQL on AWS RDS, version 14.18

      What behavior do you expect?

      In case an INSERT/UPDATE violates a constraint of the DB, record should be discarded.

      What behavior do you see?

      Instead of discarding the records, the connector returns the error:

       

      ERROR: duplicate key value violates unique constraint "pkg_info_pkg_format_type_name_uniq_idx"  Detail: Key (pkg_format_type, name)=(***REDACTED***) already exists.

      Below the full stacktrace:

       

       

      postgrescdcsink-os-project-snapshots-cdc-sink.task[0]: org.apache.kafka.connect.errors.ConnectException: Exiting WorkerSinkTask due to unrecoverable exception.at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:636)at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:345)at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:247)at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:216)at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:226)at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:281)at org.apache.kafka.connect.runtime.isolation.Plugins.lambda$withClassLoader$1(Plugins.java:238)at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539)at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)at java.base/java.lang.Thread.run(Thread.java:840)Caused by: org.apache.kafka.connect.errors.ConnectException: JDBC sink connector failureat io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:121)at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:606)... 11 moreCaused by: org.apache.kafka.connect.errors.ConnectException: Failed to process a sink recordat io.debezium.connector.jdbc.JdbcChangeEventSink.flushBufferWithRetries(JdbcChangeEventSink.java:243)at io.debezium.connector.jdbc.JdbcChangeEventSink.flushBufferRecordsWithRetries(JdbcChangeEventSink.java:210)at io.debezium.connector.jdbc.JdbcChangeEventSink.execute(JdbcChangeEventSink.java:135)at io.debezium.connector.jdbc.JdbcSinkConnectorTask.put(JdbcSinkConnectorTask.java:128)... 12 moreCaused by: org.hibernate.exception.ConstraintViolationException: error executing work [Batch entry 22 INSERT INTO "public"."pkg_info" ("id","pkg_format_type","name","latest_version","first_published_date","latest_version_published_date","deprecated","deprecated_versions","created","modified") VALUES (***REDACTED***) ON CONFLICT ("id") DO UPDATE SET "pkg_format_type"=EXCLUDED."pkg_format_type","name"=EXCLUDED."name","latest_version"=EXCLUDED."latest_version","first_published_date"=EXCLUDED."first_published_date","latest_version_published_date"=EXCLUDED."latest_version_published_date","deprecated"=EXCLUDED."deprecated","deprecated_versions"=EXCLUDED."deprecated_versions","created"=EXCLUDED."created","modified"=EXCLUDED."modified" was aborted: ERROR: duplicate key value violates unique constraint "pkg_info_pkg_format_type_name_uniq_idx"  Detail: Key (pkg_format_type, name)=(***REDACTED***) already exists.  Call getNextException to see other errors in the batch.] [n/a]at org.hibernate.exception.internal.SQLStateConversionDelegate.convert(SQLStateConversionDelegate.java:97)at org.hibernate.exception.internal.StandardSQLExceptionConverter.convert(StandardSQLExceptionConverter.java:58)at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:108)at org.hibernate.engine.jdbc.spi.SqlExceptionHelper.convert(SqlExceptionHelper.java:94)at org.hibernate.engine.jdbc.internal.JdbcCoordinatorImpl.coordinateWork(JdbcCoordinatorImpl.java:308)at org.hibernate.internal.AbstractSharedSessionContract.doWork(AbstractSharedSessionContract.java:1053)at org.hibernate.internal.AbstractSharedSessionContract.doWork(AbstractSharedSessionContract.java:1041)at io.debezium.connector.jdbc.RecordWriter.write(RecordWriter.java:50)at io.debezium.connector.jdbc.JdbcChangeEventSink.flushBuffer(JdbcChangeEventSink.java:259)at io.debezium.connector.jdbc.JdbcChangeEventSink.flushBufferWithRetries(JdbcChangeEventSink.java:234)... 15 moreCaused by: java.sql.BatchUpdateException: Batch entry 22 INSERT INTO "public"."pkg_info" ("id","pkg_format_type","name","latest_version","first_published_date","latest_version_published_date","deprecated","deprecated_versions","created","modified") VALUES (('12379812'::int8),('npm'),('@oneplatformdev/hooks'),('0.1.0-91'),('2025-01-27 08:03:25.105+00'::timestamp with time zone),('2025-10-13 14:18:59.329+00'::timestamp with time zone),('FALSE'::boolean),cast(('[]') as json),('2025-10-13 14:30:30.156+00'::timestamp with time zone),('2025-10-13 14:30:30.156+00'::timestamp with time zone)) ON CONFLICT ("id") DO UPDATE SET "pkg_format_type"=EXCLUDED."pkg_format_type","name"=EXCLUDED."name","latest_version"=EXCLUDED."latest_version","first_published_date"=EXCLUDED."first_published_date","latest_version_published_date"=EXCLUDED."latest_version_published_date","deprecated"=EXCLUDED."deprecated","deprecated_versions"=EXCLUDED."deprecated_versions","created"=EXCLUDED."created","modified"=EXCLUDED."modified" was aborted: ERROR: duplicate key value violates unique constraint "pkg_info_pkg_format_type_name_uniq_idx"  Detail: Key (pkg_format_type, name)=(npm, @oneplatformdev/hooks) already exists.  Call getNextException to see other errors in the batch.at org.postgresql.jdbc.BatchResultHandler.handleError(BatchResultHandler.java:165)at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2422)at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2154)at org.postgresql.core.v3.QueryExecutorImpl.flushIfDeadlockRisk(QueryExecutorImpl.java:1502)at org.postgresql.core.v3.QueryExecutorImpl.sendQuery(QueryExecutorImpl.java:1527)at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:566)at org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:886)at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:910)at org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1778)at com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeBatch(NewProxyPreparedStatement.java:2544)at io.debezium.connector.jdbc.RecordWriter.lambda$processBatch$0(RecordWriter.java:88)at org.hibernate.jdbc.WorkExecutor.executeWork(WorkExecutor.java:39)at org.hibernate.internal.AbstractSharedSessionContract.lambda$doWork$4(AbstractSharedSessionContract.java:1042)at org.hibernate.engine.jdbc.internal.JdbcCoordinatorImpl.coordinateWork(JdbcCoordinatorImpl.java:303)... 20 moreCaused by: org.postgresql.util.PSQLException: ERROR: duplicate key value violates unique constraint "pkg_info_pkg_format_type_name_uniq_idx"  Detail: Key (pkg_format_type, name)=(npm, @oneplatformdev/hooks) already exists.at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2734)at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2421)... 32 more

       

       

      Do you see the same behaviour using the latest released Debezium version?

      Currently can't update for reason of compatibility with other connector in the KC cluster.

      Do you have the connector logs, ideally from start till finish?

      (You might be asked later to provide DEBUG/TRACE level log)

      Provided above a stacktrace

      How to reproduce the issue using our tutorial deployment?

      I don't think I can provide this.

      Feature request or enhancement

      For feature requests or enhancements, provide this information, please:

      Which use case/requirement will be addressed by the proposed feature?

      <Your answer>

      Implementation ideas (optional)

      <Your answer>

              Unassigned Unassigned
              detronizator@gmail.com Ivan De Marino (Inactive)
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

                Created:
                Updated: