Uploaded image for project: 'Debezium'
  1. Debezium
  2. DBZ-7774

ActivateTracingSpan SMT propagates db-log-write span instead of debezium-read span

XMLWordPrintable

    • False
    • None
    • False

      In order to make your issue reports as actionable as possible, please provide the following information, depending on the issue type.

      Bug repDeort

      For bug reports, provide this information, please:

      What Debezium connector do you use and what version?

      2.6.0.Final, PostgresConnector

      What is the connector configuration?

       

      {
        "name": "orders-cloud-v1-connector",
        "config": {
          "connector.class": "io.debezium.connector.postgresql.PostgresConnector",
          "plugin.name": "pgoutput",
          "snapshot.mode": "initial",
          "topic.prefix": "orders-cloud-v1",
          "topic.heartbeat.prefix": "sys_debezium-heartbeat",
      
          "table.include.list": "public.orders",
          "message.key.columns": "public.orders:id",
          "slot.name": "orders_slot",
          "publication.name": "orders_publication",
          "publication.autocreate.mode": "disabled",
      
          "database.dbname": "productsdb",
          "database.hostname": "localhost",
          "database.port": "5432",
          "database.user": "postgres",
          "database.password": "postgres",
      
          "key.converter" : "org.apache.kafka.connect.storage.StringConverter",
          "key.converter.schemas.enable": false,
          "value.converter": "org.apache.kafka.connect.json.JsonConverter",
          "value.converter.schemas.enable": false,
      
          "decimal.handling.mode": "precise",
          "tombstones.on.delete": false,
          "null.handling.mode": "keep",
          "heartbeat.interval.ms": "15000",
      
          "transforms": "filterSignalEvents,tracing,unwrap,ReplaceField,moveHeadersToValue,dropPrefix",
          "transforms.addMetadataHeaders.type": "org.apache.kafka.connect.transforms.HeaderFrom$Value",
          "transforms.addMetadataHeaders.fields": "source,op,transaction",
          "transforms.addMetadataHeaders.headers": "source,op,transaction",
          "transforms.addMetadataHeaders.operation": "copy",
          "transforms.addMetadataHeaders.predicate": "isHeartbeat",
          "transforms.addMetadataHeaders.negate": true,
          "transforms.unwrap.type": "io.debezium.transforms.ExtractNewRecordState",
          "transforms.unwrap.add.fields": "op,lsn",
          "transforms.unwrap.delete.handling.mode": "rewrite",
          "transforms.unwrap.drop.tombstones": "true",
          "transforms.setSchemaMetadata.type": "org.apache.kafka.connect.transforms.SetSchemaMetadata$Value",
          "transforms.setSchemaMetadata.schema.name": "music.Order",
          "transforms.setSchemaMetadata.predicate": "isHeartbeat",
          "transforms.setSchemaMetadata.negate": true,
          "transforms.dropPrefix.type": "org.apache.kafka.connect.transforms.RegexRouter",
          "transforms.dropPrefix.regex": "orders-cloud-v1\\.public\\.(.*)",
          "transforms.dropPrefix.replacement": "music.subscriptions.order.events",
          "transforms.dropPrefix.predicate": "isHeartbeat",
          "transforms.dropPrefix.negate": true,
          "transforms.filterSignalEvents.type": "org.apache.kafka.connect.transforms.Filter",
          "transforms.filterSignalEvents.predicate": "isDebeziumSignal",
          "transforms.tracing.type": "io.debezium.transforms.tracing.ActivateTracingSpan",
          "transforms.tracing.tracing.with.context.field.only": "true",
          "transforms.ReplaceField.type": "org.apache.kafka.connect.transforms.ReplaceField$Value",
          "transforms.ReplaceField.exclude": "tracingspancontext",
          "transforms.moveHeadersToValue.type": "io.debezium.transforms.HeaderToValue",
          "transforms.moveHeadersToValue.operation": "copy",
          "transforms.moveHeadersToValue.headers": "traceparent",
          "transforms.moveHeadersToValue.fields": "traceparent",
      
          "predicates": "isHeartbeat,isDebeziumSignal",
          "predicates.isHeartbeat.type": "org.apache.kafka.connect.transforms.predicates.TopicNameMatches",
          "predicates.isHeartbeat.pattern": "sys_debezium-heartbeat.*",
          "predicates.isDebeziumSignal.type": "org.apache.kafka.connect.transforms.predicates.TopicNameMatches",
          "predicates.isDebeziumSignal.pattern": "orders-cloud-v1\\.public\\.debezium_signal",
      
          "signal.enabled.channels": "source",
          "signal.data.collection": "public.debezium_signal"
        }
      } 

       

       

       

      What is the captured database version and mode of depoyment?

      (E.g. on-premises, with a specific cloud provider, etc.)

      Postgres 15

      What behaviour do you expect?

      Debezium ActivateTracingSpan smt inserts header traceparent with debezium-read operation's span as it's written in documentation

      What behaviour do you see?

      Debezium ActivateTracingSpan smt inserts header traceparent with db-log-write operation's span.

      As we can see debezium-read span is 1e611f3956e57496

      In Offset Explorer however there is span 7ae54a0fcd89eaa4:

      Actually I see the same behaviour with EventRouter SMT. If i add Debezium Interceptor it will add onSend span as child of db-log-write span and will send traceparent with db-log-write span to kafka. 

      Do you see the same behaviour using the latest relesead Debezium version?

      (Ideally, also verify with latest Alpha/Beta/CR version)

      <Your answer>

      Do you have the connector logs, ideally from start till finish?

      (You might be asked later to provide DEBUG/TRACE level log)

      How to reproduce the issue using our tutorial deployment?

      Create table orders:

       

      create table orders
      (
          id                 uuid not null
              constraint order_pk
                  primary key,
          client_id          uuid not null,
          tracingspancontext varchar(255)
      ); 
      -- create publication and replication slot
      select pg_create_logical_replication_slot('orders_slot', 'pgoutput');
      create publication orders_publication for table orders;

      start kafka-connect with open telemetry jars on classpath

       

      register connector with given config via kafka-connect rest api

      Feature request or enhancement

      For feature requests or enhancements, provide this information, please:

      Which use case/requirement will be addressed by the proposed feature?

      OpenTelemetry Integration

      Implementation ideas (optional)

      https://github.com/debezium/debezium/blob/2.6/debezium-core/src/main/java/io/debezium/transforms/tracing/TracingSpanUtil.java#L73 

      Probably this issue appears because debezium-read span is started and closed via finally block within db-log-write operation. Context is put to record headers after debezium-read closure so  db-log-write span is written to traceparent. Perhaps context should be added to headers before debezium-read closure

            Unassigned Unassigned
            tantoglow Georgii Bocharov
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

              Created:
              Updated: