[2025-02-17 14:21:31,312] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Stopping task source_os_inter_scenario_avro_mongodb_for_bigdata-0 (org.apache.kafka.connect.runtime.Worker:960) [2025-02-17 14:21:31,312] INFO [Worker clientId=connect-1, groupId=connect-avro-8823] Starting task source_os_inter_scenario_avro_mongodb_for_bigdata-0 (org.apache.kafka.connect.runtime.distributed.DistributedHerder:1428) "[2025-02-17 14:21:31,315] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.connect.avro.AvroConverterConfig:376)" "[2025-02-17 14:21:31,314] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] ConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mongodb.MongoDbConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = null name = source_os_inter_scenario_avro_mongodb_for_bigdata predicates = [] tasks.max = 1 transforms = [] value.converter = null (org.apache.kafka.connect.runtime.ConnectorConfig:376)" "[2025-02-17 14:21:31,314] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] TaskConfig values: task.class = class io.debezium.connector.mongodb.MongoDbConnectorTask (org.apache.kafka.connect.runtime.TaskConfig:376)" [2025-02-17 14:21:31,315] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Instantiated task source_os_inter_scenario_avro_mongodb_for_bigdata-0 with version 3.0.2.Final of type io.debezium.connector.mongodb.MongoDbConnectorTask (org.apache.kafka.connect.runtime.Worker:550) [2025-02-17 14:21:31,313] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating task source_os_inter_scenario_avro_mongodb_for_bigdata-0 (org.apache.kafka.connect.runtime.Worker:534) "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] AvroDataConfig values: connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false schemas.cache.config = 1000 scrub.invalid.names = false (io.confluent.connect.avro.AvroDataConfig:376)" "[2025-02-17 14:21:31,317] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] ProducerConfig values: acks = -1 batch.size = 16384 bootstrap.servers = [10.110.24.232:9092, 10.110.24.233:9092, 10.110.24.234:9092, 10.110.24.235:9092, 10.110.24.236:9092] buffer.memory = 100814346 client.dns.lookup = use_all_dns_ips client.id = connector-producer-source_os_inter_scenario_avro_mongodb_for_bigdata-0 compression.type = lz4 connections.max.idle.ms = 540000 delivery.timeout.ms = 2147483647 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 9223372036854775807 max.in.flight.requests.per.connection = 1 max.request.size = 100814346 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = SCRAM-SHA-256 sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer (org.apache.kafka.clients.producer.ProducerConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroSerializerConfig:376)" [2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Set up the value converter class io.confluent.connect.avro.AvroConverter for task source_os_inter_scenario_avro_mongodb_for_bigdata-0 using the worker config (org.apache.kafka.connect.runtime.Worker:570) "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mongodb.MongoDbConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = null name = source_os_inter_scenario_avro_mongodb_for_bigdata predicates = [] tasks.max = 1 topic.creation.groups = [] transforms = [] value.converter = null (org.apache.kafka.connect.runtime.SourceConnectorConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] AvroDataConfig values: connect.meta.data = true discard.type.doc.default = false enhanced.avro.schema.support = false generalized.sum.type.support = false schemas.cache.config = 1000 scrub.invalid.names = false (io.confluent.connect.avro.AvroDataConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mongodb.MongoDbConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = null name = source_os_inter_scenario_avro_mongodb_for_bigdata predicates = [] tasks.max = 1 topic.creation.groups = [] transforms = [] value.converter = null (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] specific.avro.reader = false use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroDeserializerConfig:376)" [2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Set up the key converter class io.confluent.connect.avro.AvroConverter for task source_os_inter_scenario_avro_mongodb_for_bigdata-0 using the worker config (org.apache.kafka.connect.runtime.Worker:564) "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] KafkaAvroDeserializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.use.logical.type.converters = false basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] specific.avro.reader = false use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroDeserializerConfig:376)" "[2025-02-17 14:21:31,317] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] EnrichedConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mongodb.MongoDbConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = null name = source_os_inter_scenario_avro_mongodb_for_bigdata predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 5 topic.creation.default.replication.factor = 3 topic.creation.groups = [] transforms = [] value.converter = null (org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] AvroConverterConfig values: auto.register.schemas = true basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.connect.avro.AvroConverterConfig:376)" "[2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] KafkaAvroSerializerConfig values: auto.register.schemas = true avro.reflection.allow.null = false avro.remove.java.properties = false avro.use.logical.type.converters = false basic.auth.credentials.source = USER_INFO basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN bearer.auth.token = [hidden] context.name.strategy = class io.confluent.kafka.serializers.context.NullContextNameStrategy id.compatibility.strict = true key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy latest.compatibility.strict = true max.schemas.per.subject = 1000 normalize.schemas = false proxy.host = proxy.port = -1 schema.reflection = false schema.registry.basic.auth.user.info = [hidden] schema.registry.ssl.cipher.suites = null schema.registry.ssl.enabled.protocols = [TLSv1.2, TLSv1.3] schema.registry.ssl.endpoint.identification.algorithm = https schema.registry.ssl.engine.factory.class = null schema.registry.ssl.key.password = null schema.registry.ssl.keymanager.algorithm = SunX509 schema.registry.ssl.keystore.certificate.chain = null schema.registry.ssl.keystore.key = null schema.registry.ssl.keystore.location = null schema.registry.ssl.keystore.password = null schema.registry.ssl.keystore.type = JKS schema.registry.ssl.protocol = TLSv1.3 schema.registry.ssl.provider = null schema.registry.ssl.secure.random.implementation = null schema.registry.ssl.trustmanager.algorithm = PKIX schema.registry.ssl.truststore.certificates = null schema.registry.ssl.truststore.location = null schema.registry.ssl.truststore.password = null schema.registry.ssl.truststore.type = JKS schema.registry.url = [https://apigw-data.ghtklab.com/data/schema] use.latest.version = false use.schema.id = -1 value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy (io.confluent.kafka.serializers.KafkaAvroSerializerConfig:376)" "[2025-02-17 14:21:31,317] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] EnrichedSourceConnectorConfig values: config.action.reload = restart connector.class = io.debezium.connector.mongodb.MongoDbConnector errors.log.enable = false errors.log.include.messages = false errors.retry.delay.max.ms = 60000 errors.retry.timeout = 0 errors.tolerance = none header.converter = null key.converter = null name = source_os_inter_scenario_avro_mongodb_for_bigdata predicates = [] tasks.max = 1 topic.creation.default.exclude = [] topic.creation.default.include = [.*] topic.creation.default.partitions = 5 topic.creation.default.replication.factor = 3 topic.creation.groups = [] transforms = [] value.converter = null (org.apache.kafka.connect.runtime.SourceConnectorConfig$EnrichedSourceConnectorConfig:376)" [2025-02-17 14:21:31,316] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task source_os_inter_scenario_avro_mongodb_for_bigdata-0 using the worker config (org.apache.kafka.connect.runtime.Worker:578) [2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka version: 7.2.2-ce (org.apache.kafka.common.utils.AppInfoParser:119) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'config.providers.file.class' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,320] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'key.converter.basic.auth.credentials.source' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.commit.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.task' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'status.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.commit.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.connector' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'offset.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka commitId: f53287cf2d7ff684 (org.apache.kafka.common.utils.AppInfoParser:120) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'value.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.type' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.version' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.type' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'expose.internal.connect.endpoints' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,318] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka commitId: f53287cf2d7ff684 (org.apache.kafka.common.utils.AppInfoParser:120) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.version' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'plugin.path' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'key.converter.basic.auth.user.info' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'value.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka startTimeMs: 1739776891318 (org.apache.kafka.common.utils.AppInfoParser:121) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) "[2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] TracerConfig values: trace.records.enable = false trace.records.header.converter = class org.apache.kafka.connect.storage.SimpleHeaderConverter trace.records.key.converter = class org.apache.kafka.connect.json.JsonConverter trace.records.predicates = [] trace.records.topic = connect-traces trace.records.topic.partition = 1 trace.records.topic.replication.factor = 3 trace.records.transforms = [] trace.records.value.converter = class org.apache.kafka.connect.json.JsonConverter (org.apache.kafka.connect.runtime.tracing.TracerConfig:376)" [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.connector' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'config.storage.topic' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'rest.extension.classes' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'config.providers' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'key.converter' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) "[2025-02-17 14:21:31,318] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] AdminClientConfig values: bootstrap.servers = [10.110.24.232:9092, 10.110.24.233:9092, 10.110.24.234:9092, 10.110.24.235:9092, 10.110.24.236:9092] client.dns.lookup = use_all_dns_ips client.id = connector-adminclient-source_os_inter_scenario_avro_mongodb_for_bigdata-0 connections.max.idle.ms = 300000 default.api.timeout.ms = 60000 host.resolver.class = class org.apache.kafka.clients.DefaultHostResolver metadata.max.age.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 receive.buffer.bytes = 65536 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = SCRAM-SHA-256 sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SASL_PLAINTEXT security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS (org.apache.kafka.clients.admin.AdminClientConfig:376)" [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'listeners' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,320] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'key.converter.schema.registry.url' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'value.converter.basic.auth.user.info' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'value.converter.basic.auth.credentials.source' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'basic.auth.credentials.source' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.resource.task' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'config.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'basic.auth.user.info' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. (org.apache.kafka.clients.producer.ProducerConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'session.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,318] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka version: 7.2.2-ce (org.apache.kafka.common.utils.AppInfoParser:119) [2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Kafka startTimeMs: 1739776891320 (org.apache.kafka.common.utils.AppInfoParser:121) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,319] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The configuration 'status.storage.replication.factor' was supplied but isn't a known config. (org.apache.kafka.clients.admin.AdminClientConfig:384) [2025-02-17 14:21:31,338] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,369] INFO [Producer clientId=producer-2] Resetting the last seen epoch of partition connect-status-avro-8823-0 to 0 since the associated topicId changed from null to kJjb5h4BRESCN4-NMCPZ0g (org.apache.kafka.clients.Metadata:402) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] topic.creation.default.replication.factor = 3 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] tombstones.on.delete = false (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] producer.override.max.request.size = 100814346 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Starting MongoDbConnectorTask with configuration: (io.debezium.connector.common.BaseSourceTask:245) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] capture.mode = change_streams_update_full (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,337] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:31,369] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = change-event-source-coordinator (io.debezium.util.Threads:270) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,369] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-change-event-source-coordinator (io.debezium.util.Threads:287) [2025-02-17 14:21:31,342] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=2213500, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583843573575} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] mongodb.password = ******** (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1415448, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583839768603} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,369] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Last recorded offset is no longer available on the server. (io.debezium.connector.mongodb.MongoDbConnectorTask:295) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] topic.creation.default.partitions = 5 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,369] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] WorkerSourceTask{id=source_os_inter_scenario_avro_mongodb_for_bigdata-0} Source task finished initialization and start (org.apache.kafka.connect.runtime.WorkerSourceTask:231) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] heartbeat.interval.ms = 300000 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] mongodb.user = etl_streaming (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] producer.override.buffer.memory = 100814346 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] task.class = io.debezium.connector.mongodb.MongoDbConnectorTask (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] name = source_os_inter_scenario_avro_mongodb_for_bigdata (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] topic.creation.default.retention.ms = 259200000 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] snapshot.mode = never (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,340] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,337] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,333] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found previous partition offset MongoDbPartition [sourcePartition={server_id=os_inter_scenario_avro_mongodb_bigdata}]: {sec=1739745313, ord=28, resume_token=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=} (io.debezium.connector.common.BaseSourceTask:527) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] collection.include.list = os_inter_scenario.* (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,369] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = blocking-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] topic.prefix = os_inter_scenario_avro_mongodb_bigdata (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found existing offset for at {sec=1739745313, ord=28, resume_token=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=} (io.debezium.connector.mongodb.connection.MongoDbConnection:202) [2025-02-17 14:21:31,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Initializing: org.apache.kafka.connect.runtime.TransformationChain{} (org.apache.kafka.connect.runtime.Worker:678) [2025-02-17 14:21:31,341] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1144715, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583842294764} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] [Producer clientId=connector-producer-source_os_inter_scenario_avro_mongodb_for_bigdata-0] Cluster ID: gx9uTj1DQ4-WNbx22Wuzpw (org.apache.kafka.clients.Metadata:287) [2025-02-17 14:21:31,370] INFO [Producer clientId=producer-2] Resetting the last seen epoch of partition connect-status-avro-8823-4 to 0 since the associated topicId changed from null to kJjb5h4BRESCN4-NMCPZ0g (org.apache.kafka.clients.Metadata:402) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] connect.max.attempts = 4 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,370] INFO [Producer clientId=producer-2] Resetting the last seen epoch of partition connect-status-avro-8823-1 to 0 since the associated topicId changed from null to kJjb5h4BRESCN4-NMCPZ0g (org.apache.kafka.clients.Metadata:402) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom topic naming strategy plugin: io.debezium.schema.DefaultTopicNamingStrategy (io.debezium.config.CommonConnectorConfig:1401) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] tasks.max = 1 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] errors.max.retries = 5 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,336] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] producer.override.compression.type = lz4 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] cursor.oversize.handling.mode = fail (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,322] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1385798, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583840255504} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,334] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = SignalProcessor (io.debezium.util.Threads:270) [2025-02-17 14:21:31,341] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1592097, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583843124883} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] database.include.list = os_inter_scenario (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,340] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,342] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,339] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1263611, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583839855586} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] mongodb.connection.string = ******** (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,335] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found a not connector specific implementation io.debezium.snapshot.lock.NoLockingSupport for lock mode no_locking_support (io.debezium.snapshot.SnapshotLockProvider:82) [2025-02-17 14:21:31,369] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The connector is trying to read change stream starting at MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]], but this is no longer available on the server. Reconfigure the connector to use a snapshot when needed if you want to recover. If not the connector will streaming from the last available position in the log (io.debezium.connector.mongodb.MongoDbConnectorTask:308) [2025-02-17 14:21:31,340] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,370] INFO [Producer clientId=producer-2] Resetting the last seen epoch of partition connect-status-avro-8823-2 to 0 since the associated topicId changed from null to kJjb5h4BRESCN4-NMCPZ0g (org.apache.kafka.clients.Metadata:402) [2025-02-17 14:21:31,362] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Invalid resume token present, snapshot will be performed' (io.debezium.connector.mongodb.connection.MongoDbConnection:224) [2025-02-17 14:21:31,337] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@53cc9fb4}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] connector.class = io.debezium.connector.mongodb.MongoDbConnector (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,340] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] max.queue.size.in.bytes = 100814346 (io.debezium.connector.common.BaseSourceTask:247) [2025-02-17 14:21:31,373] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Reading change stream (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:100) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,375] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,371] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,370] INFO [Producer clientId=producer-2] Resetting the last seen epoch of partition connect-status-avro-8823-3 to 0 since the associated topicId changed from null to kJjb5h4BRESCN4-NMCPZ0g (org.apache.kafka.clients.Metadata:402) [2025-02-17 14:21:31,371] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No incremental snapshot in progress, no action needed on start (io.debezium.connector.mongodb.snapshot.MongoDbIncrementalSnapshotChangeEventSource:262) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Metrics registered (io.debezium.pipeline.ChangeEventSourceCoordinator:137) [2025-02-17 14:21:31,378] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,376] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1126144, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583877890917} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] WorkerSourceTask{id=source_os_inter_scenario_avro_mongodb_for_bigdata-0} Executing source task (org.apache.kafka.connect.runtime.WorkerSourceTask:248) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor started. Scheduling it every 5000ms (io.debezium.pipeline.signal.SignalProcessor:105) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = mongodb named = incremental-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:31,375] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,376] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=808492, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583877742138} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] A previous offset indicating a completed snapshot has been found. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:144) [2025-02-17 14:21:31,405] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting fetcher thread termination (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:457) [2025-02-17 14:21:31,378] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher submitted for execution: io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher@442520eb @ java.util.concurrent.ThreadPoolExecutor@41bf0015[Running, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:367) [2025-02-17 14:21:31,378] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Canonical address hn-fornix-production-logistic-mongodb-linelv-02:27017 does not match server address. Removing 10.110.98.198:27017 from client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,373] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@551eaef1}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,373] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = replicator-fetcher (io.debezium.util.Threads:270) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1071260, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583875877207} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,373] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1109775, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583876070529} (org.mongodb.driver.cluster:71) "[2025-02-17 14:21:31,391] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher thread has failed (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:273) com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""EQGweXyzqk7oWDTrR6Al3AgvfP8="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842)" [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'true' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-SignalProcessor (io.debezium.util.Threads:287) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-replicator-fetcher-0 (io.debezium.util.Threads:287) "[2025-02-17 14:21:31,406] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Error while attempting to Reading change stream: Unable to fetch change stream events (io.debezium.connector.mongodb.connection.MongoDbConnections:52) io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""EQGweXyzqk7oWDTrR6Al3AgvfP8="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] According to the connector configuration, no snapshot will occur. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:151) [2025-02-17 14:21:31,379] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1105534, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583880709445} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,373] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,378] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1262827, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:31 ICT 2025, lastUpdateTimeNanos=10798583875916273} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Snapshot ended with SnapshotResult [status=SKIPPED, offset=MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]]] (io.debezium.pipeline.ChangeEventSourceCoordinator:298) [2025-02-17 14:21:31,372] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Starting streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:323) [2025-02-17 14:21:31,374] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Resuming streaming from token '0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=' (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:207) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:31,370] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Context created (io.debezium.pipeline.ChangeEventSourceCoordinator:140) [2025-02-17 14:21:31,379] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) "[2025-02-17 14:21:31,406] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Streaming failed (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:89) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""EQGweXyzqk7oWDTrR6Al3AgvfP8="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:31,870] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Debezium ServiceRegistry stopped. (io.debezium.service.DefaultServiceRegistry:105) "[2025-02-17 14:21:31,870] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] WorkerSourceTask{id=source_os_inter_scenario_avro_mongodb_for_bigdata-0} failed to poll records from SourceTask. Will retry operation. (org.apache.kafka.connect.runtime.WorkerSourceTask:311) org.apache.kafka.connect.errors.RetriableException: An exception occurred in the change event producer. This connector will be restarted. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:63) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:90) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) ... 9 more Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""EQGweXyzqk7oWDTrR6Al3AgvfP8="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:31,406] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Retry 1 of 5 retries will be attempted (io.debezium.pipeline.ErrorHandler:125) [2025-02-17 14:21:31,870] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Going to restart connector after 10 sec. after a retriable exception (io.debezium.connector.common.BaseSourceTask:429) [2025-02-17 14:21:31,406] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'false' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:21:31,870] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting end of restart backoff period after a retriable error (io.debezium.connector.common.BaseSourceTask:408) [2025-02-17 14:21:31,870] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor stopped (io.debezium.pipeline.signal.SignalProcessor:127) "[2025-02-17 14:21:31,406] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Producer failure (io.debezium.pipeline.ErrorHandler:52) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""EQGweXyzqk7oWDTrR6Al3AgvfP8="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776891, ""i"": 113}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:31,406] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Finished streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:325) [2025-02-17 14:21:41,320] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0|offsets] Couldn't commit processed log positions with the source database due to a concurrent connector shutdown or restart (io.debezium.connector.common.BaseSourceTask:499) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,876] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:41,880] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,891] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Invalid resume token present, snapshot will be performed' (io.debezium.connector.mongodb.connection.MongoDbConnection:224) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Context created (io.debezium.pipeline.ChangeEventSourceCoordinator:140) [2025-02-17 14:21:41,891] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Last recorded offset is no longer available on the server. (io.debezium.connector.mongodb.MongoDbConnectorTask:295) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1666092, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594379615120} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,870] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Attempting to restart task. (io.debezium.connector.common.BaseSourceTask:400) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,891] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The connector is trying to read change stream starting at MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]], but this is no longer available on the server. Reconfigure the connector to use a snapshot when needed if you want to recover. If not the connector will streaming from the last available position in the log (io.debezium.connector.mongodb.MongoDbConnectorTask:308) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = mongodb named = incremental-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor started. Scheduling it every 5000ms (io.debezium.pipeline.signal.SignalProcessor:105) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,879] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=932563, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594380958834} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] A previous offset indicating a completed snapshot has been found. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:144) [2025-02-17 14:21:41,875] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found a not connector specific implementation io.debezium.snapshot.lock.NoLockingSupport for lock mode no_locking_support (io.debezium.snapshot.SnapshotLockProvider:82) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = change-event-source-coordinator (io.debezium.util.Threads:270) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No incremental snapshot in progress, no action needed on start (io.debezium.connector.mongodb.snapshot.MongoDbIncrementalSnapshotChangeEventSource:262) [2025-02-17 14:21:41,879] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=988651, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594381193431} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = blocking-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1396970, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594379587621} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,877] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,880] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1436394, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594381256524} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,877] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,876] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Metrics registered (io.debezium.pipeline.ChangeEventSourceCoordinator:137) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-SignalProcessor (io.debezium.util.Threads:287) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Snapshot ended with SnapshotResult [status=SKIPPED, offset=MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]]] (io.debezium.pipeline.ChangeEventSourceCoordinator:298) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,879] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'true' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Successfully restarted task (io.debezium.connector.common.BaseSourceTask:402) [2025-02-17 14:21:41,892] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-change-event-source-coordinator (io.debezium.util.Threads:287) [2025-02-17 14:21:41,876] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@4f6ee640}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:41,876] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom topic naming strategy plugin: io.debezium.schema.DefaultTopicNamingStrategy (io.debezium.config.CommonConnectorConfig:1401) [2025-02-17 14:21:41,893] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] According to the connector configuration, no snapshot will occur. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:151) [2025-02-17 14:21:41,878] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,877] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1151567, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594378806718} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,875] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found existing offset for at {sec=1739745313, ord=28, resume_token=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=} (io.debezium.connector.mongodb.connection.MongoDbConnection:202) [2025-02-17 14:21:41,871] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,872] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = SignalProcessor (io.debezium.util.Threads:270) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Starting streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:323) [2025-02-17 14:21:41,895] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Reading change stream (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:100) [2025-02-17 14:21:41,895] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) "[2025-02-17 14:21:41,913] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Error while attempting to Reading change stream: Unable to fetch change stream events (io.debezium.connector.mongodb.connection.MongoDbConnections:52) io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 272}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ZvG/79FcK6A9QfUsU6KHhmpY0FU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 271}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" "[2025-02-17 14:21:41,913] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Producer failure (io.debezium.pipeline.ErrorHandler:52) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 272}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ZvG/79FcK6A9QfUsU6KHhmpY0FU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 271}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=769944, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594397811668} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,897] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=903033, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594397751376} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,897] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,899] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,913] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Retry 2 of 5 retries will be attempted (io.debezium.pipeline.ErrorHandler:125) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,912] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting fetcher thread termination (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:457) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,895] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Resuming streaming from token '0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=' (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:207) "[2025-02-17 14:21:41,913] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Streaming failed (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:89) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 272}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ZvG/79FcK6A9QfUsU6KHhmpY0FU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 271}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:41,898] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=960217, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594399706953} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,897] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-replicator-fetcher-0 (io.debezium.util.Threads:287) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=898614, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594398155158} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,895] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@656a4b02}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,897] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,899] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1257245, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594400322676} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,897] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = replicator-fetcher (io.debezium.util.Threads:270) [2025-02-17 14:21:41,898] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1222940, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:41 ICT 2025, lastUpdateTimeNanos=10798594399618921} (org.mongodb.driver.cluster:71) "[2025-02-17 14:21:41,907] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher thread has failed (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:273) com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 272}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ZvG/79FcK6A9QfUsU6KHhmpY0FU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 271}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842)" [2025-02-17 14:21:41,894] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:41,896] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher submitted for execution: io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher@7f41e5cd @ java.util.concurrent.ThreadPoolExecutor@3fcbe3bd[Running, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:367) [2025-02-17 14:21:41,895] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:41,913] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Finished streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:325) "[2025-02-17 14:21:42,393] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] WorkerSourceTask{id=source_os_inter_scenario_avro_mongodb_for_bigdata-0} failed to poll records from SourceTask. Will retry operation. (org.apache.kafka.connect.runtime.WorkerSourceTask:311) org.apache.kafka.connect.errors.RetriableException: An exception occurred in the change event producer. This connector will be restarted. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:63) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:90) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) ... 9 more Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 272}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ZvG/79FcK6A9QfUsU6KHhmpY0FU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776901, ""i"": 271}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:42,392] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Going to restart connector after 10 sec. after a retriable exception (io.debezium.connector.common.BaseSourceTask:429) [2025-02-17 14:21:42,392] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor stopped (io.debezium.pipeline.signal.SignalProcessor:127) [2025-02-17 14:21:42,392] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Debezium ServiceRegistry stopped. (io.debezium.service.DefaultServiceRegistry:105) [2025-02-17 14:21:42,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting end of restart backoff period after a retriable error (io.debezium.connector.common.BaseSourceTask:408) [2025-02-17 14:21:41,913] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'false' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:21:41,734] INFO [AdminClient clientId=adminclient-8] Node 233 disconnected. (org.apache.kafka.clients.NetworkClient:1047) [2025-02-17 14:20:38,700] INFO [source_new_bag_mongo_avro_mongodb_for_bigdata|task-0] [AdminClient clientId=connector-adminclient-source_new_bag_mongo_avro_mongodb_for_bigdata-0] Node 232 disconnected. (org.apache.kafka.clients.NetworkClient:1047) [2025-02-17 14:21:51,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0|offsets] Couldn't commit processed log positions with the source database due to a concurrent connector shutdown or restart (io.debezium.connector.common.BaseSourceTask:499) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,394] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom topic naming strategy plugin: io.debezium.schema.DefaultTopicNamingStrategy (io.debezium.config.CommonConnectorConfig:1401) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Attempting to restart task. (io.debezium.connector.common.BaseSourceTask:400) [2025-02-17 14:21:52,397] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found a not connector specific implementation io.debezium.snapshot.lock.NoLockingSupport for lock mode no_locking_support (io.debezium.snapshot.SnapshotLockProvider:82) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,394] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,394] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,393] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,395] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = SignalProcessor (io.debezium.util.Threads:270) [2025-02-17 14:21:52,394] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,414] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Last recorded offset is no longer available on the server. (io.debezium.connector.mongodb.MongoDbConnectorTask:295) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Metrics registered (io.debezium.pipeline.ChangeEventSourceCoordinator:137) [2025-02-17 14:21:52,399] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,416] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-SignalProcessor (io.debezium.util.Threads:287) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = mongodb named = incremental-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:52,413] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Invalid resume token present, snapshot will be performed' (io.debezium.connector.mongodb.connection.MongoDbConnection:224) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,399] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1077595, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604900986318} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,398] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@516493d}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:52,421] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1613437, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604922990342} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,420] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,399] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = replicator-fetcher (io.debezium.util.Threads:270) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] A previous offset indicating a completed snapshot has been found. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:144) [2025-02-17 14:21:52,398] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,401] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1024963, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604902946517} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Context created (io.debezium.pipeline.ChangeEventSourceCoordinator:140) [2025-02-17 14:21:52,418] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Resuming streaming from token '0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=' (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:207) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher submitted for execution: io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher@3984c401 @ java.util.concurrent.ThreadPoolExecutor@3eb74307[Running, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:367) [2025-02-17 14:21:52,418] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@2a80985d}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Starting streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:323) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found existing offset for at {sec=1739745313, ord=28, resume_token=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=} (io.debezium.connector.mongodb.connection.MongoDbConnection:202) [2025-02-17 14:21:52,418] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:52,420] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The connector is trying to read change stream starting at MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]], but this is no longer available on the server. Reconfigure the connector to use a snapshot when needed if you want to recover. If not the connector will streaming from the last available position in the log (io.debezium.connector.mongodb.MongoDbConnectorTask:308) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,418] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = change-event-source-coordinator (io.debezium.util.Threads:270) [2025-02-17 14:21:52,398] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,421] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Snapshot ended with SnapshotResult [status=SKIPPED, offset=MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]]] (io.debezium.pipeline.ChangeEventSourceCoordinator:298) [2025-02-17 14:21:52,397] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1056768, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604920812698} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,399] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:21:52,402] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=888345, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604920341909} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,421] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=903636, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604922570832} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,402] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1112395, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604903273961} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-replicator-fetcher-0 (io.debezium.util.Threads:287) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No incremental snapshot in progress, no action needed on start (io.debezium.connector.mongodb.snapshot.MongoDbIncrementalSnapshotChangeEventSource:262) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=953602, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604901250165} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,423] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=2090883, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604924770354} (org.mongodb.driver.cluster:71) "[2025-02-17 14:21:52,431] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher thread has failed (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:273) com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ga214hLtekdbYscHn5z07HiT4Xo="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842)" [2025-02-17 14:21:52,420] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] According to the connector configuration, no snapshot will occur. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:151) [2025-02-17 14:21:52,420] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = blocking-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:21:52,416] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor started. Scheduling it every 5000ms (io.debezium.pipeline.signal.SignalProcessor:105) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Successfully restarted task (io.debezium.connector.common.BaseSourceTask:402) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1357980, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604901456118} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,415] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'true' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:21:52,419] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,400] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:21:52,414] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-change-event-source-coordinator (io.debezium.util.Threads:287) [2025-02-17 14:21:52,418] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Reading change stream (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:100) [2025-02-17 14:21:52,417] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,401] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1179647, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:21:52 ICT 2025, lastUpdateTimeNanos=10798604902962006} (org.mongodb.driver.cluster:71) [2025-02-17 14:21:52,914] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Going to restart connector after 10 sec. after a retriable exception (io.debezium.connector.common.BaseSourceTask:429) [2025-02-17 14:21:52,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Debezium ServiceRegistry stopped. (io.debezium.service.DefaultServiceRegistry:105) "[2025-02-17 14:21:52,435] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Streaming failed (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:89) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ga214hLtekdbYscHn5z07HiT4Xo="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:52,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting end of restart backoff period after a retriable error (io.debezium.connector.common.BaseSourceTask:408) [2025-02-17 14:21:52,434] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting fetcher thread termination (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:457) [2025-02-17 14:21:52,435] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Retry 3 of 5 retries will be attempted (io.debezium.pipeline.ErrorHandler:125) [2025-02-17 14:21:52,435] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'false' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) "[2025-02-17 14:21:52,435] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Producer failure (io.debezium.pipeline.ErrorHandler:52) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ga214hLtekdbYscHn5z07HiT4Xo="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:52,435] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Finished streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:325) "[2025-02-17 14:21:52,915] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] WorkerSourceTask{id=source_os_inter_scenario_avro_mongodb_for_bigdata-0} failed to poll records from SourceTask. Will retry operation. (org.apache.kafka.connect.runtime.WorkerSourceTask:311) org.apache.kafka.connect.errors.RetriableException: An exception occurred in the change event producer. This connector will be restarted. at io.debezium.pipeline.ErrorHandler.setProducerThrowable(ErrorHandler.java:63) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:90) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) ... 9 more Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ga214hLtekdbYscHn5z07HiT4Xo="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:21:52,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor stopped (io.debezium.pipeline.signal.SignalProcessor:127) "[2025-02-17 14:21:52,434] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Error while attempting to Reading change stream: Unable to fetch change stream events (io.debezium.connector.mongodb.connection.MongoDbConnections:52) io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""ga214hLtekdbYscHn5z07HiT4Xo="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776912, ""i"": 178}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=949964, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615423282380} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,923] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,924] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,919] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-02:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,923] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,924] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1436154, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615425533566} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1321758, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615423837464} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,932] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Invalid resume token present, snapshot will be performed' (io.debezium.connector.mongodb.connection.MongoDbConnection:224) [2025-02-17 14:22:02,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Attempting to restart task. (io.debezium.connector.common.BaseSourceTask:400) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,923] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1527359, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615424212897} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,933] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Last recorded offset is no longer available on the server. (io.debezium.connector.mongodb.MongoDbConnectorTask:295) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:01,321] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0|offsets] Couldn't commit processed log positions with the source database due to a concurrent connector shutdown or restart (io.debezium.connector.common.BaseSourceTask:499) [2025-02-17 14:22:02,919] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found a not connector specific implementation io.debezium.snapshot.lock.NoLockingSupport for lock mode no_locking_support (io.debezium.snapshot.SnapshotLockProvider:82) [2025-02-17 14:22:02,923] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=869478, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615425166042} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,923] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom topic naming strategy plugin: io.debezium.schema.DefaultTopicNamingStrategy (io.debezium.config.CommonConnectorConfig:1401) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,916] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,933] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] The connector is trying to read change stream starting at MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]], but this is no longer available on the server. Reconfigure the connector to use a snapshot when needed if you want to recover. If not the connector will streaming from the last available position in the log (io.debezium.connector.mongodb.MongoDbConnectorTask:308) [2025-02-17 14:22:02,915] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,921] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@437779c9}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:22:02,924] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=868865, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615425420657} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,920] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Found existing offset for at {sec=1739745313, ord=28, resume_token=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=} (io.debezium.connector.mongodb.connection.MongoDbConnection:202) [2025-02-17 14:22:02,922] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Snapshot ended with SnapshotResult [status=SKIPPED, offset=MongoDbOffsetContext [sourceInfo=SourceInfo [initialSync=false, collectionId=null, position=Position [ts=Timestamp{value=7472149222704283676, seconds=1739745313, inc=28}, changeStreamSessionTxnId=null, resumeToken=0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=]]]] (io.debezium.pipeline.ChangeEventSourceCoordinator:298) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-SignalProcessor (io.debezium.util.Threads:287) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Effective change stream pipeline: [{"$replaceRoot": {"newRoot": {"event": "$$ROOT", "namespace": {"$concat": ["$ns.db", ".", "$ns.coll"]}}}}, {"$match": {"$and": [{"$and": [{"event.ns.db": {"$regularExpression": {"pattern": "os_inter_scenario", "options": "i"}}}, {"namespace": {"$regularExpression": {"pattern": "os_inter_scenario.*", "options": "i"}}}]}, {"event.operationType": {"$in": ["insert", "update", "replace", "delete"]}}]}}, {"$replaceRoot": {"newRoot": "$event"}}] (io.debezium.connector.mongodb.ChangeStreamPipelineFactory:56) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.198:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Reading change stream (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:100) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Starting streaming (io.debezium.pipeline.ChangeEventSourceCoordinator:323) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Metrics registered (io.debezium.pipeline.ChangeEventSourceCoordinator:137) [2025-02-17 14:22:02,940] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-02:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1263468, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615442034176} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,933] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = blocking-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:22:02,940] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-03:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1133284, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615442009796} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,937] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.198:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] A previous offset indicating a completed snapshot has been found. (io.debezium.connector.mongodb.MongoDbSnapshotChangeEventSource:144) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'true' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.199:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,935] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-01:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server hn-fornix-production-logistic-mongodb-linelv-03:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Creating thread debezium-mongodbconnector-os_inter_scenario_avro_mongodb_bigdata-replicator-fetcher-0 (io.debezium.util.Threads:287) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = mongodb named = incremental-snapshot (io.debezium.util.Threads:270) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = replicator-fetcher (io.debezium.util.Threads:270) "[2025-02-17 14:22:02,949] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Fetcher thread has failed (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:273) com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 884}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""Wh9IIK5zxyYtsDFqM+7Z+HN3nTU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 883}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842)" [2025-02-17 14:22:02,941] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary hn-fornix-production-logistic-mongodb-linelv-01:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,941] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=hn-fornix-production-logistic-mongodb-linelv-01:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1415364, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615442540984} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.198:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1269634, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-02:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=673cb590144a2b6ad2caf340, counter=3}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615440537014} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No server chosen by ReadPreferenceServerSelector{readPreference=primary} from cluster description ClusterDescription{type=REPLICA_SET, connectionMode=MULTIPLE, serverDescriptions=[ServerDescription{address=10.110.98.199:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.198:27017, type=UNKNOWN, state=CONNECTING}, ServerDescription{address=10.110.98.197:27017, type=UNKNOWN, state=CONNECTING}]}. Waiting for 30000 ms before timing out (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,935] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] No incremental snapshot in progress, no action needed on start (io.debezium.connector.mongodb.snapshot.MongoDbIncrementalSnapshotChangeEventSource:262) [2025-02-17 14:22:02,937] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.199:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Discovered replica set primary 10.110.98.197:27017 with max election id 7fffffff0000000000000011 and max set version 1 (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Resuming streaming from token '0QAAAAJfZGF0YQDBAAAAODI2N0IyNjgyMTAwMDAwMDFDMkIwNDJDMDEwMDI5NkU1QTEwMDQyREJGNUZFMUQwMzQ0NjgwQjFCNzlEMUUxMTUwMkYxNzQ2M0M2RjcwNjU3MjYxNzQ2OTZGNkU1NDc5NzA2NTAwM0M3NTcwNjQ2MTc0NjUwMDQ2NjQ2RjYzNzU2RDY1NkU3NDRCNjU3OTAwNDY2NDVGNjk2NDAwNjQ2N0IxREU5QzJFRjZGQkFCNjIwOTgxOEEwMDAwMkIwMjA0AAA=' (io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource:207) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.197:27017, type=REPLICA_SET_PRIMARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=894304, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-01:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=7fffffff0000000000000011, setVersion=1, topologyVersion=TopologyVersion{processId=66facde78a44cde354088f08, counter=11}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615439961611} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,933] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Requested thread factory for component MongoDbConnector, id = os_inter_scenario_avro_mongodb_bigdata named = change-event-source-coordinator (io.debezium.util.Threads:270) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,938] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] MongoClient with metadata {"driver": {"name": "mongo-java-driver|sync", "version": "4.11.0"}, "os": {"type": "Linux", "name": "Linux", "architecture": "amd64", "version": "5.4.0-196-generic"}, "platform": "Java/Oracle Corporation/17.0.13+10-LTS-268"} created with settings MongoClientSettings{readPreference=primary, writeConcern=WriteConcern{w=null, wTimeout=null ms, journal=null}, retryWrites=true, retryReads=true, readConcern=ReadConcern{level=null}, credential=MongoCredential{mechanism=null, userName='etl_streaming', source='admin', password=, mechanismProperties=}, transportSettings=null, streamFactoryFactory=null, commandListeners=[], codecRegistry=ProvidersCodecRegistry{codecProviders=[ValueCodecProvider{}, BsonValueCodecProvider{}, DBRefCodecProvider{}, DBObjectCodecProvider{}, DocumentCodecProvider{}, CollectionCodecProvider{}, IterableCodecProvider{}, MapCodecProvider{}, GeoJsonCodecProvider{}, GridFSFileCodecProvider{}, Jsr310CodecProvider{}, JsonObjectCodecProvider{}, BsonCodecProvider{}, EnumCodecProvider{}, com.mongodb.client.model.mql.ExpressionCodecProvider@336a48a1, com.mongodb.Jep395RecordCodecProvider@24b54547, com.mongodb.KotlinCodecProvider@49b53705]}, loggerSettings=LoggerSettings{maxDocumentLength=1000}, clusterSettings={hosts=[10.110.98.197:27017, 10.110.98.198:27017, 10.110.98.199:27017], srvServiceName=mongodb, mode=MULTIPLE, requiredClusterType=REPLICA_SET, requiredReplicaSetName='inter_scenario', serverSelector='null', clusterListeners='[]', serverSelectionTimeout='30000 ms', localThreshold='15 ms'}, socketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=0, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, heartbeatSocketSettings=SocketSettings{connectTimeoutMS=10000, readTimeoutMS=10000, receiveBufferSize=0, proxySettings=ProxySettings{host=null, port=null, username=null, password=null}}, connectionPoolSettings=ConnectionPoolSettings{maxSize=100, minSize=0, maxWaitTimeMS=120000, maxConnectionLifeTimeMS=0, maxConnectionIdleTimeMS=0, maintenanceInitialDelayMS=0, maintenanceFrequencyMS=60000, connectionPoolListeners=[], maxConnecting=2}, serverSettings=ServerSettings{heartbeatFrequencyMS=10000, minHeartbeatFrequencyMS=500, serverListeners='[]', serverMonitorListeners='[]'}, sslSettings=SslSettings{enabled=false, invalidHostNameAllowed=false, context=javax.net.ssl.SSLContext@7154753d}, applicationName='null', compressorList=[], uuidRepresentation=STANDARD, serverApi=null, autoEncryptionSettings=null, dnsClient=null, inetAddressResolver=null, contextProvider=null} (org.mongodb.driver.client:71) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Loading the custom source info struct maker plugin: io.debezium.connector.mongodb.MongoDbSourceInfoStructMaker (io.debezium.config.CommonConnectorConfig:1701) [2025-02-17 14:22:02,934] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Successfully restarted task (io.debezium.connector.common.BaseSourceTask:402) [2025-02-17 14:22:02,936] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor started. Scheduling it every 5000ms (io.debezium.pipeline.signal.SignalProcessor:105) [2025-02-17 14:22:02,937] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Adding discovered server 10.110.98.197:27017 to client view of cluster (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Monitor thread successfully connected to server with description ServerDescription{address=10.110.98.199:27017, type=REPLICA_SET_SECONDARY, state=CONNECTED, ok=true, minWireVersion=0, maxWireVersion=21, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1308629, setName='inter_scenario', canonicalAddress=hn-fornix-production-logistic-mongodb-linelv-03:27017, hosts=[hn-fornix-production-logistic-mongodb-linelv-02:27017, hn-fornix-production-logistic-mongodb-linelv-03:27017, hn-fornix-production-logistic-mongodb-linelv-01:27017], passives=[], arbiters=[], primary='hn-fornix-production-logistic-mongodb-linelv-01:27017', tagSet=TagSet{[]}, electionId=null, setVersion=1, topologyVersion=TopologyVersion{processId=66ce07fe2f1788d78be09072, counter=59}, lastWriteDate=Mon Feb 17 14:22:02 ICT 2025, lastUpdateTimeNanos=10798615440707218} (org.mongodb.driver.cluster:71) [2025-02-17 14:22:02,939] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Server 10.110.98.197:27017 is no longer a member of the replica set. Removing from client view of cluster. (org.mongodb.driver.cluster:71) [2025-02-17 14:22:03,434] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Debezium ServiceRegistry stopped. (io.debezium.service.DefaultServiceRegistry:105) [2025-02-17 14:22:03,434] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] SignalProcessor stopped (io.debezium.pipeline.signal.SignalProcessor:127) "[2025-02-17 14:22:02,955] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Error while attempting to Reading change stream: Unable to fetch change stream events (io.debezium.connector.mongodb.connection.MongoDbConnections:52) io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 884}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""Wh9IIK5zxyYtsDFqM+7Z+HN3nTU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 883}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:22:03,434] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Going to restart connector after 10 sec. after a retriable exception (io.debezium.connector.common.BaseSourceTask:429) [2025-02-17 14:22:02,954] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Awaiting fetcher thread termination (io.debezium.connector.mongodb.events.BufferingChangeStreamCursor:457) "[2025-02-17 14:22:02,955] ERROR [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Producer failure (io.debezium.pipeline.ErrorHandler:52) io.debezium.DebeziumException: Error while attempting to Reading change stream at io.debezium.connector.mongodb.connection.MongoDbConnections.lambda$eventSourcingErrorHandler$1(MongoDbConnections.java:53) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:111) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:88) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:84) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.execute(MongoDbStreamingChangeEventSource.java:37) at io.debezium.pipeline.ChangeEventSourceCoordinator.streamEvents(ChangeEventSourceCoordinator.java:324) at io.debezium.pipeline.ChangeEventSourceCoordinator.executeChangeEventSources(ChangeEventSourceCoordinator.java:203) at io.debezium.pipeline.ChangeEventSourceCoordinator.lambda$start$0(ChangeEventSourceCoordinator.java:143) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.base/java.lang.Thread.run(Thread.java:842) Caused by: io.debezium.DebeziumException: Unable to fetch change stream events at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.poll(BufferingChangeStreamCursor.java:235) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.pollWithDelay(BufferingChangeStreamCursor.java:405) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor.tryNext(BufferingChangeStreamCursor.java:374) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.readChangeStream(MongoDbStreamingChangeEventSource.java:107) at io.debezium.connector.mongodb.MongoDbStreamingChangeEventSource.lambda$execute$0(MongoDbStreamingChangeEventSource.java:85) at io.debezium.connector.mongodb.connection.MongoDbConnection.lambda$execute$0(MongoDbConnection.java:89) at io.debezium.connector.mongodb.connection.MongoDbConnection.execute(MongoDbConnection.java:105) ... 11 more Caused by: com.mongodb.MongoCommandException: Command failed with error 280 (ChangeStreamFatalError): 'To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage' on server hn-fornix-production-logistic-mongodb-linelv-01:27017. The full response is {""errorLabels"": [""NonResumableChangeStreamError""], ""ok"": 0.0, ""errmsg"": ""To resume from a split event, the $changeStream pipeline must include a $changeStreamSplitLargeEvent stage"", ""code"": 280, ""codeName"": ""ChangeStreamFatalError"", ""$clusterTime"": {""clusterTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 884}}, ""signature"": {""hash"": {""$binary"": {""base64"": ""Wh9IIK5zxyYtsDFqM+7Z+HN3nTU="", ""subType"": ""00""}}, ""keyId"": 7427895936975634460}}, ""operationTime"": {""$timestamp"": {""t"": 1739776922, ""i"": 883}}} at com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:205) at com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:454) at com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:372) at com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:114) at com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:765) at com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:76) at com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:209) at com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:115) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:83) at com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:74) at com.mongodb.internal.connection.DefaultServer$OperationCountTrackingConnection.command(DefaultServer.java:299) at com.mongodb.internal.operation.SyncOperationHelper.createReadCommandAndExecute(SyncOperationHelper.java:273) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$3(SyncOperationHelper.java:191) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$0(SyncOperationHelper.java:127) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.lambda$withSourceAndConnection$1(SyncOperationHelper.java:126) at com.mongodb.internal.operation.SyncOperationHelper.withSuppliedResource(SyncOperationHelper.java:152) at com.mongodb.internal.operation.SyncOperationHelper.withSourceAndConnection(SyncOperationHelper.java:125) at com.mongodb.internal.operation.SyncOperationHelper.lambda$executeRetryableRead$4(SyncOperationHelper.java:189) at com.mongodb.internal.operation.SyncOperationHelper.lambda$decorateReadWithRetries$12(SyncOperationHelper.java:292) at com.mongodb.internal.async.function.RetryingSyncSupplier.get(RetryingSyncSupplier.java:67) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:194) at com.mongodb.internal.operation.SyncOperationHelper.executeRetryableRead(SyncOperationHelper.java:176) at com.mongodb.internal.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:193) at com.mongodb.internal.operation.ChangeStreamOperation.lambda$execute$0(ChangeStreamOperation.java:187) at com.mongodb.internal.operation.SyncOperationHelper.withReadConnectionSource(SyncOperationHelper.java:99) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:185) at com.mongodb.internal.operation.ChangeStreamOperation.execute(ChangeStreamOperation.java:54) at com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:153) at com.mongodb.client.internal.ChangeStreamIterableImpl.execute(ChangeStreamIterableImpl.java:212) at com.mongodb.client.internal.ChangeStreamIterableImpl.cursor(ChangeStreamIterableImpl.java:187) at io.debezium.connector.mongodb.events.BufferingChangeStreamCursor$EventFetcher.run(BufferingChangeStreamCursor.java:260) ... 5 more" [2025-02-17 14:22:02,955] INFO [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Connected metrics set to 'false' (io.debezium.pipeline.ChangeEventSourceCoordinator:477) [2025-02-17 14:22:02,955] WARN [source_os_inter_scenario_avro_mongodb_for_bigdata|task-0] Retry 4 of 5 retries will be attempted (io.debezium.pipeline.ErrorHandler:125) [2025-02-17 14:21:59,356] INFO [AdminClient clientId=adminclient-8] Node 234 disconnected. (org.apache.kafka.clients.NetworkClient:1047)