18:26:49.515 [main] INFO io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - Initializing Cassandra connector task context ... 18:26:49.700 [main] INFO org.apache.cassandra.config.YamlConfigurationLoader - Configuration location: file:/app/debezium/cassandra.yaml 18:26:49.701 [main] DEBUG org.apache.cassandra.config.YamlConfigurationLoader - Loading settings from file:/app/debezium/cassandra.yaml 18:26:49.778 [main] DEBUG io.netty.util.internal.logging.InternalLoggerFactory - Using SLF4J as the default logging framework 18:26:49.791 [main] DEBUG io.netty.util.internal.PlatformDependent0 - -Dio.netty.noUnsafe: false 18:26:49.791 [main] DEBUG io.netty.util.internal.PlatformDependent0 - Java version: 11 18:26:49.792 [main] DEBUG io.netty.util.internal.PlatformDependent0 - sun.misc.Unsafe.theUnsafe: available 18:26:49.793 [main] DEBUG io.netty.util.internal.PlatformDependent0 - sun.misc.Unsafe.copyMemory: available 18:26:49.793 [main] DEBUG io.netty.util.internal.PlatformDependent0 - java.nio.Buffer.address: available 18:26:49.795 [main] DEBUG io.netty.util.internal.PlatformDependent0 - direct buffer constructor: unavailable java.lang.UnsupportedOperationException: Reflective setAccessible(true) disabled at io.netty.util.internal.ReflectionUtil.trySetAccessible(ReflectionUtil.java:31) at io.netty.util.internal.PlatformDependent0$4.run(PlatformDependent0.java:238) at java.base/java.security.AccessController.doPrivileged(Native Method) at io.netty.util.internal.PlatformDependent0.(PlatformDependent0.java:232) at io.netty.util.internal.PlatformDependent.isAndroid(PlatformDependent.java:293) at io.netty.util.internal.PlatformDependent.(PlatformDependent.java:92) at io.netty.handler.ssl.OpenSsl.(OpenSsl.java:124) at org.apache.cassandra.security.SSLFactory.(SSLFactory.java:100) at org.apache.cassandra.config.EncryptionOptions.(EncryptionOptions.java:242) at org.apache.cassandra.config.Config.(Config.java:284) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:330) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:308) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:301) at org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:167) at org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:331) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObjectNoCheck(BaseConstructor.java:230) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:219) at org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:173) at org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:157) at org.yaml.snakeyaml.Yaml.loadFromReader(Yaml.java:472) at org.yaml.snakeyaml.Yaml.loadAs(Yaml.java:466) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:206) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:134) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:109) at org.apache.cassandra.config.DatabaseDescriptor.loadConfig(DatabaseDescriptor.java:301) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:216) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:187) at io.debezium.connector.cassandra.CassandraConnectorTask$Cassandra4SchemaLoader.load(CassandraConnectorTask.java:26) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:51) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) 18:26:49.796 [main] DEBUG io.netty.util.internal.PlatformDependent0 - java.nio.Bits.unaligned: available, true 18:26:49.797 [main] DEBUG io.netty.util.internal.PlatformDependent0 - jdk.internal.misc.Unsafe.allocateUninitializedArray(int): available 18:26:49.797 [main] DEBUG io.netty.util.internal.PlatformDependent0 - java.nio.DirectByteBuffer.(long, int): unavailable 18:26:49.797 [main] DEBUG io.netty.util.internal.PlatformDependent - sun.misc.Unsafe: available 18:26:49.799 [main] DEBUG io.netty.util.internal.PlatformDependent - maxDirectMemory: 8589934592 bytes (maybe) 18:26:49.799 [main] DEBUG io.netty.util.internal.PlatformDependent - -Dio.netty.tmpdir: /tmp (java.io.tmpdir) 18:26:49.799 [main] DEBUG io.netty.util.internal.PlatformDependent - -Dio.netty.bitMode: 64 (sun.arch.data.model) 18:26:49.800 [main] DEBUG io.netty.util.internal.PlatformDependent - -Dio.netty.maxDirectMemory: -1 bytes 18:26:49.800 [main] DEBUG io.netty.util.internal.PlatformDependent - -Dio.netty.uninitializedArrayAllocationThreshold: 1024 18:26:49.800 [main] DEBUG io.netty.util.internal.CleanerJava9 - java.nio.ByteBuffer.cleaner(): available 18:26:49.801 [main] DEBUG io.netty.util.internal.PlatformDependent - -Dio.netty.noPreferDirect: false 18:26:49.803 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - -Dio.netty.native.workdir: /tmp (io.netty.tmpdir) 18:26:49.803 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - -Dio.netty.native.deleteLibAfterLoading: true 18:26:49.804 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - -Dio.netty.native.tryPatchShadedId: true 18:26:49.805 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - Unable to load the library 'netty_tcnative_linux_x86_64', trying other loading mechanism. java.lang.UnsatisfiedLinkError: no netty_tcnative_linux_x86_64 in java.library.path: [/usr/java/packages/lib, /usr/lib64, /lib64, /lib, /usr/lib] at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2670) at java.base/java.lang.Runtime.loadLibrary0(Runtime.java:830) at java.base/java.lang.System.loadLibrary(System.java:1873) at io.netty.util.internal.NativeLibraryUtil.loadLibrary(NativeLibraryUtil.java:38) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.netty.util.internal.NativeLibraryLoader$1.run(NativeLibraryLoader.java:385) at java.base/java.security.AccessController.doPrivileged(Native Method) at io.netty.util.internal.NativeLibraryLoader.loadLibraryByHelper(NativeLibraryLoader.java:377) at io.netty.util.internal.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:341) at io.netty.util.internal.NativeLibraryLoader.load(NativeLibraryLoader.java:136) at io.netty.util.internal.NativeLibraryLoader.loadFirstAvailable(NativeLibraryLoader.java:96) at io.netty.handler.ssl.OpenSsl.loadTcNative(OpenSsl.java:592) at io.netty.handler.ssl.OpenSsl.(OpenSsl.java:136) at org.apache.cassandra.security.SSLFactory.(SSLFactory.java:100) at org.apache.cassandra.config.EncryptionOptions.(EncryptionOptions.java:242) at org.apache.cassandra.config.Config.(Config.java:284) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:330) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:308) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:301) at org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:167) at org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:331) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObjectNoCheck(BaseConstructor.java:230) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:219) at org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:173) at org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:157) at org.yaml.snakeyaml.Yaml.loadFromReader(Yaml.java:472) at org.yaml.snakeyaml.Yaml.loadAs(Yaml.java:466) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:206) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:134) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:109) at org.apache.cassandra.config.DatabaseDescriptor.loadConfig(DatabaseDescriptor.java:301) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:216) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:187) at io.debezium.connector.cassandra.CassandraConnectorTask$Cassandra4SchemaLoader.load(CassandraConnectorTask.java:26) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:51) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) 18:26:49.805 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - netty_tcnative_linux_x86_64 cannot be loaded from java.library.path, now trying export to -Dio.netty.native.workdir: /tmp java.lang.UnsatisfiedLinkError: no netty_tcnative_linux_x86_64 in java.library.path: [/usr/java/packages/lib, /usr/lib64, /lib64, /lib, /usr/lib] at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2670) at java.base/java.lang.Runtime.loadLibrary0(Runtime.java:830) at java.base/java.lang.System.loadLibrary(System.java:1873) at io.netty.util.internal.NativeLibraryUtil.loadLibrary(NativeLibraryUtil.java:38) at io.netty.util.internal.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:351) at io.netty.util.internal.NativeLibraryLoader.load(NativeLibraryLoader.java:136) at io.netty.util.internal.NativeLibraryLoader.loadFirstAvailable(NativeLibraryLoader.java:96) at io.netty.handler.ssl.OpenSsl.loadTcNative(OpenSsl.java:592) at io.netty.handler.ssl.OpenSsl.(OpenSsl.java:136) at org.apache.cassandra.security.SSLFactory.(SSLFactory.java:100) at org.apache.cassandra.config.EncryptionOptions.(EncryptionOptions.java:242) at org.apache.cassandra.config.Config.(Config.java:284) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:330) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:308) at org.yaml.snakeyaml.constructor.BaseConstructor.newInstance(BaseConstructor.java:301) at org.yaml.snakeyaml.constructor.Constructor$ConstructMapping.construct(Constructor.java:167) at org.yaml.snakeyaml.constructor.Constructor$ConstructYamlObject.construct(Constructor.java:331) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObjectNoCheck(BaseConstructor.java:230) at org.yaml.snakeyaml.constructor.BaseConstructor.constructObject(BaseConstructor.java:219) at org.yaml.snakeyaml.constructor.BaseConstructor.constructDocument(BaseConstructor.java:173) at org.yaml.snakeyaml.constructor.BaseConstructor.getSingleData(BaseConstructor.java:157) at org.yaml.snakeyaml.Yaml.loadFromReader(Yaml.java:472) at org.yaml.snakeyaml.Yaml.loadAs(Yaml.java:466) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:206) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:134) at org.apache.cassandra.config.YamlConfigurationLoader.loadConfig(YamlConfigurationLoader.java:109) at org.apache.cassandra.config.DatabaseDescriptor.loadConfig(DatabaseDescriptor.java:301) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:216) at org.apache.cassandra.config.DatabaseDescriptor.toolInitialization(DatabaseDescriptor.java:187) at io.debezium.connector.cassandra.CassandraConnectorTask$Cassandra4SchemaLoader.load(CassandraConnectorTask.java:26) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:51) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) Suppressed: java.lang.UnsatisfiedLinkError: no netty_tcnative_linux_x86_64 in java.library.path: [/usr/java/packages/lib, /usr/lib64, /lib64, /lib, /usr/lib] at java.base/java.lang.ClassLoader.loadLibrary(ClassLoader.java:2670) at java.base/java.lang.Runtime.loadLibrary0(Runtime.java:830) at java.base/java.lang.System.loadLibrary(System.java:1873) at io.netty.util.internal.NativeLibraryUtil.loadLibrary(NativeLibraryUtil.java:38) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.netty.util.internal.NativeLibraryLoader$1.run(NativeLibraryLoader.java:385) at java.base/java.security.AccessController.doPrivileged(Native Method) at io.netty.util.internal.NativeLibraryLoader.loadLibraryByHelper(NativeLibraryLoader.java:377) at io.netty.util.internal.NativeLibraryLoader.loadLibrary(NativeLibraryLoader.java:341) ... 33 common frames omitted 18:26:49.828 [main] DEBUG io.netty.util.internal.NativeLibraryLoader - Successfully loaded the library /tmp/libnetty_tcnative_linux_x86_641097797716924964190.so 18:26:49.828 [main] DEBUG io.netty.handler.ssl.OpenSsl - Initialize netty-tcnative using engine: 'default' 18:26:49.828 [main] DEBUG io.netty.handler.ssl.OpenSsl - netty-tcnative using native library: BoringSSL 18:26:49.975 [main] DEBUG io.netty.util.ResourceLeakDetector - -Dio.netty.leakDetection.level: simple 18:26:49.975 [main] DEBUG io.netty.util.ResourceLeakDetector - -Dio.netty.leakDetection.targetRecords: 4 18:26:49.982 [main] DEBUG io.netty.buffer.AbstractByteBuf - -Dio.netty.buffer.checkAccessible: true 18:26:49.982 [main] DEBUG io.netty.buffer.AbstractByteBuf - -Dio.netty.buffer.checkBounds: true 18:26:49.983 [main] DEBUG io.netty.util.ResourceLeakDetectorFactory - Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@5c18016b 18:26:50.001 [main] DEBUG io.netty.util.internal.InternalThreadLocalMap - -Dio.netty.threadLocalMap.stringBuilder.initialSize: 1024 18:26:50.001 [main] DEBUG io.netty.util.internal.InternalThreadLocalMap - -Dio.netty.threadLocalMap.stringBuilder.maxSize: 4096 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.numHeapArenas: 16 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.numDirectArenas: 16 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.pageSize: 8192 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxOrder: 11 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.chunkSize: 16777216 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.smallCacheSize: 256 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.normalCacheSize: 64 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxCachedBufferCapacity: 32768 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.cacheTrimInterval: 8192 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.cacheTrimIntervalMillis: 0 18:26:50.003 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.useCacheForAllThreads: true 18:26:50.004 [main] DEBUG io.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxCachedByteBuffersPerChunk: 1023 18:26:50.012 [main] DEBUG io.netty.buffer.ByteBufUtil - -Dio.netty.allocator.type: pooled 18:26:50.012 [main] DEBUG io.netty.buffer.ByteBufUtil - -Dio.netty.threadLocalDirectBufferSize: 0 18:26:50.012 [main] DEBUG io.netty.buffer.ByteBufUtil - -Dio.netty.maxThreadLocalCharBufferSize: 16384 18:26:50.015 [main] DEBUG io.netty.util.ResourceLeakDetectorFactory - Loaded default ResourceLeakDetector: io.netty.util.ResourceLeakDetector@b78a709 18:26:50.020 [main] DEBUG io.netty.util.Recycler - -Dio.netty.recycler.maxCapacityPerThread: 4096 18:26:50.020 [main] DEBUG io.netty.util.Recycler - -Dio.netty.recycler.maxSharedCapacityFactor: 2 18:26:50.020 [main] DEBUG io.netty.util.Recycler - -Dio.netty.recycler.linkCapacity: 16 18:26:50.020 [main] DEBUG io.netty.util.Recycler - -Dio.netty.recycler.ratio: 8 18:26:50.020 [main] DEBUG io.netty.util.Recycler - -Dio.netty.recycler.delayedQueue.ratio: 8 18:26:50.042 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 => ECDHE-ECDSA-AES128-GCM-SHA256 18:26:50.042 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 => ECDHE-ECDSA-AES128-GCM-SHA256 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 => ECDHE-RSA-AES128-GCM-SHA256 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_RSA_WITH_AES_128_GCM_SHA256 => ECDHE-RSA-AES128-GCM-SHA256 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 => ECDHE-ECDSA-AES256-GCM-SHA384 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 => ECDHE-ECDSA-AES256-GCM-SHA384 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 => ECDHE-RSA-AES256-GCM-SHA384 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_RSA_WITH_AES_256_GCM_SHA384 => ECDHE-RSA-AES256-GCM-SHA384 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-ECDSA-CHACHA20-POLY1305 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-ECDSA-CHACHA20-POLY1305 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-RSA-CHACHA20-POLY1305 18:26:50.043 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-RSA-CHACHA20-POLY1305 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_PSK_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-PSK-CHACHA20-POLY1305 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_PSK_WITH_CHACHA20_POLY1305_SHA256 => ECDHE-PSK-CHACHA20-POLY1305 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA => ECDHE-ECDSA-AES128-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_ECDSA_WITH_AES_128_CBC_SHA => ECDHE-ECDSA-AES128-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA => ECDHE-RSA-AES128-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_RSA_WITH_AES_128_CBC_SHA => ECDHE-RSA-AES128-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_PSK_WITH_AES_128_CBC_SHA => ECDHE-PSK-AES128-CBC-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_PSK_WITH_AES_128_CBC_SHA => ECDHE-PSK-AES128-CBC-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA => ECDHE-ECDSA-AES256-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_ECDSA_WITH_AES_256_CBC_SHA => ECDHE-ECDSA-AES256-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA => ECDHE-RSA-AES256-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_RSA_WITH_AES_256_CBC_SHA => ECDHE-RSA-AES256-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_ECDHE_PSK_WITH_AES_256_CBC_SHA => ECDHE-PSK-AES256-CBC-SHA 18:26:50.044 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_ECDHE_PSK_WITH_AES_256_CBC_SHA => ECDHE-PSK-AES256-CBC-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_RSA_WITH_AES_128_GCM_SHA256 => AES128-GCM-SHA256 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_RSA_WITH_AES_128_GCM_SHA256 => AES128-GCM-SHA256 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_RSA_WITH_AES_256_GCM_SHA384 => AES256-GCM-SHA384 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_RSA_WITH_AES_256_GCM_SHA384 => AES256-GCM-SHA384 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_RSA_WITH_AES_128_CBC_SHA => AES128-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_RSA_WITH_AES_128_CBC_SHA => AES128-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_PSK_WITH_AES_128_CBC_SHA => PSK-AES128-CBC-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_PSK_WITH_AES_128_CBC_SHA => PSK-AES128-CBC-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_RSA_WITH_AES_256_CBC_SHA => AES256-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_RSA_WITH_AES_256_CBC_SHA => AES256-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_PSK_WITH_AES_256_CBC_SHA => PSK-AES256-CBC-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_PSK_WITH_AES_256_CBC_SHA => PSK-AES256-CBC-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: TLS_RSA_WITH_3DES_EDE_CBC_SHA => DES-CBC3-SHA 18:26:50.045 [main] DEBUG io.netty.handler.ssl.CipherSuiteConverter - Cipher suite mapping: SSL_RSA_WITH_3DES_EDE_CBC_SHA => DES-CBC3-SHA 18:26:50.046 [main] DEBUG io.netty.handler.ssl.OpenSsl - Supported protocols (OpenSSL): [SSLv2Hello, TLSv1, TLSv1.1, TLSv1.2, TLSv1.3] 18:26:50.046 [main] DEBUG io.netty.handler.ssl.OpenSsl - Default cipher suites (OpenSSL): [TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, TLS_RSA_WITH_AES_128_GCM_SHA256, TLS_RSA_WITH_AES_128_CBC_SHA, TLS_RSA_WITH_AES_256_CBC_SHA, TLS_AES_128_GCM_SHA256, TLS_AES_256_GCM_SHA384, TLS_AES_128_GCM_SHA256, TLS_AES_256_GCM_SHA384] 18:26:50.149 [main] INFO org.apache.cassandra.config.Config - Node configuration:[allocate_tokens_for_keyspace=null; allocate_tokens_for_local_replication_factor=3; allow_extra_insecure_udfs=false; allow_insecure_udfs=false; audit_logging_options=AuditLogOptions{enabled=false, logger='BinAuditLogger', included_keyspaces='', excluded_keyspaces='system,system_schema,system_virtual_schema', included_categories='', excluded_categories='', included_users='', excluded_users='', audit_logs_dir='./audit/', archive_command='', roll_cycle='HOURLY', block=true, max_queue_weight=268435456, max_log_size=17179869184}; authenticator=PasswordAuthenticator; authorizer=CassandraAuthorizer; auto_bootstrap=true; auto_optimise_full_repair_streams=false; auto_optimise_inc_repair_streams=false; auto_optimise_preview_repair_streams=false; auto_snapshot=true; autocompaction_on_startup_enabled=true; automatic_sstable_upgrade=false; back_pressure_enabled=false; back_pressure_strategy=null; batch_size_fail_threshold_in_kb=50; batch_size_warn_threshold_in_kb=5; batchlog_replay_throttle_in_kb=1024; block_for_peers_in_remote_dcs=false; block_for_peers_timeout_in_secs=10; broadcast_address=null; broadcast_rpc_address=10.28.193.59; buffer_pool_use_heap_if_exhausted=false; cache_load_timeout_seconds=30; cas_contention_timeout_in_ms=1000; cdc_enabled=true; cdc_free_space_check_interval_ms=250; cdc_raw_directory=/mnt/resource/cassandra/commitlog/cdc_raw; cdc_total_space_in_mb=0; check_for_duplicate_rows_during_compaction=true; check_for_duplicate_rows_during_reads=true; client_encryption_options=; cluster_name=daq-cassandra-cdc-dev; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=die; commitlog_compression=null; commitlog_directory=/mnt/resource/cassandra/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_group_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=96; concurrent_compactors=4; concurrent_counter_writes=64; concurrent_materialized_view_builders=1; concurrent_materialized_view_writes=32; concurrent_reads=64; concurrent_replicates=null; concurrent_validations=0; concurrent_writes=96; consecutive_message_errors_threshold=1; corrupted_tombstone_strategy=disabled; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=60000; cross_node_timeout=true; data_file_directories=[Ljava.lang.String;@33617539; diagnostic_events_enabled=false; disk_access_mode=mmap_index_only; disk_failure_policy=die; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=false; dynamic_snitch_badness_threshold=1.0; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_drop_compact_storage=false; enable_materialized_views=false; enable_sasi_indexes=false; enable_scripted_user_defined_functions=false; enable_transient_replication=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; endpoint_snitch=SimpleSnitch; file_cache_enabled=false; file_cache_round_up=null; file_cache_size_in_mb=2048; flush_compression=fast; force_new_prepared_statement_behaviour=false; full_query_logging_options=FullQueryLoggerOptions{log_dir='', archive_command='', roll_cycle='HOURLY', block=true, max_queue_weight=268435456, max_log_size=17179869184}; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=/app/cassandra/hints; hints_flush_period_in_ms=10000; ideal_consistency_level=null; incremental_backups=false; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_range_tombstone_list_allocation_size=1; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_application_receive_queue_capacity_in_bytes=4194304; internode_application_receive_queue_reserve_endpoint_capacity_in_bytes=134217728; internode_application_receive_queue_reserve_global_capacity_in_bytes=536870912; internode_application_send_queue_capacity_in_bytes=4194304; internode_application_send_queue_reserve_endpoint_capacity_in_bytes=134217728; internode_application_send_queue_reserve_global_capacity_in_bytes=536870912; internode_authenticator=null; internode_compression=all; internode_max_message_size_in_bytes=null; internode_socket_receive_buffer_size_in_bytes=0; internode_socket_send_buffer_size_in_bytes=0; internode_streaming_tcp_user_timeout_in_ms=300000; internode_tcp_connect_timeout_in_ms=2000; internode_tcp_user_timeout_in_ms=30000; key_cache_keys_to_save=2147483647; key_cache_migrate_during_compaction=true; key_cache_save_period=14400; key_cache_size_in_mb=null; keyspace_count_warn_threshold=40; listen_address=10.28.193.59; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; local_system_data_file_directory=null; max_concurrent_automatic_sstable_upgrades=1; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=offheap_objects; memtable_cleanup_threshold=null; memtable_flush_writers=2; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_allow_older_protocols=true; native_transport_flush_in_batches_legacy=false; native_transport_idle_timeout_in_ms=0; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_concurrent_requests_in_bytes=-1; native_transport_max_concurrent_requests_in_bytes_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_negotiable_protocol_version=null; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; native_transport_receive_queue_capacity_in_bytes=1048576; network_authorizer=AllowAllNetworkAuthorizer; networking_cache_size_in_mb=null; num_tokens=16; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; periodic_commitlog_sync_lag_block_in_ms=null; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=60000; phi_convict_threshold=12.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; range_tombstone_list_growth_factor=1.5; read_request_timeout_in_ms=5000; reject_repair_compaction_threshold=2147483647; repair_command_pool_full_strategy=queue; repair_command_pool_size=0; repair_session_max_tree_depth=null; repair_session_space_in_mb=null; repaired_data_tracking_for_partition_reads_enabled=false; repaired_data_tracking_for_range_reads_enabled=false; report_unconfirmed_repaired_data_mismatches=false; request_timeout_in_ms=10000; role_manager=CassandraRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=60000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=0.0.0.0; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; saved_caches_directory=/mnt/resource/cassandra/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=10.28.193.12,10.28.193.9}; server_encryption_options=; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; snapshot_links_per_second=0; snapshot_on_duplicate_row_detection=false; snapshot_on_repaired_data_mismatch=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; storage_port=7000; stream_entire_sstables=true; stream_throughput_outbound_megabits_per_sec=200; streaming_connections_per_host=1; streaming_keep_alive_period_in_secs=300; table_count_warn_threshold=150; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions@2c177f9e; trickle_fsync=true; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; use_offheap_merkle_trees=true; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; validation_preview_purge_head_start_in_sec=3600; windows_timer_interval=1; write_request_timeout_in_ms=2000] 18:26:50.150 [main] DEBUG org.apache.cassandra.config.DatabaseDescriptor - Syncing log with a period of 10000 18:26:50.150 [main] INFO org.apache.cassandra.config.DatabaseDescriptor - DiskAccessMode is standard, indexAccessMode is mmap 18:26:50.150 [main] INFO org.apache.cassandra.config.DatabaseDescriptor - Global memtable on-heap threshold is enabled at 2048MB 18:26:50.150 [main] INFO org.apache.cassandra.config.DatabaseDescriptor - Global memtable off-heap threshold is enabled at 2048MB 18:26:50.165 [main] INFO org.apache.cassandra.config.DatabaseDescriptor - cdc_enabled is true. Starting casssandra node with Change-Data-Capture enabled. 18:26:50.184 [main] INFO org.apache.cassandra.utils.FBUtilities - InetAddress.getLocalHost() was used to resolve listen_address to cass-1087560719-3-1491280003/10.28.193.59, double check this is correct. Please check your node's config and set the listen_address in cassandra.yaml accordingly if applicable. 18:26:50.629 [main] INFO org.apache.cassandra.cql3.QueryProcessor - Initialized prepared statement caches with 32 MB 18:26:50.756 [main] INFO org.apache.cassandra.utils.MonotonicClock$AbstractEpochSamplingClock - Scheduling approximate time conversion task with an interval of 10000 milliseconds 18:26:50.758 [main] INFO org.apache.cassandra.utils.MonotonicClock$SampledClock - Scheduling approximate time-check task with a precision of 2 milliseconds 18:26:50.765 [main] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy system_schema params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.LocalStrategy}} 18:26:50.827 [main] DEBUG org.apache.cassandra.db.Keyspace - New replication settings for keyspace system_schema - invalidating disk boundary caches 18:26:50.893 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.keyspaces 18:26:50.955 [SSTableBatchOpen:2] INFO org.apache.cassandra.utils.memory.BufferPools - Global buffer pool limit is 2.000GiB for chunk-cache and 128.000MiB for networking 18:26:50.979 [SSTableBatchOpen:3] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/nb-21-big (0.316KiB) 18:26:50.979 [SSTableBatchOpen:1] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/nb-23-big (0.120KiB) 18:26:50.979 [SSTableBatchOpen:2] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/nb-22-big (0.120KiB) 18:26:51.014 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.keyspaces 18:26:51.021 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.022 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.keyspaces 18:26:51.108 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.tables 18:26:51.116 [SSTableBatchOpen:3] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/nb-15-big (2.100KiB) 18:26:51.116 [SSTableBatchOpen:1] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/nb-14-big (2.098KiB) 18:26:51.116 [SSTableBatchOpen:2] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/nb-13-big (3.517KiB) 18:26:51.122 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.tables 18:26:51.122 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.122 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.tables 18:26:51.134 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.columns 18:26:51.138 [SSTableBatchOpen:3] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/nb-14-big (4.948KiB) 18:26:51.138 [SSTableBatchOpen:2] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/nb-13-big (7.727KiB) 18:26:51.138 [SSTableBatchOpen:1] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/nb-15-big (4.948KiB) 18:26:51.141 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.columns 18:26:51.141 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.142 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.columns 18:26:51.149 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.triggers 18:26:51.150 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.triggers 18:26:51.150 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.150 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.triggers 18:26:51.157 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.dropped_columns 18:26:51.160 [SSTableBatchOpen:1] INFO org.apache.cassandra.io.sstable.format.SSTableReaderBuilder - Opening /app/cassandra/data/system_schema/dropped_columns-5e7583b5f3f43af19a39b7e1d6f5f11f/nb-6-big (0.085KiB) 18:26:51.162 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.dropped_columns 18:26:51.162 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.162 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.dropped_columns 18:26:51.167 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.views 18:26:51.167 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.views 18:26:51.168 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.168 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.views 18:26:51.174 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.types 18:26:51.174 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.types 18:26:51.174 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.175 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.types 18:26:51.180 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.functions 18:26:51.180 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.functions 18:26:51.180 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.180 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.functions 18:26:51.185 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.aggregates 18:26:51.185 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.aggregates 18:26:51.185 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.185 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.aggregates 18:26:51.190 [main] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing system_schema.indexes 18:26:51.191 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for system_schema.indexes 18:26:51.191 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:51.191 [main] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for system_schema.indexes 18:26:51.571 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.logging.InternalLoggerFactory - Using SLF4J as the default logging framework 18:26:51.572 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.InternalThreadLocalMap - -Dio.netty.threadLocalMap.stringBuilder.initialSize: 1024 18:26:51.572 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.InternalThreadLocalMap - -Dio.netty.threadLocalMap.stringBuilder.maxSize: 4096 18:26:51.650 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Could not load com.esri.core.geometry.ogc.OGCGeometry with loader null: java.lang.ClassNotFoundException: com.esri.core.geometry.ogc.OGCGeometry java.lang.ClassNotFoundException: com.esri.core.geometry.ogc.OGCGeometry at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522) at java.base/java.lang.Class.forName0(Native Method) at java.base/java.lang.Class.forName(Class.java:315) at com.datastax.oss.driver.internal.core.util.Reflection.loadClass(Reflection.java:57) at com.datastax.oss.driver.internal.core.util.DefaultDependencyChecker.lambda$isPresent$0(DefaultDependencyChecker.java:48) at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1705) at com.datastax.oss.driver.internal.core.util.DefaultDependencyChecker.isPresent(DefaultDependencyChecker.java:41) at com.datastax.dse.driver.internal.core.type.codec.DseTypeCodecsRegistrar.registerDseCodecs(DseTypeCodecsRegistrar.java:32) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.buildCodecRegistry(DefaultDriverContext.java:523) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.(DefaultDriverContext.java:260) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildContext(SessionBuilder.java:968) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildDefaultSessionAsync(SessionBuilder.java:904) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildAsync(SessionBuilder.java:817) at com.datastax.oss.driver.api.core.session.SessionBuilder.build(SessionBuilder.java:835) at io.debezium.connector.cassandra.CassandraClient.(CassandraClient.java:37) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:56) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) 18:26:51.650 [main] DEBUG com.datastax.dse.driver.internal.core.type.codec.DseTypeCodecsRegistrar - ESRI was not found on the classpath: geo codecs will not be available 18:26:51.655 [main] INFO com.datastax.oss.driver.internal.core.DefaultMavenCoordinates - DataStax Java driver for Apache Cassandra(R) (com.datastax.oss:java-driver-core-shaded) version 4.14.0 18:26:51.656 [main] DEBUG com.datastax.oss.driver.internal.core.session.DefaultSession - Creating new session s0 (1 live instances) 18:26:51.663 [main] DEBUG com.datastax.oss.driver.shaded.netty.channel.MultithreadEventLoopGroup - -Dio.netty.eventLoopThreads: 16 18:26:51.685 [main] DEBUG com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoop - -Dio.netty.noKeySetOptimization: false 18:26:51.685 [main] DEBUG com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoop - -Dio.netty.selectorAutoRebuildThreshold: 512 18:26:51.699 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - -Dio.netty.noUnsafe: false 18:26:51.700 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - Java version: 11 18:26:51.700 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - sun.misc.Unsafe.theUnsafe: available 18:26:51.700 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - sun.misc.Unsafe.copyMemory: available 18:26:51.701 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - java.nio.Buffer.address: available 18:26:51.701 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - direct buffer constructor: unavailable java.lang.UnsupportedOperationException: Reflective setAccessible(true) disabled at com.datastax.oss.driver.shaded.netty.util.internal.ReflectionUtil.trySetAccessible(ReflectionUtil.java:31) at com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0$4.run(PlatformDependent0.java:238) at java.base/java.security.AccessController.doPrivileged(Native Method) at com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0.(PlatformDependent0.java:232) at com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent.isAndroid(PlatformDependent.java:294) at com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent.(PlatformDependent.java:93) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoop.newTaskQueue0(NioEventLoop.java:279) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoop.newTaskQueue(NioEventLoop.java:150) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoop.(NioEventLoop.java:138) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoopGroup.newChild(NioEventLoopGroup.java:146) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoopGroup.newChild(NioEventLoopGroup.java:37) at com.datastax.oss.driver.shaded.netty.util.concurrent.MultithreadEventExecutorGroup.(MultithreadEventExecutorGroup.java:84) at com.datastax.oss.driver.shaded.netty.util.concurrent.MultithreadEventExecutorGroup.(MultithreadEventExecutorGroup.java:58) at com.datastax.oss.driver.shaded.netty.util.concurrent.MultithreadEventExecutorGroup.(MultithreadEventExecutorGroup.java:47) at com.datastax.oss.driver.shaded.netty.channel.MultithreadEventLoopGroup.(MultithreadEventLoopGroup.java:59) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoopGroup.(NioEventLoopGroup.java:86) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoopGroup.(NioEventLoopGroup.java:81) at com.datastax.oss.driver.shaded.netty.channel.nio.NioEventLoopGroup.(NioEventLoopGroup.java:68) at com.datastax.oss.driver.internal.core.context.DefaultNettyOptions.(DefaultNettyOptions.java:84) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.buildNettyOptions(DefaultDriverContext.java:472) at com.datastax.oss.driver.internal.core.util.concurrent.LazyReference.get(LazyReference.java:55) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.getNettyOptions(DefaultDriverContext.java:851) at com.datastax.oss.driver.internal.core.session.DefaultSession.(DefaultSession.java:116) at com.datastax.oss.driver.internal.core.session.DefaultSession.init(DefaultSession.java:88) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildDefaultSessionAsync(SessionBuilder.java:903) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildAsync(SessionBuilder.java:817) at com.datastax.oss.driver.api.core.session.SessionBuilder.build(SessionBuilder.java:835) at io.debezium.connector.cassandra.CassandraClient.(CassandraClient.java:37) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:56) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) 18:26:51.702 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - java.nio.Bits.unaligned: available, true 18:26:51.702 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - jdk.internal.misc.Unsafe.allocateUninitializedArray(int): available 18:26:51.702 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent0 - java.nio.DirectByteBuffer.(long, int): unavailable 18:26:51.702 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - sun.misc.Unsafe: available 18:26:51.703 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - maxDirectMemory: 8589934592 bytes (maybe) 18:26:51.703 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - -Dio.netty.tmpdir: /tmp (java.io.tmpdir) 18:26:51.703 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - -Dio.netty.bitMode: 64 (sun.arch.data.model) 18:26:51.704 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - -Dio.netty.maxDirectMemory: -1 bytes 18:26:51.704 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - -Dio.netty.uninitializedArrayAllocationThreshold: 1024 18:26:51.704 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.CleanerJava9 - java.nio.ByteBuffer.cleaner(): available 18:26:51.704 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - -Dio.netty.noPreferDirect: false 18:26:51.712 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.internal.PlatformDependent - org.jctools-core.MpscChunkedArrayQueue: available 18:26:51.720 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetector - -Dcom.datastax.oss.driver.shaded.netty.leakDetection.level: simple 18:26:51.721 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetector - -Dcom.datastax.oss.driver.shaded.netty.leakDetection.targetRecords: 4 18:26:51.722 [main] DEBUG com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetectorFactory - Loaded default ResourceLeakDetector: com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetector@72b2c5ed 18:26:51.735 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a ReconnectionPolicy from config option advanced.reconnection-policy.class 18:26:51.735 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name ExponentialReconnectionPolicy 18:26:51.735 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.connection.ExponentialReconnectionPolicy 18:26:51.747 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@7b8aebd0 for class com.datastax.oss.driver.internal.core.metadata.DistanceEvent 18:26:51.747 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 for class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent 18:26:51.749 [main] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] No server-side filtering 18:26:51.750 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.metadata.MetadataManager$$Lambda$485/0x0000000800614440@da09250 for class com.datastax.oss.driver.internal.core.config.ConfigChangeEvent 18:26:51.751 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@112c2930 for class com.datastax.oss.driver.internal.core.channel.ChannelEvent 18:26:51.752 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@733534f9 for class com.datastax.oss.driver.internal.core.metadata.TopologyEvent 18:26:51.752 [main] WARN com.datastax.oss.driver.internal.core.context.InternalDriverContext - Option METADATA_SCHEMA_CHANGE_LISTENER_CLASS has been deprecated and will be removed in a future release; please use option METADATA_SCHEMA_CHANGE_LISTENER_CLASSES instead. 18:26:51.752 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a SchemaChangeListener from config option advanced.schema-change-listener.class 18:26:51.752 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name NoopSchemaChangeListener 18:26:51.752 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.metadata.schema.NoopSchemaChangeListener 18:26:51.755 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@27bcb4ad for class com.datastax.oss.driver.internal.core.metadata.schema.events.AggregateChangeEvent 18:26:51.756 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@5f3b84bd for class com.datastax.oss.driver.internal.core.metadata.schema.events.FunctionChangeEvent 18:26:51.756 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@5e68be2 for class com.datastax.oss.driver.internal.core.metadata.schema.events.KeyspaceChangeEvent 18:26:51.757 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@45eab322 for class com.datastax.oss.driver.internal.core.metadata.schema.events.TableChangeEvent 18:26:51.757 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@6bd92538 for class com.datastax.oss.driver.internal.core.metadata.schema.events.TypeChangeEvent 18:26:51.758 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@372954e1 for class com.datastax.oss.driver.internal.core.metadata.schema.events.ViewChangeEvent 18:26:51.758 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff for class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent 18:26:51.765 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Could not load org.apache.tinkerpop.gremlin.process.traversal.dsl.graph.GraphTraversal with loader null: java.lang.ClassNotFoundException: org.apache.tinkerpop.gremlin.process.traversal.dsl.graph.GraphTraversal java.lang.ClassNotFoundException: org.apache.tinkerpop.gremlin.process.traversal.dsl.graph.GraphTraversal at java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) at java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522) at java.base/java.lang.Class.forName0(Native Method) at java.base/java.lang.Class.forName(Class.java:315) at com.datastax.oss.driver.internal.core.util.Reflection.loadClass(Reflection.java:57) at com.datastax.oss.driver.internal.core.util.DefaultDependencyChecker.lambda$isPresent$0(DefaultDependencyChecker.java:48) at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1705) at com.datastax.oss.driver.internal.core.util.DefaultDependencyChecker.isPresent(DefaultDependencyChecker.java:41) at com.datastax.oss.driver.internal.core.session.BuiltInRequestProcessors.createDefaultProcessors(BuiltInRequestProcessors.java:47) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.buildRequestProcessorRegistry(DefaultDriverContext.java:512) at com.datastax.oss.driver.internal.core.util.concurrent.LazyReference.get(LazyReference.java:55) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.getRequestProcessorRegistry(DefaultDriverContext.java:905) at com.datastax.oss.driver.internal.core.session.DefaultSession.(DefaultSession.java:121) at com.datastax.oss.driver.internal.core.session.DefaultSession.init(DefaultSession.java:88) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildDefaultSessionAsync(SessionBuilder.java:903) at com.datastax.oss.driver.api.core.session.SessionBuilder.buildAsync(SessionBuilder.java:817) at com.datastax.oss.driver.api.core.session.SessionBuilder.build(SessionBuilder.java:835) at io.debezium.connector.cassandra.CassandraClient.(CassandraClient.java:37) at io.debezium.connector.cassandra.CassandraConnectorContext.(CassandraConnectorContext.java:56) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.run(CassandraConnectorTaskTemplate.java:104) at io.debezium.connector.cassandra.CassandraConnectorTaskTemplate.main(CassandraConnectorTaskTemplate.java:64) at io.debezium.connector.cassandra.CassandraConnectorTask.main(CassandraConnectorTask.java:43) 18:26:51.765 [main] DEBUG com.datastax.oss.driver.internal.core.session.BuiltInRequestProcessors - Tinkerpop was not found on the classpath: graph extensions will not be available 18:26:51.769 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@13004dd8 for class com.datastax.oss.driver.internal.core.metadata.DistanceEvent 18:26:51.769 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a for class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent 18:26:51.770 [main] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@e01a26b for class com.datastax.oss.driver.internal.core.metadata.TopologyEvent 18:26:51.770 [main] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Prepared statements cache configured to use weak values 18:26:51.774 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a MetricsFactory from config option advanced.metrics.factory.class 18:26:51.774 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name DefaultMetricsFactory 18:26:51.774 [main] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.metrics.DefaultMetricsFactory 18:26:51.778 [main] DEBUG com.datastax.oss.driver.internal.core.metrics.DropwizardMetricsFactory - [s0] All metrics are disabled, Session.getMetrics will be empty 18:26:51.778 [main] DEBUG com.datastax.oss.driver.internal.core.metrics.DefaultMetricsFactory - [s0] Using DropwizardMetricsFactory 18:26:51.781 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.DefaultSession - [s0] Starting initialization 18:26:51.789 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a LoadBalancingPolicy from config option basic.load-balancing-policy.class 18:26:51.789 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name DefaultLoadBalancingPolicy 18:26:51.789 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.loadbalancing.DefaultLoadBalancingPolicy 18:26:51.794 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a RetryPolicy from config option advanced.retry-policy.class 18:26:51.794 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name DefaultRetryPolicy 18:26:51.794 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.retry.DefaultRetryPolicy 18:26:51.795 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a SpeculativeExecutionPolicy from config option advanced.speculative-execution-policy.class 18:26:51.795 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name NoSpeculativeExecutionPolicy 18:26:51.795 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.specex.NoSpeculativeExecutionPolicy 18:26:51.795 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a AddressTranslator from config option advanced.address-translator.class 18:26:51.796 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name PassThroughAddressTranslator 18:26:51.796 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.addresstranslation.PassThroughAddressTranslator 18:26:51.796 [s0-admin-0] WARN com.datastax.oss.driver.internal.core.context.InternalDriverContext - Option METADATA_NODE_STATE_LISTENER_CLASS has been deprecated and will be removed in a future release; please use option METADATA_NODE_STATE_LISTENER_CLASSES instead. 18:26:51.796 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a NodeStateListener from config option advanced.node-state-listener.class 18:26:51.796 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name NoopNodeStateListener 18:26:51.796 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.metadata.NoopNodeStateListener 18:26:51.797 [s0-admin-0] WARN com.datastax.oss.driver.internal.core.context.InternalDriverContext - Option REQUEST_TRACKER_CLASS has been deprecated and will be removed in a future release; please use option REQUEST_TRACKER_CLASSES instead. 18:26:51.798 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a RequestTracker from config option advanced.request-tracker.class 18:26:51.798 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name NoopRequestTracker 18:26:51.798 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.tracker.NoopRequestTracker 18:26:51.799 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a RequestThrottler from config option advanced.throttler.class 18:26:51.799 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name PassThroughRequestThrottler 18:26:51.799 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.session.throttling.PassThroughRequestThrottler 18:26:51.799 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a AuthProvider from config option advanced.auth-provider.class 18:26:51.800 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name PlainTextAuthProvider 18:26:51.800 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.auth.PlainTextAuthProvider 18:26:51.801 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a SslEngineFactory from config option advanced.ssl-engine-factory.class 18:26:51.801 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Option is not defined, skipping 18:26:51.801 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a TimestampGenerator from config option advanced.timestamp-generator.class 18:26:51.802 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Building from unqualified name AtomicTimestampGenerator 18:26:51.802 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Trying with default package com.datastax.oss.driver.internal.core.time.AtomicTimestampGenerator 18:26:52.089 [s0-admin-0] INFO com.datastax.oss.driver.internal.core.time.Clock - Using native clock for microsecond precision 18:26:52.091 [s0-admin-0] INFO com.datastax.oss.driver.internal.core.metadata.MetadataManager - [s0] No contact points provided, defaulting to /127.0.0.1:9042 18:26:52.092 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.MetadataManager - [s0] Adding initial contact points [Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)] 18:26:52.093 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.control.ControlConnection - [s0] Initializing with event types [SCHEMA_CHANGE, STATUS_CHANGE, TOPOLOGY_CHANGE] 18:26:52.096 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 for class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent 18:26:52.098 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.control.ControlConnection - [s0] Trying to establish a connection to Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f) 18:26:52.126 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.numHeapArenas: 16 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.numDirectArenas: 16 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.pageSize: 8192 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxOrder: 11 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.chunkSize: 16777216 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.smallCacheSize: 256 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.normalCacheSize: 64 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxCachedBufferCapacity: 32768 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.cacheTrimInterval: 8192 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.cacheTrimIntervalMillis: 0 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.useCacheForAllThreads: true 18:26:52.127 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.PooledByteBufAllocator - -Dio.netty.allocator.maxCachedByteBuffersPerChunk: 1023 18:26:52.135 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.ByteBufUtil - -Dio.netty.allocator.type: pooled 18:26:52.135 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.ByteBufUtil - -Dio.netty.threadLocalDirectBufferSize: 0 18:26:52.136 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.buffer.ByteBufUtil - -Dio.netty.maxThreadLocalCharBufferSize: 16384 18:26:52.145 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.channel.DefaultChannelId - -Dio.netty.processId: 24361 (auto-detected) 18:26:52.146 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.util.NetUtil - -Djava.net.preferIPv4Stack: false 18:26:52.146 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.util.NetUtil - -Djava.net.preferIPv6Addresses: false 18:26:52.148 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.util.NetUtilInitializations - Loopback interface: lo (lo, 0:0:0:0:0:0:0:1%lo) 18:26:52.148 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.util.NetUtil - /proc/sys/net/core/somaxconn: 65000 18:26:52.149 [s0-admin-1] DEBUG com.datastax.oss.driver.shaded.netty.channel.DefaultChannelId - -Dio.netty.machineId: 00:22:48:ff:fe:0a:10:15 (auto-detected) 18:26:52.173 [s0-io-0] WARN com.datastax.oss.driver.internal.core.channel.ChannelFactory - [s0] Invalid value for advanced.connection.max-orphan-requests: 24576. It must be lower than advanced.connection.max-requests-per-connection. Defaulting to 256 (1/4 of max-requests) instead. 18:26:52.219 [s0-io-0] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|connecting...] Starting channel initialization 18:26:52.228 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.Recycler - -Dio.netty.recycler.maxCapacityPerThread: 4096 18:26:52.228 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.Recycler - -Dio.netty.recycler.maxSharedCapacityFactor: 2 18:26:52.228 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.Recycler - -Dio.netty.recycler.linkCapacity: 16 18:26:52.228 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.Recycler - -Dio.netty.recycler.ratio: 8 18:26:52.228 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.Recycler - -Dio.netty.recycler.delayedQueue.ratio: 8 18:26:52.235 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.buffer.AbstractByteBuf - -Dcom.datastax.oss.driver.shaded.netty.buffer.checkAccessible: true 18:26:52.235 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.buffer.AbstractByteBuf - -Dcom.datastax.oss.driver.shaded.netty.buffer.checkBounds: true 18:26:52.235 [s0-io-0] DEBUG com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetectorFactory - Loaded default ResourceLeakDetector: com.datastax.oss.driver.shaded.netty.util.ResourceLeakDetector@5ae48317 18:26:52.262 [s0-io-0] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x255e3dd4, L:/127.0.0.1:65002 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.262 [s0-io-0] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x255e3dd4, L:/127.0.0.1:65002 - R:/127.0.0.1:9042] step OPTIONS received response opcode=ERROR 18:26:52.264 [s0-io-0] DEBUG com.datastax.oss.driver.internal.core.channel.ChannelFactory - [s0] Failed to connect with protocol DSE_V2, retrying with DSE_V1 18:26:52.266 [s0-io-1] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|connecting...] Starting channel initialization 18:26:52.273 [s0-io-1] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x806d6997, L:/127.0.0.1:65004 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.273 [s0-io-1] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x806d6997, L:/127.0.0.1:65004 - R:/127.0.0.1:9042] step OPTIONS received response opcode=ERROR 18:26:52.273 [s0-io-1] DEBUG com.datastax.oss.driver.internal.core.channel.ChannelFactory - [s0] Failed to connect with protocol DSE_V1, retrying with V5 18:26:52.275 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|connecting...] Starting channel initialization 18:26:52.286 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.286 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] step OPTIONS received response opcode=SUPPORTED 18:26:52.290 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.290 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] step STARTUP received response opcode=AUTHENTICATE 18:26:52.407 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.407 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] step AUTH_RESPONSE received response opcode=AUTH_SUCCESS 18:26:52.413 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.413 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] step GET_CLUSTER_NAME received response opcode=RESULT 18:26:52.414 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.414 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] step REGISTER received response opcode=READY 18:26:52.415 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.control.ControlConnection - [s0] New channel opened [id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] 18:26:52.415 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.channel.ChannelEvent: ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.416 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@112c2930 of ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.416 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Processing ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.417 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Transitioning Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f) UNKNOWN=>UP (because a new connection was opened to it) 18:26:52.417 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent: NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.417 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.418 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.419 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.419 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f)) 18:26:52.419 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.DefaultTopologyMonitor - [s0] Refreshing node list 18:26:52.421 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system.local' 18:26:52.423 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system.peers_v2' 18:26:52.423 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system.local', waiting for response 18:26:52.425 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.425 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(1 x 20 columns) 18:26:52.426 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system.peers_v2', waiting for response 18:26:52.427 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.427 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(2 x 12 columns) 18:26:52.434 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.InitialNodeListRefresh - [s0] Copying contact point Node(endPoint=/127.0.0.1:9042, hostId=null, hashCode=1e06063f) 18:26:52.435 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.DefaultTokenFactoryRegistry - [s0] Detected Murmur3 partitioner (org.apache.cassandra.dht.Murmur3Partitioner) 18:26:52.436 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.InitialNodeListRefresh - [s0] Adding new node Node(endPoint=/10.28.193.12:9042, hostId=null, hashCode=49c1b183) 18:26:52.436 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.InitialNodeListRefresh - [s0] Adding new node Node(endPoint=/10.28.193.9:9042, hostId=null, hashCode=4b68db9e) 18:26:52.437 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.DefaultMetadata - [s0] Building initial token map 18:26:52.439 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.DefaultTokenMap - [s0] Rebuilt ring (48 tokens) 18:26:52.446 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.DefaultTokenMap - [s0] Computing keyspace-level data for {} 18:26:52.447 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.DefaultMetadata - [s0] Rebuilding token map took 10 ms 18:26:52.447 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent: NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent: NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.448 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 of NodeStateEvent(null=>UNKNOWN, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.449 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.DefaultProtocolVersionRegistry - [s0] Node /10.28.193.12:9042 reports Cassandra version 4.0.3 18:26:52.449 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.DefaultProtocolVersionRegistry - [s0] Excluding protocol DSE_V1 18:26:52.449 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.DefaultProtocolVersionRegistry - [s0] Excluding protocol DSE_V2 18:26:52.449 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.DefaultProtocolVersionRegistry - [s0] Node /10.28.193.9:9042 reports Cassandra version 4.0.3 18:26:52.449 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.DefaultProtocolVersionRegistry - [s0] Node /127.0.0.1:9042 reports Cassandra version 4.0.3 18:26:52.450 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.concurrent.Debouncer - [s0|metadata debouncer] Received java.util.concurrent.CompletableFuture@f16b9a0[Not completed], scheduling next flush in PT1S 18:26:52.451 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.concurrent.Debouncer - [s0|metadata debouncer] Flushing now 18:26:52.451 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.concurrent.Debouncer - [s0|metadata debouncer] Cancelled existing scheduled flush 18:26:52.451 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.MetadataManager - [s0] Starting schema refresh 18:26:52.453 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.SchemaAgreementChecker - [s0] Checking schema agreement 18:26:52.453 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT schema_version FROM system.local WHERE key='local'' 18:26:52.453 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system.peers' 18:26:52.454 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT schema_version FROM system.local WHERE key='local'', waiting for response 18:26:52.454 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system.peers', waiting for response 18:26:52.455 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.455 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(1 x 1 columns) 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 1, completing and releasing 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(2 x 9 columns) 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.SchemaAgreementChecker - [s0] Peer 32396c06-0823-4b25-a75c-c67794d3f8d1 is down, excluding from schema agreement check 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.SchemaAgreementChecker - [s0] Peer 8447928c-e9ff-4c1a-9ebd-49a045d7059c is down, excluding from schema agreement check 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.SchemaAgreementChecker - [s0] Schema agreement reached (25a9230d-abfb-3c85-89cd-c54ad0f24975), completing 18:26:52.456 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.DefaultSchemaQueriesFactory - [s0] Sending schema queries to Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f) with version 4.0.3 18:26:52.458 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] No server-side filtering 18:26:52.461 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.keyspaces' 18:26:52.462 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.types' 18:26:52.462 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.tables' 18:26:52.463 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.columns' 18:26:52.464 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.indexes' 18:26:52.464 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.views' 18:26:52.465 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.functions' 18:26:52.465 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_schema.aggregates' 18:26:52.466 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_virtual_schema.keyspaces' 18:26:52.466 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_virtual_schema.tables' 18:26:52.467 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Executing query 'SELECT * FROM system_virtual_schema.columns' 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.keyspaces', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.types', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.tables', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.columns', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.indexes', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.views', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.functions', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_schema.aggregates', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_virtual_schema.keyspaces', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_virtual_schema.tables', waiting for response 18:26:52.469 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Successfully wrote query 'SELECT * FROM system_virtual_schema.columns', waiting for response 18:26:52.470 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 8, completing and releasing 18:26:52.470 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(2 x 1 columns) 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 9, completing and releasing 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(21 x 3 columns) 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 1, completing and releasing 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(0 x 4 columns) 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 7, completing and releasing 18:26:52.471 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(0 x 8 columns) 18:26:52.472 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 6, completing and releasing 18:26:52.472 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(0 x 8 columns) 18:26:52.472 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_views' because it matches at least one regex exclude 18:26:52.472 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_virtual_schema' because it matches at least one regex exclude 18:26:52.473 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.473 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(8 x 3 columns) 18:26:52.480 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 10, completing and releasing 18:26:52.480 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(151 x 8 columns) 18:26:52.480 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 4, completing and releasing 18:26:52.480 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(0 x 5 columns) 18:26:52.480 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 5, completing and releasing 18:26:52.481 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(0 x 25 columns) 18:26:52.483 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 2, completing and releasing 18:26:52.484 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(45 x 22 columns) 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|control|id: 0x6eca3b96, L:/127.0.0.1:65006 - R:/127.0.0.1:9042] Got last response on in-flight stream id 3, completing and releasing 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.adminrequest.AdminRequestHandler - [s0] Got response ROWS(310 x 8 columns) 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_auth' because it matches at least one regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering in 'daqcassandra' because it does not match any regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_schema' because it matches at least one regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_distributed' because it matches at least one regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system' because it is excluded by name 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering in 'reaper_db' because it does not match any regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering out 'system_traces' because it matches at least one regex exclude 18:26:52.486 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.RuleBasedKeyspaceFilter - [s0] Filtering in 'daq_cassandra_cdc' because it does not match any regex exclude 18:26:52.491 [s0-io-2] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.queries.CassandraSchemaQueries - [s0] Schema queries took 33 ms 18:26:52.504 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.schema.parsing.CassandraSchemaParser - [s0] Schema parsing took 10 ms 18:26:52.505 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.DefaultMetadata - [s0] Refreshing token map (only schema has changed) 18:26:52.505 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.DefaultTokenMap - [s0] Computing keyspace-level data for {daqcassandra={class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}, reaper_db={class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}, daq_cassandra_cdc={class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.505 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.DefaultTokenMap - [s0] Computing new keyspace-level data for {class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3} 18:26:52.510 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.token.KeyspaceTokenMap - [s0] Computing keyspace-level data for {class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3} took 3 ms 18:26:52.510 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.DefaultMetadata - [s0] Rebuilding token map took 4 ms 18:26:52.510 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.MetadataManager - [s0] Applying schema refresh took 5 ms 18:26:52.510 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Initializing policies 18:26:52.511 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.OptionalLocalDcHelper - [s0|default] Local DC not set, DC awareness will be disabled 18:26:52.511 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.MandatoryLocalDcHelper - [s0|default] Local DC set from implicit contact point Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f): westus 18:26:52.512 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a NodeDistanceEvaluator from config option basic.load-balancing-policy.evaluator.class 18:26:52.512 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Option is not defined, skipping 18:26:52.512 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Creating a Predicate from config option basic.load-balancing-policy.filter.class 18:26:52.512 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.util.Reflection - Option is not defined, skipping 18:26:52.513 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.DefaultNodeDistanceEvaluatorHelper - [s0|default] Evaluator did not assign a distance to node Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) 18:26:52.513 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] com.datastax.oss.driver.internal.core.loadbalancing.DefaultLoadBalancingPolicy@1e14a8c8 suggested Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) to LOCAL, checking what other policies said 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Shortest distance across all policies is LOCAL 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) was IGNORED, changing to LOCAL 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.DistanceEvent: DistanceEvent(LOCAL, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@13004dd8 of DistanceEvent(LOCAL, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@7b8aebd0 of DistanceEvent(LOCAL, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.DefaultNodeDistanceEvaluatorHelper - [s0|default] Evaluator did not assign a distance to node Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) 18:26:52.514 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] com.datastax.oss.driver.internal.core.loadbalancing.DefaultLoadBalancingPolicy@1e14a8c8 suggested Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) to LOCAL, checking what other policies said 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Shortest distance across all policies is LOCAL 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) was IGNORED, changing to LOCAL 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.DistanceEvent: DistanceEvent(LOCAL, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@13004dd8 of DistanceEvent(LOCAL, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@7b8aebd0 of DistanceEvent(LOCAL, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.DefaultNodeDistanceEvaluatorHelper - [s0|default] Evaluator did not assign a distance to node Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] com.datastax.oss.driver.internal.core.loadbalancing.DefaultLoadBalancingPolicy@1e14a8c8 suggested Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f) to LOCAL, checking what other policies said 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Shortest distance across all policies is LOCAL 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper - [s0] Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f) was IGNORED, changing to LOCAL 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.DistanceEvent: DistanceEvent(LOCAL, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@13004dd8 of DistanceEvent(LOCAL, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.515 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@7b8aebd0 of DistanceEvent(LOCAL, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.517 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.config.typesafe.DefaultDriverConfigLoader - [s0] Scheduling periodic reloading with interval PT5M 18:26:52.517 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Starting initialization 18:26:52.517 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Creating a pool for Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) 18:26:52.520 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@50b3b5dc for class com.datastax.oss.driver.internal.core.config.ConfigChangeEvent 18:26:52.521 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.12:9042] Trying to create 1 missing channels 18:26:52.524 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Creating a pool for Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) 18:26:52.524 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@37004fb3 for class com.datastax.oss.driver.internal.core.config.ConfigChangeEvent 18:26:52.524 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|connecting...] Starting channel initialization 18:26:52.524 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.9:9042] Trying to create 1 missing channels 18:26:52.524 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Creating a pool for Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f) 18:26:52.524 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Registering com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@20a45f6e for class com.datastax.oss.driver.internal.core.config.ConfigChangeEvent 18:26:52.524 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/127.0.0.1:9042] Trying to create 1 missing channels 18:26:52.532 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|connecting...] Starting channel initialization 18:26:52.533 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.533 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|connecting...] Starting channel initialization 18:26:52.533 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] step STARTUP received response opcode=AUTHENTICATE 18:26:52.539 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.539 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] step STARTUP received response opcode=AUTHENTICATE 18:26:52.540 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.540 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] step STARTUP received response opcode=AUTHENTICATE 18:26:52.629 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.629 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] step AUTH_RESPONSE received response opcode=AUTH_SUCCESS 18:26:52.631 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.631 [s0-io-3] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] step GET_CLUSTER_NAME received response opcode=RESULT 18:26:52.631 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.12:9042] New channel added [id: 0xef578f0f, L:/10.28.193.59:20364 - R:/10.28.193.12:9042] 18:26:52.632 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.channel.ChannelEvent: ChannelEvent(OPENED, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@112c2930 of ChannelEvent(OPENED, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Processing ChannelEvent(OPENED, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Transitioning Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) UNKNOWN=>UP (because a new connection was opened to it) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent: NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.DefaultNodeDistanceEvaluatorHelper - [s0|default] Evaluator did not assign a distance to node Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183) 18:26:52.633 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.12:9042] Reconnection attempt complete, 1/1 channels 18:26:52.632 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.633 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.12:9042, hostId=32396c06-0823-4b25-a75c-c67794d3f8d1, hashCode=49c1b183)) 18:26:52.645 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.645 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] step AUTH_RESPONSE received response opcode=AUTH_SUCCESS 18:26:52.646 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.646 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] step AUTH_RESPONSE received response opcode=AUTH_SUCCESS 18:26:52.646 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.646 [s0-io-5] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] step GET_CLUSTER_NAME received response opcode=RESULT 18:26:52.646 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/127.0.0.1:9042] New channel added [id: 0x0ce7ae4b, L:/127.0.0.1:65012 - R:/127.0.0.1:9042] 18:26:52.646 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.channel.ChannelEvent: ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.647 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@112c2930 of ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.647 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/127.0.0.1:9042] Reconnection attempt complete, 1/1 channels 18:26:52.647 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Processing ChannelEvent(OPENED, Node(endPoint=/127.0.0.1:9042, hostId=6d76543c-6ce5-4a5b-93bc-22eca9ee9ed3, hashCode=1e06063f)) 18:26:52.648 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.InFlightHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] Got last response on in-flight stream id 0, completing and releasing 18:26:52.648 [s0-io-4] DEBUG com.datastax.oss.driver.internal.core.channel.ProtocolInitHandler - [s0|id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] step GET_CLUSTER_NAME received response opcode=RESULT 18:26:52.648 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.9:9042] New channel added [id: 0xf3e407c1, L:/10.28.193.59:23078 - R:/10.28.193.9:9042] 18:26:52.648 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.channel.ChannelEvent: ChannelEvent(OPENED, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.648 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@112c2930 of ChannelEvent(OPENED, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.648 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Processing ChannelEvent(OPENED, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.648 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.metadata.NodeStateManager - [s0] Transitioning Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) UNKNOWN=>UP (because a new connection was opened to it) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Firing an instance of class com.datastax.oss.driver.internal.core.metadata.NodeStateEvent: NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@173511ff of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.metadata.LoadBalancingPolicyWrapper$$Lambda$531/0x00000008006b2840@2a49ca83 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.loadbalancing.helper.DefaultNodeDistanceEvaluatorHelper - [s0|default] Evaluator did not assign a distance to node Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@65d9e72a of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.context.EventBus - [s0] Notifying com.datastax.oss.driver.internal.core.util.concurrent.RunOrSchedule$$Lambda$480/0x0000000800615040@270d5060 of NodeStateEvent(UNKNOWN=>UP, Node(endPoint=/10.28.193.9:9042, hostId=8447928c-e9ff-4c1a-9ebd-49a045d7059c, hashCode=4b68db9e)) 18:26:52.649 [s0-admin-1] DEBUG com.datastax.oss.driver.internal.core.pool.ChannelPool - [s0|/10.28.193.9:9042] Reconnection attempt complete, 1/1 channels 18:26:52.650 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] All pools have finished initializing 18:26:52.650 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.PoolManager - [s0] Initialization complete, ready 18:26:52.650 [s0-admin-0] DEBUG com.datastax.oss.driver.internal.core.session.DefaultSession - [s0] Initialization complete, ready 18:26:52.784 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Initializing SchemaHolder ... 18:26:52.792 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Added the schema for daqcassandra.chotest to table schema cache. 18:26:52.808 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.809 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.814 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.814 [s0-admin-0] DEBUG org.apache.cassandra.db.Keyspace - New replication settings for keyspace daqcassandra - invalidating disk boundary caches 18:26:52.816 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added keyspace [CREATE KEYSPACE daqcassandra WITH replication = { 'class' : 'org.apache.cassandra.locator.NetworkTopologyStrategy', 'westus': '3' } AND durable_writes = true;] to schema instance. 18:26:52.817 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Updated the schema for daqcassandra.chotest in table schema cache. 18:26:52.817 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Table daqcassandra.chotest detected to be added! 18:26:52.823 [s0-admin-0] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing daqcassandra.chotest 18:26:52.823 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for daqcassandra.chotest 18:26:52.824 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:52.824 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for daqcassandra.chotest 18:26:52.832 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.832 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.832 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.833 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added table [CREATE TABLE daqcassandra.chotest ( name text, choice boolean, date timestamp, address inet, dbl double, lval bigint, ival int, uid timeuuid, value blob, PRIMARY KEY ((name, choice), date, address, dbl, lval, ival, uid) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.1 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.LeveledCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';] to schema instance. 18:26:52.833 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Added the schema for daqcassandra.person to table schema cache. 18:26:52.833 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.833 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.833 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.834 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added keyspace [CREATE KEYSPACE daqcassandra WITH replication = { 'class' : 'org.apache.cassandra.locator.NetworkTopologyStrategy', 'westus': '3' } AND durable_writes = true;] to schema instance. 18:26:52.834 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Updated the schema for daqcassandra.person in table schema cache. 18:26:52.834 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Table daqcassandra.person detected to be added! 18:26:52.837 [s0-admin-0] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing daqcassandra.person 18:26:52.838 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for daqcassandra.person 18:26:52.838 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:52.838 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for daqcassandra.person 18:26:52.842 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.842 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.842 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.842 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added table [CREATE TABLE daqcassandra.person ( id int, name text, phone int, state text, PRIMARY KEY (id) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.01 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';] to schema instance. 18:26:52.842 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Added the schema for daqcassandra.standard1 to table schema cache. 18:26:52.843 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.843 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.843 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.843 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added keyspace [CREATE KEYSPACE daqcassandra WITH replication = { 'class' : 'org.apache.cassandra.locator.NetworkTopologyStrategy', 'westus': '3' } AND durable_writes = true;] to schema instance. 18:26:52.843 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Updated the schema for daqcassandra.standard1 in table schema cache. 18:26:52.843 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Table daqcassandra.standard1 detected to be added! 18:26:52.847 [s0-admin-0] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing daqcassandra.standard1 18:26:52.847 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for daqcassandra.standard1 18:26:52.847 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:52.847 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for daqcassandra.standard1 18:26:52.850 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daqcassandra params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.850 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.851 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daqcassandra 18:26:52.851 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added table [CREATE TABLE daqcassandra.standard1 ( name text, choice boolean, date timestamp, address inet, dbl double, lval bigint, ival int, uid timeuuid, value blob, PRIMARY KEY ((name, choice), date, address, dbl, lval, ival, uid) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.1 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.LeveledCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';] to schema instance. 18:26:52.851 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Added the schema for daq_cassandra_cdc.person to table schema cache. 18:26:52.852 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daq_cassandra_cdc params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.852 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.852 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daq_cassandra_cdc 18:26:52.852 [s0-admin-0] DEBUG org.apache.cassandra.db.Keyspace - New replication settings for keyspace daq_cassandra_cdc - invalidating disk boundary caches 18:26:52.853 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added keyspace [CREATE KEYSPACE daq_cassandra_cdc WITH replication = { 'class' : 'org.apache.cassandra.locator.NetworkTopologyStrategy', 'westus': '3' } AND durable_writes = true;] to schema instance. 18:26:52.853 [s0-admin-0] INFO io.debezium.connector.cassandra.SchemaHolder - Updated the schema for daq_cassandra_cdc.person in table schema cache. 18:26:52.853 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Table daq_cassandra_cdc.person detected to be added! 18:26:52.857 [s0-admin-0] INFO org.apache.cassandra.db.ColumnFamilyStore - Initializing daq_cassandra_cdc.person 18:26:52.858 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Refreshing disk boundary cache for daq_cassandra_cdc.person 18:26:52.858 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Got local ranges [] (ringVersion = 0) 18:26:52.858 [s0-admin-0] DEBUG org.apache.cassandra.db.DiskBoundaryManager - Updating boundaries from null to DiskBoundaries{directories=[DataDirectory{location=/app/cassandra/data}], positions=null, ringVersion=0, directoriesVersion=0} for daq_cassandra_cdc.person 18:26:52.861 [s0-admin-0] INFO org.apache.cassandra.db.Keyspace - Creating replication strategy daq_cassandra_cdc params KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.NetworkTopologyStrategy, westus=3}} 18:26:52.861 [s0-admin-0] INFO org.apache.cassandra.locator.NetworkTopologyStrategy - Configured datacenter replicas are westus:rf(3) 18:26:52.862 [s0-admin-0] WARN org.apache.cassandra.locator.AbstractReplicationStrategy - Ignoring Unrecognized strategy option {westus} passed to NetworkTopologyStrategy for keyspace daq_cassandra_cdc 18:26:52.862 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Added table [CREATE TABLE daq_cassandra_cdc.person ( id int, name text, phone int, state text, PRIMARY KEY (id) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.01 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';] to schema instance. 18:26:52.864 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - CDC enabled entities: [CREATE TABLE daqcassandra.person ( id int, name text, phone int, state text, PRIMARY KEY (id) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.01 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';, CREATE TABLE daqcassandra.standard1 ( name text, choice boolean, date timestamp, address inet, dbl double, lval bigint, ival int, uid timeuuid, value blob, PRIMARY KEY ((name, choice), date, address, dbl, lval, ival, uid) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.1 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.LeveledCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';, CREATE TABLE daqcassandra.chotest ( name text, choice boolean, date timestamp, address inet, dbl double, lval bigint, ival int, uid timeuuid, value blob, PRIMARY KEY ((name, choice), date, address, dbl, lval, ival, uid) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.1 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.LeveledCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';, CREATE TABLE daq_cassandra_cdc.person ( id int, name text, phone int, state text, PRIMARY KEY (id) ) WITH additional_write_policy = '99p' AND bloom_filter_fp_chance = 0.01 AND caching = {'keys':'ALL','rows_per_partition':'NONE'} AND cdc = true AND comment = '' AND compaction = {'class':'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy','max_threshold':'32','min_threshold':'4'} AND compression = {'chunk_length_in_kb':'16','class':'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND default_time_to_live = 0 AND extensions = {} AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair = 'BLOCKING' AND speculative_retry = '99p';] 18:26:52.864 [s0-admin-0] INFO io.debezium.connector.cassandra.Cassandra4SchemaChangeListener - Initialized SchemaHolder. 18:26:52.888 [main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: acks = -1 batch.size = 16384 bootstrap.servers = [kafka-1475864931-1-1563145828.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-2-1563145831.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-4-1563145837.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-5-1563145840.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-6-1563145843.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-7-1563145846.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-8-1563145849.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-9-1563145852.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-10-1563145855.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-11-1563145858.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-12-1563145861.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-13-1563145864.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-14-1563145867.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093, kafka-1475864931-15-1563145870.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093] buffer.memory = 33554432 client.dns.lookup = use_all_dns_ips client.id = producer-1 compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 120000 enable.idempotence = true interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 30000 retries = 2147483647 retry.backoff.ms = 100 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.connect.timeout.ms = null sasl.login.read.timeout.ms = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.login.retry.backoff.max.ms = 10000 sasl.login.retry.backoff.ms = 100 sasl.mechanism = GSSAPI sasl.oauthbearer.clock.skew.seconds = 30 sasl.oauthbearer.expected.audience = null sasl.oauthbearer.expected.issuer = null sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 sasl.oauthbearer.jwks.endpoint.url = null sasl.oauthbearer.scope.claim.name = scope sasl.oauthbearer.sub.claim.name = sub sasl.oauthbearer.token.endpoint.url = null security.protocol = SSL security.providers = null send.buffer.bytes = 131072 socket.connection.setup.timeout.max.ms = 30000 socket.connection.setup.timeout.ms = 10000 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.3] ssl.endpoint.identification.algorithm = https ssl.engine.factory.class = null ssl.key.password = [hidden] ssl.keymanager.algorithm = SunX509 ssl.keystore.certificate.chain = null ssl.keystore.key = null ssl.keystore.location = /app/ssl/DataAcquisition.kafka-clients.walmart.net_keyStore.jks ssl.keystore.password = [hidden] ssl.keystore.type = JKS ssl.protocol = TLSv1.3 ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.certificates = null ssl.truststore.location = /app/ssl/DataAcquisition.kafka-clients.walmart.net_trustStore.jks ssl.truststore.password = [hidden] ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer 18:26:53.383 [main] DEBUG org.apache.kafka.common.security.ssl.DefaultSslEngineFactory - Created SSL context with keystore SecurityStore(path=/app/ssl/DataAcquisition.kafka-clients.walmart.net_keyStore.jks, modificationTime=Fri Jun 10 22:46:30 UTC 2022), truststore SecurityStore(path=/app/ssl/DataAcquisition.kafka-clients.walmart.net_trustStore.jks, modificationTime=Fri Jun 10 22:46:30 UTC 2022), provider SunJSSE. 18:26:53.417 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.producer.internals.Sender - [Producer clientId=producer-1] Starting Kafka producer I/O thread. 18:26:53.418 [main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.1.0 18:26:53.418 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Initialize connection to node kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: -3 rack: null) for sending metadata request 18:26:53.418 [main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 37edeed0777bacb3 18:26:53.418 [main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1657218413417 18:26:53.419 [main] DEBUG org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=producer-1] Kafka producer started 18:26:53.420 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.ClientUtils - Resolved host kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net as 10.103.24.232 18:26:53.420 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Initiating connection to node kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: -3 rack: null) using address kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net/10.103.24.232 18:26:53.421 [main] INFO io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - Starting processor group ... 18:26:53.438 [main] INFO org.apache.kafka.connect.json.JsonConverterConfig - JsonConverterConfig values: converter.type = key decimal.format = BASE64 schemas.cache.size = 1000 schemas.enable = true 18:26:53.440 [main] INFO org.apache.kafka.connect.json.JsonConverterConfig - JsonConverterConfig values: converter.type = value decimal.format = BASE64 schemas.cache.size = 1000 schemas.enable = true 18:26:53.445 [main] INFO io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - Initialized Processor Group. 18:26:53.446 [pool-4-thread-2] INFO io.debezium.connector.cassandra.AbstractProcessor - Started Commit Log Post-Processor 18:26:53.447 [pool-4-thread-3] INFO io.debezium.connector.cassandra.AbstractProcessor - Started Queue Processor [0] 18:26:53.447 [main] INFO io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - Starting HTTP server ... 18:26:53.447 [main] DEBUG io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - HTTP port is 8000 18:26:53.448 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - polling records... 18:26:53.448 [pool-4-thread-4] INFO io.debezium.connector.cassandra.AbstractProcessor - Started Commit Log Processor 18:26:53.448 [pool-4-thread-4] DEBUG io.debezium.connector.cassandra.Cassandra4CommitLogProcessor - Processing commitLogFiles while initial is true 18:26:53.449 [pool-4-thread-1] INFO io.debezium.connector.cassandra.AbstractProcessor - Started Snapshot Processor 18:26:53.449 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.network.Selector - [Producer clientId=producer-1] Created socket with SO_RCVBUF = 32768, SO_SNDBUF = 131072, SO_TIMEOUT = 0 to node -3 18:26:53.449 [pool-4-thread-1] DEBUG io.debezium.connector.cassandra.SnapshotProcessor - Skipping snapshot [mode: NEVER] 18:26:53.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:53.454 [pool-4-thread-4] INFO io.debezium.connector.cassandra.Cassandra4CommitLogProcessor - Reading existing commit logs in /mnt/resource/cassandra/commitlog/cdc_raw 18:26:53.458 [main] DEBUG org.eclipse.jetty.util.log - Logging to Logger[org.eclipse.jetty.util.log] via org.eclipse.jetty.util.log.Slf4jLog 18:26:53.459 [main] INFO org.eclipse.jetty.util.log - Logging initialized @4544ms to org.eclipse.jetty.util.log.Slf4jLog 18:26:53.461 [pool-2-thread-1] INFO io.debezium.connector.cassandra.Cassandra4CommitLogProcessor$CommitLogProcessingCallable - Processing commit log /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log 18:26:53.462 [pool-4-thread-4] INFO io.debezium.connector.cassandra.AbstractDirectoryWatcher - Polling commitLog files from /mnt/resource/cassandra/commitlog/cdc_raw ... 18:26:53.462 [pool-2-thread-1] INFO io.debezium.connector.cassandra.Cassandra4CommitLogProcessor$CommitLogProcessingCallable - LogicalCommitLog{commitLogPosition=CommitLogPosition(segmentId=1652287118530, position=0), synced=33552621, completed=true, log=/mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log, index=/mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530_cdc.idx, commitLogSegmentId=1652287118530} 18:26:53.468 [pool-2-thread-1] DEBUG org.apache.cassandra.db.commitlog.CommitLogReader - Reading /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log (CL version 7, messaging version 12, compression null) 18:26:53.468 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - Server@5f935d49{STOPPED}[9.4.z-SNAPSHOT] added {QueuedThreadPool[qtp2144817630]@7fd751de{STOPPED,8<=0<=200,i=0,q=0}[org.eclipse.jetty.util.thread.TryExecutor$$Lambda$751/0x000000080080bc40@4597e6e3],AUTO} 18:26:53.485 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - HttpConnectionFactory@374b6e33[HTTP/1.1] added {HttpConfiguration@7ed7ae{32768/8192,8192/8192,https://:0,[]},POJO} 18:26:53.487 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{null,[]}{0.0.0.0:0} added {Server@5f935d49{STOPPED}[9.4.z-SNAPSHOT],UNMANAGED} 18:26:53.487 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{null,[]}{0.0.0.0:0} added {QueuedThreadPool[qtp2144817630]@7fd751de{STOPPED,8<=0<=200,i=0,q=0}[org.eclipse.jetty.util.thread.TryExecutor$$Lambda$751/0x000000080080bc40@4597e6e3],AUTO} 18:26:53.487 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{null,[]}{0.0.0.0:0} added {ScheduledExecutorScheduler@2b395581{STOPPED},AUTO} 18:26:53.487 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{null,[]}{0.0.0.0:0} added {org.eclipse.jetty.io.ArrayByteBufferPool@266da047,POJO} 18:26:53.487 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{null,[http/1.1]}{0.0.0.0:0} added {HttpConnectionFactory@374b6e33[HTTP/1.1],AUTO} 18:26:53.487 [main] DEBUG org.eclipse.jetty.server.AbstractConnector - ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:0} added HttpConnectionFactory@374b6e33[HTTP/1.1] 18:26:53.489 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:0} added {SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:0},MANAGED} 18:26:53.489 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - Server@5f935d49{STOPPED}[9.4.z-SNAPSHOT] added {ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000},AUTO} 18:26:53.505 [main] DEBUG org.eclipse.jetty.util.DecoratedObjectFactory - Adding Decorator: org.eclipse.jetty.util.DeprecationWarning@34549979 18:26:53.510 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - o.e.j.s.ServletContextHandler@2e1add6f{/,null,UNAVAILABLE} added {org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1,MANAGED} 18:26:53.514 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 added {ServletHandler@30eedaa4{STOPPED},MANAGED} 18:26:53.515 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - Server@5f935d49{STOPPED}[9.4.z-SNAPSHOT] added {o.e.j.s.ServletContextHandler@2e1add6f{/,null,UNAVAILABLE},MANAGED} 18:26:53.522 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true,AUTO} 18:26:53.523 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {[/ping]=>com.codahale.metrics.servlets.PingServlet-3193e21d,POJO} 18:26:53.523 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true,AUTO} 18:26:53.523 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {[/buildinfo]=>io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db,POJO} 18:26:53.524 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true,AUTO} 18:26:53.525 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {[/metrics]=>com.codahale.metrics.servlets.MetricsServlet-b1d19ff,POJO} 18:26:53.527 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true,AUTO} 18:26:53.527 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STOPPED} added {[/health]=>com.codahale.metrics.servlets.HealthCheckServlet-5d221b20,POJO} 18:26:53.527 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting Server@5f935d49{STOPPED}[9.4.z-SNAPSHOT] 18:26:53.528 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - Server@5f935d49{STARTING}[9.4.z-SNAPSHOT] added {ErrorHandler@70f4abc1{STOPPED},AUTO} 18:26:53.529 [main] INFO org.eclipse.jetty.server.Server - jetty-9.4.z-SNAPSHOT; built: 2018-08-30T13:59:14.071Z; git: 27208684755d94a92186989f695db2d7b21ebc51; jvm 11.0.12+7-LTS 18:26:53.539 [main] DEBUG org.eclipse.jetty.server.handler.AbstractHandler - starting Server@5f935d49{STARTING}[9.4.z-SNAPSHOT] 18:26:53.539 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting QueuedThreadPool[qtp2144817630]@7fd751de{STOPPED,8<=0<=200,i=0,q=0}[org.eclipse.jetty.util.thread.TryExecutor$$Lambda$751/0x000000080080bc40@4597e6e3] 18:26:53.540 [main] DEBUG org.eclipse.jetty.util.thread.ReservedThreadExecutor - ReservedThreadExecutor@67fac095{s=0/8,p=0} 18:26:53.540 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - QueuedThreadPool[qtp2144817630]@7fd751de{STARTING,8<=0<=200,i=0,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}] added {ReservedThreadExecutor@67fac095{s=0/8,p=0},AUTO} 18:26:53.540 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ReservedThreadExecutor@67fac095{s=0/8,p=0} 18:26:53.542 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4627ms ReservedThreadExecutor@67fac095{s=0/8,p=0} 18:26:53.551 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4636ms QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=7,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}] 18:26:53.551 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting o.e.j.s.ServletContextHandler@2e1add6f{/,null,UNAVAILABLE} 18:26:53.555 [main] DEBUG org.eclipse.jetty.http.PreEncodedHttpField - HttpField encoders loaded: [org.eclipse.jetty.http.Http1FieldPreEncoder@481ed95c] 18:26:53.559 [main] DEBUG org.eclipse.jetty.server.handler.AbstractHandler - starting o.e.j.s.ServletContextHandler@2e1add6f{/,null,STARTING} 18:26:53.559 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 18:26:53.562 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 added {org.eclipse.jetty.server.session.DefaultSessionCache@3b362f1[evict=-1,removeUnloadable=false,saveOnCreate=false,saveOnInactiveEvict=false],AUTO} 18:26:53.563 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - org.eclipse.jetty.server.session.DefaultSessionCache@3b362f1[evict=-1,removeUnloadable=false,saveOnCreate=false,saveOnInactiveEvict=false] added {org.eclipse.jetty.server.session.NullSessionDataStore@42ecc554[passivating=false,graceSec=3600],AUTO} 18:26:53.564 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting DefaultSessionIdManager@4a04ca74{STOPPED}[worker=null] 18:26:53.564 [main] INFO org.eclipse.jetty.server.session - DefaultSessionIdManager workerName=node0 18:26:53.564 [main] INFO org.eclipse.jetty.server.session - No SessionScavenger set, using defaults 18:26:53.565 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - DefaultSessionIdManager@4a04ca74{STARTING}[worker=node0] added {HouseKeeper@1850f2da{STOPPED}[interval=600000, ownscheduler=false],MANAGED} 18:26:53.565 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting HouseKeeper@1850f2da{STOPPED}[interval=600000, ownscheduler=false] 18:26:53.565 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ScheduledExecutorScheduler@38d42ab7{STOPPED} 18:26:53.565 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4650ms ScheduledExecutorScheduler@38d42ab7{STARTED} 18:26:53.565 [main] DEBUG org.eclipse.jetty.server.session - Using own scheduler for scavenging 18:26:53.565 [main] INFO org.eclipse.jetty.server.session - node0 Scavenging every 600000ms 18:26:53.567 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4652ms HouseKeeper@1850f2da{STARTED}[interval=600000, ownscheduler=true] 18:26:53.567 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4652ms DefaultSessionIdManager@4a04ca74{STARTED}[worker=node0] 18:26:53.567 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - Server@5f935d49{STARTING}[9.4.z-SNAPSHOT] added {DefaultSessionIdManager@4a04ca74{STARTED}[worker=node0],MANAGED} 18:26:53.568 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 added {DefaultSessionIdManager@4a04ca74{STARTED}[worker=node0],UNMANAGED} 18:26:53.568 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ScheduledExecutorScheduler@2e8b24a1{STOPPED} 18:26:53.568 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4652ms ScheduledExecutorScheduler@2e8b24a1{STARTED} 18:26:53.568 [main] DEBUG org.eclipse.jetty.server.handler.AbstractHandler - starting org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 18:26:53.568 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ServletHandler@30eedaa4{STOPPED} 18:26:53.570 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/health[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20[EMBEDDED:null] 18:26:53.571 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/health",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true] to PathMappings[size=1] 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/buildinfo[EMBEDDED:null] mapped to servlet=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db[EMBEDDED:null] 18:26:53.571 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/buildinfo",pathDepth=1,group=EXACT],resource=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true] to PathMappings[size=2] 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/ping[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.PingServlet-3193e21d[EMBEDDED:null] 18:26:53.571 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/ping",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true] to PathMappings[size=3] 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/metrics[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.MetricsServlet-b1d19ff[EMBEDDED:null] 18:26:53.571 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/metrics",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true] to PathMappings[size=4] 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - filterNameMap={} 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - pathFilters=null 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletFilterMap=null 18:26:53.571 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletPathMap=PathMappings[size=4] 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletNameMap={com.codahale.metrics.servlets.MetricsServlet-b1d19ff=com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true, io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true, com.codahale.metrics.servlets.PingServlet-3193e21d=com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true, com.codahale.metrics.servlets.HealthCheckServlet-5d221b20=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true} 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Adding Default404Servlet to ServletHandler@30eedaa4{STARTING} 18:26:53.572 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STARTING} added {org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a@305b0b17==org.eclipse.jetty.servlet.ServletHandler$Default404Servlet,jsp=null,order=-1,inst=false,AUTO} 18:26:53.572 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServletHandler@30eedaa4{STARTING} added {[/]=>org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a,POJO} 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/health[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20[EMBEDDED:null] 18:26:53.572 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/health",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true] to PathMappings[size=1] 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/buildinfo[EMBEDDED:null] mapped to servlet=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db[EMBEDDED:null] 18:26:53.572 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/buildinfo",pathDepth=1,group=EXACT],resource=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true] to PathMappings[size=2] 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/ping[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.PingServlet-3193e21d[EMBEDDED:null] 18:26:53.572 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/ping",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true] to PathMappings[size=3] 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/metrics[EMBEDDED:null] mapped to servlet=com.codahale.metrics.servlets.MetricsServlet-b1d19ff[EMBEDDED:null] 18:26:53.572 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/metrics",pathDepth=1,group=EXACT],resource=com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true] to PathMappings[size=4] 18:26:53.572 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - Path=/[EMBEDDED:null] mapped to servlet=org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a[EMBEDDED:null] 18:26:53.572 [main] DEBUG org.eclipse.jetty.http.pathmap.PathMappings - Added MappedResource[pathSpec=ServletPathSpec["/",pathDepth=-1,group=DEFAULT],resource=org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a@305b0b17==org.eclipse.jetty.servlet.ServletHandler$Default404Servlet,jsp=null,order=-1,inst=false] to PathMappings[size=5] 18:26:53.573 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - filterNameMap={} 18:26:53.573 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - pathFilters=null 18:26:53.573 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletFilterMap=null 18:26:53.573 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletPathMap=PathMappings[size=5] 18:26:53.573 [main] DEBUG org.eclipse.jetty.servlet.ServletHandler - servletNameMap={com.codahale.metrics.servlets.MetricsServlet-b1d19ff=com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true, org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a=org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a@305b0b17==org.eclipse.jetty.servlet.ServletHandler$Default404Servlet,jsp=null,order=-1,inst=false, io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db=io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true, com.codahale.metrics.servlets.PingServlet-3193e21d=com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true, com.codahale.metrics.servlets.HealthCheckServlet-5d221b20=com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true} 18:26:53.573 [main] DEBUG org.eclipse.jetty.server.handler.AbstractHandler - starting ServletHandler@30eedaa4{STARTING} 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4658ms ServletHandler@30eedaa4{STARTED} 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting org.eclipse.jetty.server.session.DefaultSessionCache@3b362f1[evict=-1,removeUnloadable=false,saveOnCreate=false,saveOnInactiveEvict=false] 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting org.eclipse.jetty.server.session.NullSessionDataStore@42ecc554[passivating=false,graceSec=3600] 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4658ms org.eclipse.jetty.server.session.NullSessionDataStore@42ecc554[passivating=false,graceSec=3600] 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4658ms org.eclipse.jetty.server.session.DefaultSessionCache@3b362f1[evict=-1,removeUnloadable=false,saveOnCreate=false,saveOnInactiveEvict=false] 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4658ms org.eclipse.jetty.server.session.SessionHandler1182463628==dftMaxIdleSec=-1 18:26:53.573 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true 18:26:53.574 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4658ms com.codahale.metrics.servlets.HealthCheckServlet-5d221b20@2b0dd0a3==com.codahale.metrics.servlets.HealthCheckServlet,jsp=null,order=-1,inst=true 18:26:53.575 [main] DEBUG org.eclipse.jetty.servlet.ServletHolder - Servlet.init com.codahale.metrics.servlets.HealthCheckServlet@20fcea37 for com.codahale.metrics.servlets.HealthCheckServlet-5d221b20 18:26:53.579 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true 18:26:53.579 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4664ms com.codahale.metrics.servlets.MetricsServlet-b1d19ff@b4351f61==com.codahale.metrics.servlets.MetricsServlet,jsp=null,order=-1,inst=true 18:26:53.579 [main] DEBUG org.eclipse.jetty.servlet.ServletHolder - Servlet.init com.codahale.metrics.servlets.MetricsServlet@68a426c3 for com.codahale.metrics.servlets.MetricsServlet-b1d19ff 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4666ms com.codahale.metrics.servlets.PingServlet-3193e21d@988f0ee1==com.codahale.metrics.servlets.PingServlet,jsp=null,order=-1,inst=true 18:26:53.582 [main] DEBUG org.eclipse.jetty.servlet.ServletHolder - Servlet.init com.codahale.metrics.servlets.PingServlet@756c67cd for com.codahale.metrics.servlets.PingServlet-3193e21d 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4666ms io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db@68b74e1a==io.debezium.connector.cassandra.network.BuildInfoServlet,jsp=null,order=-1,inst=true 18:26:53.582 [main] DEBUG org.eclipse.jetty.servlet.ServletHolder - Servlet.init io.debezium.connector.cassandra.network.BuildInfoServlet@50de907a for io.debezium.connector.cassandra.network.BuildInfoServlet-e07b4db 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a@305b0b17==org.eclipse.jetty.servlet.ServletHandler$Default404Servlet,jsp=null,order=-1,inst=false 18:26:53.582 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4667ms org.eclipse.jetty.servlet.ServletHandler$Default404Servlet-2eb6d34a@305b0b17==org.eclipse.jetty.servlet.ServletHandler$Default404Servlet,jsp=null,order=-1,inst=false 18:26:53.582 [main] INFO org.eclipse.jetty.server.handler.ContextHandler - Started o.e.j.s.ServletContextHandler@2e1add6f{/,null,AVAILABLE} 18:26:53.583 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4667ms o.e.j.s.ServletContextHandler@2e1add6f{/,null,AVAILABLE} 18:26:53.583 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ErrorHandler@70f4abc1{STOPPED} 18:26:53.583 [main] DEBUG org.eclipse.jetty.server.handler.AbstractHandler - starting ErrorHandler@70f4abc1{STARTING} 18:26:53.583 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4667ms ErrorHandler@70f4abc1{STARTED} 18:26:53.583 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} 18:26:53.584 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {sun.nio.ch.ServerSocketChannelImpl[/0:0:0:0:0:0:0:0:8000],POJO} 18:26:53.585 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ScheduledExecutorScheduler@2b395581{STOPPED} 18:26:53.585 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4669ms ScheduledExecutorScheduler@2b395581{STARTED} 18:26:53.585 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting HttpConnectionFactory@374b6e33[HTTP/1.1] 18:26:53.585 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4670ms HttpConnectionFactory@374b6e33[HTTP/1.1] 18:26:53.585 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} 18:26:53.590 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.590071Z added {SelectorProducer@3f322610,POJO} 18:26:53.591 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.59088Z added {QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}],UNMANAGED} 18:26:53.591 [main] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.591169Z created 18:26:53.591 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ManagedSelector@135e49b2{STOPPED} id=0 keys=-1 selected=-1 updates=0 added {EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.591455Z,MANAGED} 18:26:53.591 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {ManagedSelector@135e49b2{STOPPED} id=0 keys=-1 selected=-1 updates=0,AUTO} 18:26:53.591 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.591902Z added {SelectorProducer@10b87ff6,POJO} 18:26:53.592 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.592098Z added {QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}],UNMANAGED} 18:26:53.592 [main] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.592328Z created 18:26:53.592 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ManagedSelector@63262071{STOPPED} id=1 keys=-1 selected=-1 updates=0 added {EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.592524Z,MANAGED} 18:26:53.592 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {ManagedSelector@63262071{STOPPED} id=1 keys=-1 selected=-1 updates=0,AUTO} 18:26:53.592 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.592807Z added {SelectorProducer@1282f784,POJO} 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.592976Z added {QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}],UNMANAGED} 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.593203Z created 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ManagedSelector@41167ded{STOPPED} id=2 keys=-1 selected=-1 updates=0 added {EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.593397Z,MANAGED} 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {ManagedSelector@41167ded{STOPPED} id=2 keys=-1 selected=-1 updates=0,AUTO} 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.593678Z added {SelectorProducer@36238b12,POJO} 18:26:53.593 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.593836Z added {QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}],UNMANAGED} 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.594037Z created 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ManagedSelector@33187485{STOPPED} id=3 keys=-1 selected=-1 updates=0 added {EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.594242Z,MANAGED} 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {ManagedSelector@33187485{STOPPED} id=3 keys=-1 selected=-1 updates=0,AUTO} 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ManagedSelector@135e49b2{STOPPED} id=0 keys=-1 selected=-1 updates=0 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.594555Z 18:26:53.594 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4679ms EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=8,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.594716Z 18:26:53.595 [main] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - queue org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@77ff14ce 18:26:53.595 [qtp2144817630-56] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - run org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@77ff14ce 18:26:53.595 [main] DEBUG org.eclipse.jetty.io.ManagedSelector - Queued change org.eclipse.jetty.io.ManagedSelector$Start@74764622 on ManagedSelector@135e49b2{STARTING} id=0 keys=0 selected=0 updates=0 18:26:53.596 [qtp2144817630-56] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@2bfc8558/SelectorProducer@3f322610/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=7,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.595912Z tryProduce false 18:26:53.597 [qtp2144817630-56] DEBUG org.eclipse.jetty.io.ManagedSelector - updateable 1 18:26:53.597 [qtp2144817630-56] DEBUG org.eclipse.jetty.io.ManagedSelector - update org.eclipse.jetty.io.ManagedSelector$Start@74764622 18:26:53.597 [qtp2144817630-56] DEBUG org.eclipse.jetty.io.ManagedSelector - updates 0 18:26:53.597 [qtp2144817630-56] DEBUG org.eclipse.jetty.io.ManagedSelector - Selector sun.nio.ch.EPollSelectorImpl@73120b00 waiting with 0 keys 18:26:53.597 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4682ms ManagedSelector@135e49b2{STARTED} id=0 keys=0 selected=0 updates=0 18:26:53.597 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ManagedSelector@63262071{STOPPED} id=1 keys=-1 selected=-1 updates=0 18:26:53.597 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=7,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.597665Z 18:26:53.597 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4682ms EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=7,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.597836Z 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - queue org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@7ba06506 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - run org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@7ba06506 18:26:53.598 [main] DEBUG org.eclipse.jetty.io.ManagedSelector - Queued change org.eclipse.jetty.io.ManagedSelector$Start@63dc3420 on ManagedSelector@63262071{STARTING} id=1 keys=0 selected=0 updates=0 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@5bb97fe7/SelectorProducer@10b87ff6/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=6,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.598276Z tryProduce false 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.io.ManagedSelector - updateable 1 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.io.ManagedSelector - update org.eclipse.jetty.io.ManagedSelector$Start@63dc3420 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.io.ManagedSelector - updates 0 18:26:53.598 [qtp2144817630-57] DEBUG org.eclipse.jetty.io.ManagedSelector - Selector sun.nio.ch.EPollSelectorImpl@31642a3d waiting with 0 keys 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4683ms ManagedSelector@63262071{STARTED} id=1 keys=0 selected=0 updates=0 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ManagedSelector@41167ded{STOPPED} id=2 keys=-1 selected=-1 updates=0 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=6,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.598655Z 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4683ms EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=6,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.598813Z 18:26:53.598 [main] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - queue org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@525b8922 18:26:53.599 [main] DEBUG org.eclipse.jetty.io.ManagedSelector - Queued change org.eclipse.jetty.io.ManagedSelector$Start@145a821d on ManagedSelector@41167ded{STARTING} id=2 keys=0 selected=0 updates=0 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - run org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@525b8922 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@2d4f67e/SelectorProducer@1282f784/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=5,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.599707Z tryProduce false 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.io.ManagedSelector - updateable 1 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.io.ManagedSelector - update org.eclipse.jetty.io.ManagedSelector$Start@145a821d 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.io.ManagedSelector - updates 0 18:26:53.599 [qtp2144817630-58] DEBUG org.eclipse.jetty.io.ManagedSelector - Selector sun.nio.ch.EPollSelectorImpl@4f943367 waiting with 0 keys 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4684ms ManagedSelector@41167ded{STARTED} id=2 keys=0 selected=0 updates=0 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting ManagedSelector@33187485{STOPPED} id=3 keys=-1 selected=-1 updates=0 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - starting EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=5,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.600192Z 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4684ms EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=5,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.600367Z 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - queue org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@6caa4dc5 18:26:53.600 [main] DEBUG org.eclipse.jetty.io.ManagedSelector - Queued change org.eclipse.jetty.io.ManagedSelector$Start@6ed51944 on ManagedSelector@33187485{STARTING} id=3 keys=0 selected=0 updates=0 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - run org.eclipse.jetty.io.ManagedSelector$$Lambda$769/0x000000080085b840@6caa4dc5 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.util.thread.strategy.EatWhatYouKill - EatWhatYouKill@13e1e816/SelectorProducer@36238b12/IDLE/p=false/QueuedThreadPool[qtp2144817630]@7fd751de{STARTED,8<=8<=200,i=4,q=0}[ReservedThreadExecutor@67fac095{s=0/8,p=0}][pc=0,pic=0,pec=0,epc=0]@2022-07-07T18:26:53.600751Z tryProduce false 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.io.ManagedSelector - updateable 1 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.io.ManagedSelector - update org.eclipse.jetty.io.ManagedSelector$Start@6ed51944 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.io.ManagedSelector - updates 0 18:26:53.600 [qtp2144817630-59] DEBUG org.eclipse.jetty.io.ManagedSelector - Selector sun.nio.ch.EPollSelectorImpl@12fd5a53 waiting with 0 keys 18:26:53.600 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4685ms ManagedSelector@33187485{STARTED} id=3 keys=0 selected=0 updates=0 18:26:53.601 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4685ms SelectorManager@ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} 18:26:53.601 [main] DEBUG org.eclipse.jetty.util.component.ContainerLifeCycle - ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} added {acceptor-0@58e02359,POJO} 18:26:53.601 [main] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - queue acceptor-0@58e02359 18:26:53.601 [main] INFO org.eclipse.jetty.server.AbstractConnector - Started ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} 18:26:53.602 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4686ms ServerConnector@5ef2cbe6{HTTP/1.1,[http/1.1]}{0.0.0.0:8000} 18:26:53.602 [qtp2144817630-60] DEBUG org.eclipse.jetty.util.thread.QueuedThreadPool - run acceptor-0@58e02359 18:26:53.602 [main] INFO org.eclipse.jetty.server.Server - Started @4687ms 18:26:53.602 [main] DEBUG org.eclipse.jetty.util.component.AbstractLifeCycle - STARTED @4687ms Server@5f935d49{STARTED}[9.4.z-SNAPSHOT] 18:26:53.602 [main] INFO io.debezium.connector.cassandra.CassandraConnectorTaskTemplate - Starting JMX reporter ... 18:26:53.734 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Completed connection to node -3. Fetching API versions. 18:26:53.933 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.network.SslTransportLayer - [SslTransportLayer channelId=-3 key=channel=java.nio.channels.SocketChannel[connection-pending remote=kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net/10.103.24.232:9093], selector=sun.nio.ch.EPollSelectorImpl@273e9b9f, interestOps=8, readyOps=0] SSL handshake completed successfully with peerHost 'kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net' peerPort 9093 peerPrincipal 'CN=kafka-1475864931-3-1563145834.westus2.prod.us.walmart.net, OU=WeC, O=Wal-Mart Stores Inc., L=San Bruno, ST=California, C=US' cipherSuite 'TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256' 18:26:53.934 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.common.network.Selector - [Producer clientId=producer-1] Successfully authenticated with kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net/10.103.24.232 18:26:53.934 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Initiating API versions fetch from node -3. 18:26:53.956 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Sending API_VERSIONS request with header RequestHeader(apiKey=API_VERSIONS, apiVersion=3, clientId=producer-1, correlationId=0) and timeout 30000 to node -3: ApiVersionsRequestData(clientSoftwareName='apache-kafka-java', clientSoftwareVersion='3.1.0') 18:26:54.006 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Received API_VERSIONS response from node -3 for request with header RequestHeader(apiKey=API_VERSIONS, apiVersion=3, clientId=producer-1, correlationId=0): ApiVersionsResponseData(errorCode=0, apiKeys=[ApiVersion(apiKey=0, minVersion=0, maxVersion=9), ApiVersion(apiKey=1, minVersion=0, maxVersion=12), ApiVersion(apiKey=2, minVersion=0, maxVersion=6), ApiVersion(apiKey=3, minVersion=0, maxVersion=11), ApiVersion(apiKey=4, minVersion=0, maxVersion=5), ApiVersion(apiKey=5, minVersion=0, maxVersion=3), ApiVersion(apiKey=6, minVersion=0, maxVersion=7), ApiVersion(apiKey=7, minVersion=0, maxVersion=3), ApiVersion(apiKey=8, minVersion=0, maxVersion=8), ApiVersion(apiKey=9, minVersion=0, maxVersion=7), ApiVersion(apiKey=10, minVersion=0, maxVersion=3), ApiVersion(apiKey=11, minVersion=0, maxVersion=7), ApiVersion(apiKey=12, minVersion=0, maxVersion=4), ApiVersion(apiKey=13, minVersion=0, maxVersion=4), ApiVersion(apiKey=14, minVersion=0, maxVersion=5), ApiVersion(apiKey=15, minVersion=0, maxVersion=5), ApiVersion(apiKey=16, minVersion=0, maxVersion=4), ApiVersion(apiKey=17, minVersion=0, maxVersion=1), ApiVersion(apiKey=18, minVersion=0, maxVersion=3), ApiVersion(apiKey=19, minVersion=0, maxVersion=7), ApiVersion(apiKey=20, minVersion=0, maxVersion=6), ApiVersion(apiKey=21, minVersion=0, maxVersion=2), ApiVersion(apiKey=22, minVersion=0, maxVersion=4), ApiVersion(apiKey=23, minVersion=0, maxVersion=4), ApiVersion(apiKey=24, minVersion=0, maxVersion=3), ApiVersion(apiKey=25, minVersion=0, maxVersion=3), ApiVersion(apiKey=26, minVersion=0, maxVersion=3), ApiVersion(apiKey=27, minVersion=0, maxVersion=1), ApiVersion(apiKey=28, minVersion=0, maxVersion=3), ApiVersion(apiKey=29, minVersion=0, maxVersion=2), ApiVersion(apiKey=30, minVersion=0, maxVersion=2), ApiVersion(apiKey=31, minVersion=0, maxVersion=2), ApiVersion(apiKey=32, minVersion=0, maxVersion=4), ApiVersion(apiKey=33, minVersion=0, maxVersion=2), ApiVersion(apiKey=34, minVersion=0, maxVersion=2), ApiVersion(apiKey=35, minVersion=0, maxVersion=2), ApiVersion(apiKey=36, minVersion=0, maxVersion=2), ApiVersion(apiKey=37, minVersion=0, maxVersion=3), ApiVersion(apiKey=38, minVersion=0, maxVersion=2), ApiVersion(apiKey=39, minVersion=0, maxVersion=2), ApiVersion(apiKey=40, minVersion=0, maxVersion=2), ApiVersion(apiKey=41, minVersion=0, maxVersion=2), ApiVersion(apiKey=42, minVersion=0, maxVersion=2), ApiVersion(apiKey=43, minVersion=0, maxVersion=2), ApiVersion(apiKey=44, minVersion=0, maxVersion=1), ApiVersion(apiKey=45, minVersion=0, maxVersion=0), ApiVersion(apiKey=46, minVersion=0, maxVersion=0), ApiVersion(apiKey=47, minVersion=0, maxVersion=0), ApiVersion(apiKey=48, minVersion=0, maxVersion=1), ApiVersion(apiKey=49, minVersion=0, maxVersion=1), ApiVersion(apiKey=50, minVersion=0, maxVersion=0), ApiVersion(apiKey=51, minVersion=0, maxVersion=0), ApiVersion(apiKey=56, minVersion=0, maxVersion=0), ApiVersion(apiKey=57, minVersion=0, maxVersion=0), ApiVersion(apiKey=60, minVersion=0, maxVersion=0), ApiVersion(apiKey=61, minVersion=0, maxVersion=0)], throttleTimeMs=0, supportedFeatures=[], finalizedFeaturesEpoch=0, finalizedFeatures=[]) 18:26:54.049 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Node -3 has finalized features epoch: 0, finalized features: [], supported features: [], API versions: (Produce(0): 0 to 9 [usable: 9], Fetch(1): 0 to 12 [usable: 12], ListOffsets(2): 0 to 6 [usable: 6], Metadata(3): 0 to 11 [usable: 11], LeaderAndIsr(4): 0 to 5 [usable: 5], StopReplica(5): 0 to 3 [usable: 3], UpdateMetadata(6): 0 to 7 [usable: 7], ControlledShutdown(7): 0 to 3 [usable: 3], OffsetCommit(8): 0 to 8 [usable: 8], OffsetFetch(9): 0 to 7 [usable: 7], FindCoordinator(10): 0 to 3 [usable: 3], JoinGroup(11): 0 to 7 [usable: 7], Heartbeat(12): 0 to 4 [usable: 4], LeaveGroup(13): 0 to 4 [usable: 4], SyncGroup(14): 0 to 5 [usable: 5], DescribeGroups(15): 0 to 5 [usable: 5], ListGroups(16): 0 to 4 [usable: 4], SaslHandshake(17): 0 to 1 [usable: 1], ApiVersions(18): 0 to 3 [usable: 3], CreateTopics(19): 0 to 7 [usable: 7], DeleteTopics(20): 0 to 6 [usable: 6], DeleteRecords(21): 0 to 2 [usable: 2], InitProducerId(22): 0 to 4 [usable: 4], OffsetForLeaderEpoch(23): 0 to 4 [usable: 4], AddPartitionsToTxn(24): 0 to 3 [usable: 3], AddOffsetsToTxn(25): 0 to 3 [usable: 3], EndTxn(26): 0 to 3 [usable: 3], WriteTxnMarkers(27): 0 to 1 [usable: 1], TxnOffsetCommit(28): 0 to 3 [usable: 3], DescribeAcls(29): 0 to 2 [usable: 2], CreateAcls(30): 0 to 2 [usable: 2], DeleteAcls(31): 0 to 2 [usable: 2], DescribeConfigs(32): 0 to 4 [usable: 4], AlterConfigs(33): 0 to 2 [usable: 2], AlterReplicaLogDirs(34): 0 to 2 [usable: 2], DescribeLogDirs(35): 0 to 2 [usable: 2], SaslAuthenticate(36): 0 to 2 [usable: 2], CreatePartitions(37): 0 to 3 [usable: 3], CreateDelegationToken(38): 0 to 2 [usable: 2], RenewDelegationToken(39): 0 to 2 [usable: 2], ExpireDelegationToken(40): 0 to 2 [usable: 2], DescribeDelegationToken(41): 0 to 2 [usable: 2], DeleteGroups(42): 0 to 2 [usable: 2], ElectLeaders(43): 0 to 2 [usable: 2], IncrementalAlterConfigs(44): 0 to 1 [usable: 1], AlterPartitionReassignments(45): 0 [usable: 0], ListPartitionReassignments(46): 0 [usable: 0], OffsetDelete(47): 0 [usable: 0], DescribeClientQuotas(48): 0 to 1 [usable: 1], AlterClientQuotas(49): 0 to 1 [usable: 1], DescribeUserScramCredentials(50): 0 [usable: 0], AlterUserScramCredentials(51): 0 [usable: 0], AlterIsr(56): 0 [usable: 0], UpdateFeatures(57): 0 [usable: 0], DescribeCluster(60): 0 [usable: 0], DescribeProducers(61): 0 [usable: 0], DescribeTransactions(65): UNSUPPORTED, ListTransactions(66): UNSUPPORTED, AllocateProducerIds(67): UNSUPPORTED). 18:26:54.050 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Sending metadata request MetadataRequestData(topics=[], allowAutoTopicCreation=true, includeClusterAuthorizedOperations=false, includeTopicAuthorizedOperations=false) to node kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: -3 rack: null) 18:26:54.050 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Sending METADATA request with header RequestHeader(apiKey=METADATA, apiVersion=11, clientId=producer-1, correlationId=1) and timeout 30000 to node -3: MetadataRequestData(topics=[], allowAutoTopicCreation=true, includeClusterAuthorizedOperations=false, includeTopicAuthorizedOperations=false) 18:26:54.079 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.NetworkClient - [Producer clientId=producer-1] Received METADATA response from node -3 for request with header RequestHeader(apiKey=METADATA, apiVersion=11, clientId=producer-1, correlationId=1): MetadataResponseData(throttleTimeMs=0, brokers=[MetadataResponseBroker(nodeId=2106540296, host='kafka-1475864931-13-1563145864.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='1'), MetadataResponseBroker(nodeId=928058233, host='kafka-1475864931-1-1563145828.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='0'), MetadataResponseBroker(nodeId=1420740148, host='kafka-1475864931-9-1563145852.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='2'), MetadataResponseBroker(nodeId=498548389, host='kafka-1475864931-11-1563145858.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='2'), MetadataResponseBroker(nodeId=1156618488, host='kafka-1475864931-10-1563145855.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='0'), MetadataResponseBroker(nodeId=624691423, host='kafka-1475864931-7-1563145846.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='1'), MetadataResponseBroker(nodeId=2024303373, host='kafka-1475864931-8-1563145849.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='1'), MetadataResponseBroker(nodeId=982633903, host='kafka-1475864931-15-1563145870.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='0'), MetadataResponseBroker(nodeId=875732220, host='kafka-1475864931-5-1563145840.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='2'), MetadataResponseBroker(nodeId=446590054, host='kafka-1475864931-12-1563145861.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='2'), MetadataResponseBroker(nodeId=1541242006, host='kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='1'), MetadataResponseBroker(nodeId=906322439, host='kafka-1475864931-4-1563145837.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='1'), MetadataResponseBroker(nodeId=129848562, host='kafka-1475864931-6-1563145843.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='0'), MetadataResponseBroker(nodeId=394146119, host='kafka-1475864931-2-1563145831.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='2'), MetadataResponseBroker(nodeId=1006058978, host='kafka-1475864931-14-1563145867.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net', port=9093, rack='0')], clusterId='9p8m3OXlQwSSHPTjYCHTJQ', controllerId=1541242006, topics=[], clusterAuthorizedOperations=-2147483648) 18:26:54.082 [kafka-producer-network-thread | producer-1] INFO org.apache.kafka.clients.Metadata - [Producer clientId=producer-1] Cluster ID: 9p8m3OXlQwSSHPTjYCHTJQ 18:26:54.082 [kafka-producer-network-thread | producer-1] DEBUG org.apache.kafka.clients.Metadata - [Producer clientId=producer-1] Updated cluster metadata updateVersion 2 to MetadataCache{clusterId='9p8m3OXlQwSSHPTjYCHTJQ', nodes={906322439=kafka-1475864931-4-1563145837.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 906322439 rack: 1), 624691423=kafka-1475864931-7-1563145846.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 624691423 rack: 1), 2024303373=kafka-1475864931-8-1563145849.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 2024303373 rack: 1), 2106540296=kafka-1475864931-13-1563145864.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 2106540296 rack: 1), 928058233=kafka-1475864931-1-1563145828.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 928058233 rack: 0), 1156618488=kafka-1475864931-10-1563145855.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 1156618488 rack: 0), 1541242006=kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 1541242006 rack: 1), 875732220=kafka-1475864931-5-1563145840.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 875732220 rack: 2), 129848562=kafka-1475864931-6-1563145843.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 129848562 rack: 0), 498548389=kafka-1475864931-11-1563145858.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 498548389 rack: 2), 1006058978=kafka-1475864931-14-1563145867.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 1006058978 rack: 0), 446590054=kafka-1475864931-12-1563145861.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 446590054 rack: 2), 394146119=kafka-1475864931-2-1563145831.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 394146119 rack: 2), 1420740148=kafka-1475864931-9-1563145852.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 1420740148 rack: 2), 982633903=kafka-1475864931-15-1563145870.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 982633903 rack: 0)}, partitions=[], controller=kafka-1475864931-3-1563145834.wus.kafka-taas-shared5-stg.ms-df-messaging.stg-az-westus2-1.westus2.prod.us.walmart.net:9093 (id: 1541242006 rack: 1)} 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.449 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:54.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - polling records... 18:26:54.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:54.539 [pool-2-thread-1] INFO org.apache.cassandra.db.commitlog.CommitLogReader - Finished reading /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log 18:26:54.539 [pool-2-thread-1] DEBUG io.debezium.connector.base.ChangeEventQueue - Enqueuing source record 'io.debezium.connector.cassandra.EOFEvent@b95cf29' 18:26:54.540 [pool-2-thread-1] INFO io.debezium.connector.cassandra.Cassandra4CommitLogProcessor$CommitLogProcessingCallable - ProcessingResult{commitLog=LogicalCommitLog{commitLogPosition=CommitLogPosition(segmentId=1652287118530, position=0), synced=33552621, completed=true, log=/mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log, index=/mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530_cdc.idx, commitLogSegmentId=1652287118530}, result=OK, ex=none} 18:26:54.540 [pool-2-thread-1] INFO io.debezium.connector.cassandra.Cassandra4CommitLogProcessor$CommitLogProcessingCallable - Processing commit log /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118531.log 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit... 18:26:55.450 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - checking for more records... 18:26:55.451 [pool-4-thread-3] INFO io.debezium.connector.cassandra.QueueProcessor - Encountered EOF event for CommitLog-7-1652287118530.log ... 18:26:55.462 [pool-4-thread-4] INFO io.debezium.connector.cassandra.AbstractDirectoryWatcher - No commitLogFile is detected in /mnt/resource/cassandra/commitlog/cdc_raw. 18:26:55.462 [pool-4-thread-4] DEBUG io.debezium.connector.cassandra.Cassandra4CommitLogProcessor - Processing commitLogFiles while initial is false 18:26:55.462 [pool-4-thread-4] INFO io.debezium.connector.cassandra.AbstractDirectoryWatcher - Polling commitLog files from /mnt/resource/cassandra/commitlog/cdc_raw ... 18:26:55.494 [pool-4-thread-3] INFO io.debezium.connector.cassandra.CommitLogUtil - Moved CommitLog file /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530.log to /tmp/debezium-connector-cassandra/test_dir/relocation/archive. 18:26:55.494 [pool-4-thread-3] INFO io.debezium.connector.cassandra.CommitLogUtil - Moved CommitLog index file /mnt/resource/cassandra/commitlog/cdc_raw/CommitLog-7-1652287118530_cdc.idx to /tmp/debezium-connector-cassandra/test_dir/relocation/archive. 18:26:55.494 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - polling records... 18:26:55.494 [pool-4-thread-3] DEBUG io.debezium.connector.base.ChangeEventQueue - no records available yet, sleeping a bit...