Uploaded image for project: 'Debezium'
  1. Debezium
  2. DBZ-2575

Improve performance for very large postgres schemas

XMLWordPrintable

    • False
    • False
    • Undefined

      Hello! We are using Debezium to track changes to a single table in Postgres and have an extremely large database (~350 schemas, ~200 tables per schema, ~15 columns per table). When Debezium starts up it seems to be loading the entire schema, which for one of our production-like environments can take up to a half hour of constant querying to load the required data. Since we really only want the events from a single table, is there a way to limit how much of the schema that Debezium needs to load? Or is there a better approach to this problem?
       
      We have snapshotting set to `never` and are using the table whitelist.
       
      This was originally reported on Gitter, and I wish there was a way to link to a thread so that I could attach it here, instead I've copied it into the comments below. I was advised to convert the thread into a feature enhancement.

            Unassigned Unassigned
            G-e88d06f1-cef7-44d4-bc07-cffc06a293a4 Adam Kerr (Inactive)
            Votes:
            0 Vote for this issue
            Watchers:
            8 Start watching this issue

              Created:
              Updated:
              Resolved: