How should the existing GBs or probably TBs of data in a large enterprise be made available for immediate availability in Kafka topics?
While I know that we can use the CDC capabilities of Kafka to build an event log or DB replica, with CDC the Kafka topics will be filled only whenever there is a change in the backend DB.
What is the best approach to build a Kafka layer which replaces the need to hit the real backend databases. How will the millions of records in existing database tables be available in Kafka topics as soon as we deploy our application integrated with Kafka CDC connector?