And, in a followup question - what is the current recommended approach to dumping such data into the deltalake/dremio/spark family of tools?
On Tue, Jan 28, 2025 at 6:39 AM Richard Beare <richard.be...@gmail.com> wrote: > Thanks - that sounds very useful. > > On Tue, Jan 28, 2025 at 2:00 AM Matt Burgess <mattyb...@gmail.com> wrote: > >> You can use ListDatabaseTables -> GenerateTableFetch to handle an >> arbitrary number of tables, and if you don't have the target tables created >> you can use UpdateDatabaseTable once the fetch of data is done to create >> the target tables if they don't exist, so something like: >> >> ListDatabaseTables -> GenerateTableFetch -> ExecuteSQL -> >> UpdateDatabaseTable -> PutDatabaseRecord >> >> Regards, >> Matt >> >> On Sun, Jan 26, 2025 at 6:17 PM Richard Beare <richard.be...@gmail.com> >> wrote: >> >>> I'll have a look - the previous workflow I developed for a single large >>> table used GenerateTableFetch followed by ExecuteSQLRecord. >>> >>> Did you develop any useful strategies for larger numbers of tables, or >>> was doing everything by hand the best way in the long run? I'm thinking of >>> code to generate parts of the workflow, or some kind of template to >>> describe key parts of the table structure. >>> >>> On Mon, Jan 27, 2025 at 9:56 AM Dennis N Brown <mr.dnbr...@gmail.com> >>> wrote: >>> >>>> I have always used "QueryDatabaseTable" for this... it does tend to get >>>> complicated when you have a large number of tables, but you have a lot of >>>> control, like setting a starting point for the "copy". >>>> >>>> On Sun, Jan 26, 2025, 17:04 Richard Beare <richard.be...@gmail.com> >>>> wrote: >>>> >>>>> Hi, >>>>> I have a project for which mirroring, potentially in a close to real >>>>> time fashion, of a complex DB might be very useful. We have only limited >>>>> access to the source systems and therefore can't use some of the more >>>>> conventional approaches for live replication. I'm aware that >>>>> GenerateTableFetch is useful for fetching recent changes from a remote >>>>> table. Does anyone have experience using it, or an alternative, for a case >>>>> where there may be hundreds of tables? If so, can you recommend strategies >>>>> for keeping the system maintainable? >>>>> Thanks >>>>> >>>>