Thanks - that sounds very useful. On Tue, Jan 28, 2025 at 2:00 AM Matt Burgess <mattyb...@gmail.com> wrote:
> You can use ListDatabaseTables -> GenerateTableFetch to handle an > arbitrary number of tables, and if you don't have the target tables created > you can use UpdateDatabaseTable once the fetch of data is done to create > the target tables if they don't exist, so something like: > > ListDatabaseTables -> GenerateTableFetch -> ExecuteSQL -> > UpdateDatabaseTable -> PutDatabaseRecord > > Regards, > Matt > > On Sun, Jan 26, 2025 at 6:17 PM Richard Beare <richard.be...@gmail.com> > wrote: > >> I'll have a look - the previous workflow I developed for a single large >> table used GenerateTableFetch followed by ExecuteSQLRecord. >> >> Did you develop any useful strategies for larger numbers of tables, or >> was doing everything by hand the best way in the long run? I'm thinking of >> code to generate parts of the workflow, or some kind of template to >> describe key parts of the table structure. >> >> On Mon, Jan 27, 2025 at 9:56 AM Dennis N Brown <mr.dnbr...@gmail.com> >> wrote: >> >>> I have always used "QueryDatabaseTable" for this... it does tend to get >>> complicated when you have a large number of tables, but you have a lot of >>> control, like setting a starting point for the "copy". >>> >>> On Sun, Jan 26, 2025, 17:04 Richard Beare <richard.be...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> I have a project for which mirroring, potentially in a close to real >>>> time fashion, of a complex DB might be very useful. We have only limited >>>> access to the source systems and therefore can't use some of the more >>>> conventional approaches for live replication. I'm aware that >>>> GenerateTableFetch is useful for fetching recent changes from a remote >>>> table. Does anyone have experience using it, or an alternative, for a case >>>> where there may be hundreds of tables? If so, can you recommend strategies >>>> for keeping the system maintainable? >>>> Thanks >>>> >>>