Thanks for the suggestions guys. The pre-filtered list is possibly one I can use.
Regards, Jeremy > On 9 Mar 2020, at 20:16, Shawn Weeks <[email protected]> wrote: > When I’ve had to do this I just skip trying to use ListFile and instead create a text file containing a list of all the files that can be used with the SplitFile and FetchFile processors to pull things in in batches. Even with filtering ListFile will iterate through a lot of files. Thanks From: Edward Armes <[email protected]> Reply-To: "[email protected]" <[email protected]> Date: Monday, March 9, 2020 at 4:43 AM To: "[email protected]" <[email protected]> Subject: Re: Listing a folder with millions of files Hi Jeremy, In this case I don't think there is an easy answer here. You may have some luck with adjusting the max runtime of the processor but without checking the the processors implementation I couldn't know for certain if that would have any effect. Edward On Mon, 9 Mar 2020, 06:34 Jeremy Pemberton-Pigott, <[email protected]> wrote: Hi, I need to list a sub-set (few 100,000) of files in a folder with millions of files (to do some historical processing). What's the best way I can do that? ListFiles is taking way too long and seems to try to dump the entire list to the flow when I test it on a smaller folder list. It would be good if the listing emitted files in smaller chunks so that the flow can start working on them. Regards, Jeremy
