I'll be doing this again this week & next looking at a few different topics.
Tomorrow (July 25th @ 10am pacific) Gris & I will be updating the PR from
my last live stream (adding Python dependency handling) -
https://www.twitch.tv/events/P92irbgYR9Sx6nMQ-lGY3g /
https://www.youtube.com/watch?v=4xD
Hi Friends,
If you missed my talk @ OSCON on TFT/TFMA + Beam on Flink (and other
related adventures), I've uploaded my slides to
https://www.slideshare.net/hkarau/powering-tensorflow-with-big-data-using-apache-beam-flink-and-spark-oscon-pdx-2018
and once the video is done I'll try and find a way t
Are you trying to read a growing file ? I don't think this scenario is well
supported. You can use FileIO.MatchAll.continuously() if you want to read a
growing list of files (where new files get added to a given directory).
If you are reading a large but fixed set of files then what you need is a
To unsubscribe, please use the -unsubscribe addresses listed on
https://beam.apache.org/community/contact-us/
On Tue, Jul 24, 2018 at 6:34 AM Chandan Biswas
wrote:
>
>
Hi,
Please consider following pipeline:-
Source is Parquet file, having hundreds of columns.
Sink is Parquet. Multiple output parquet files are generated after applying
some sql joins. Sql joins to be applied differ for each output parquet file.
Lets assume we have a sql queries generator or
Just to share my findings to others ...
I noticed that elements forwarded from WriteFileResults.
getPerDestinationOutputFilenames() have always pane index 0 and correct
timing of written window (such as EARLY, LATE).
I am not sure what is going on but following Window.into() were not working.
Ins