unsubscribe
On Fri, Aug 6, 2021 at 10:36 AM Jeremy Lewi wrote:
> Hi Folks,
>
> I'm running Beam Python (2.31.0) on Flink on Kubernetes. I'm using the
> PortableRunner and a beam job server.
> I'm using S3 for the artifacts dir. The job server is throwing exceptions
> like the one below
Hi Folks,
I'm running Beam Python (2.31.0) on Flink on Kubernetes. I'm using the
PortableRunner and a beam job server.
I'm using S3 for the artifacts dir. The job server is throwing exceptions
like the one below complaining that filesystem scheme S3 isn't registered.
I'm using the stock job
Hooray! Thanks, Etienne!
On Thu, Aug 5, 2021 at 3:11 AM Etienne Chauchot
wrote:
> Hi all,
>
> Just to let you know that Spark Structured Streaming runner was migrated
> to Spark 3.
>
> Enjoy !
>
> Etienne
>
>
Sorry. Here is the link:
https://doodle.com/poll/d57tcpt46tkvtvay?utm_source=poll_medium=link
On Wed, Aug 4, 2021 at 3:32 PM Aizhamal Nurmamat kyzy
wrote:
> Hi all,
>
> I will get this workshop scheduled for September. I am trying to figure
> out which day/time works best considering the US and
Hi Alexey,
It was a great presentation!
Regarding my perf testing, I was not doing aggregation, filtering, projection
or joining. I was simply reading all the fields of parquet and then immediately
save PCollection back to parquet.
Regarding SDF translation, is it enabled by default?
I will
It’s very likely that Spark SQL may have much better performance because of SQL
push-downs and avoiding additional ser/deser operations.
In the same time, did you try to leverage "withProjection()” in ParquetIO and
project only the fields that you needed?
Did you use ParquetIO splittable
I also want to be invited :)
On Thu, Aug 5, 2021 at 1:45 AM Koosha Hosseiny
wrote:
> Hello there
> Can I also be invited?
>
> Many thanks.
>
>
>
>
>
>
>
>
>
> --
> *From:* Bergmeier, Andreas
> *Sent:* Thursday, August 5, 2021 06:48
> *To:* user@beam.apache.org
>
Hi all,
Just to let you know that Spark Structured Streaming runner was migrated
to Spark 3.
Enjoy !
Etienne