Re: Let multiple jobs share one rdd?

2020-09-24 Thread Khalid Mammadov
Perhaps you can use Global Temp Views? https://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.createGlobalTempView On 24/09/2020 14:52, Gang Li wrote: Hi all, There are three jobs, among which the first rdd is the same. Can the first rdd be calculated once, and

Let multiple jobs share one rdd?

2020-09-24 Thread Gang Li
Hi all, There are three jobs, among which the first rdd is the same. Can the first rdd be calculated once, and then the subsequent operations will be calculated in parallel? My code is as follows: sqls =