Tentatively I'm planning on this list to start backporting. If no one sees any issues with those I'll start to make backport JIRAs for them for tracking this afternoon. SPARK-26390 ColumnPruning rule should only do column pruning SPARK-25407 Allow nested access for non-existent field for Parquet file when nested pruning is enabled SPARK-25559 Remove the unsupported predicates in Parquet when possible SPARK-25860 Replace Literal(null, _) with FalseLiteral whenever possible SPARK-27514 Skip collapsing windows with empty window expressions SPARK-25338 Ensure to call super.beforeAll() and super.afterAll() in test cases SPARK-27138 Remove AdminUtils calls (fixes deprecation) SPARK-27981 Remove `Illegal reflective access` warning for `java.nio.Bits.unaligned()` in JDK9+ SPARK-26095 Disable parallelization in make-distibution.sh. (Avoid build hanging) SPARK-25692 Remove static initialization of worker eventLoop handling chunk fetch requests within TransportContext. This fixes ChunkFetchIntegrationSuite as well SPARK-26306 More memory to de-flake SorterSuite SPARK-30199 Recover `spark.(ui|blockManager).port` from checkpoint SPARK-27676 InMemoryFileIndex should respect spark.sql.files.ignoreMissingFiles SPARK-31047 Improve file listing for ViewFileSystem SPARK-25595 Ignore corrupt Avro file if flag IGNORE_CORRUPT_FILES enabled
Maybe: SPARK-27801 Delegate to ViewFileSystem during file listing correctly Not yet merged: SPARK-31485 Barrier execution hang if insufficient resources On Thu, Apr 23, 2020 at 9:13 AM Holden Karau <hol...@pigscanfly.ca> wrote: > > > On Thu, Apr 23, 2020 at 9:07 AM edeesis <edee...@gmail.com> wrote: > >> There's other information you can obtain from the Pod metadata on a >> describe >> than just from the logs, which are typically what's being printed by the >> Application itself. > > Would get pods -w -o yaml do the trick here or is there going to be > information that wouldn’t be captured that way? > >> >> >> I've also found that Spark has some trouble obtaining the reason for a K8S >> executor death (as evident by the >> spark.kubernetes.executor.lostCheck.maxAttempts config property) >> >> I admittedly don't know what should qualify for a backport, but >> considering >> 3.0 is a major upgrade (Scala version, et al), is there any room for for >> being more generous with backporting to 2.4? > > I’d like to revisit the conversation around a Spark 2.5 as a transitional > release. I know that some people are already effectively maintaining 2.4+ > Selective new functionality backports internally. Maybe I’ll kick off that > discussion which we can have and that can help inform what we should be > putting in 2.4. > >> >> >> >> >> -- >> Sent from: http://apache-spark-developers-list.1001551.n3.nabble.com/ >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> >> -- > Twitter: https://twitter.com/holdenkarau > Books (Learning Spark, High Performance Spark, etc.): > https://amzn.to/2MaRAG9 <https://amzn.to/2MaRAG9> > YouTube Live Streams: https://www.youtube.com/user/holdenkarau > -- Twitter: https://twitter.com/holdenkarau Books (Learning Spark, High Performance Spark, etc.): https://amzn.to/2MaRAG9 <https://amzn.to/2MaRAG9> YouTube Live Streams: https://www.youtube.com/user/holdenkarau