Sean thanks for checking them!

I made one pass and re-targeted/closed some of them. Most of them are
documentation and auditing, do we need to block the release for them?

On Fri, Sep 21, 2018 at 6:01 AM Sean Owen <sro...@apache.org> wrote:

> Because we're into 2.4 release candidates, I thought I'd look at
> what's still open and targeted at 2.4.0. I presume the Blockers are
> the usual umbrellas that don't themselves block anything, but,
> confirming, there is nothing left to do there?
>
> I think that's mostly a question for Joseph and Weichen.
>
> As ever, anyone who knows these items are a) done or b) not going to
> be in 2.4, go ahead and update them.
>
>
> Blocker:
>
> SPARK-25321 ML, Graph 2.4 QA: API: New Scala APIs, docs
> SPARK-25324 ML 2.4 QA: API: Java compatibility, docs
> SPARK-25323 ML 2.4 QA: API: Python API coverage
> SPARK-25320 ML, Graph 2.4 QA: API: Binary incompatible changes
>
> Critical:
>
> SPARK-25319 Spark MLlib, GraphX 2.4 QA umbrella
> SPARK-25378 ArrayData.toArray(StringType) assume UTF8String in 2.4
> SPARK-25327 Update MLlib, GraphX websites for 2.4
> SPARK-25325 ML, Graph 2.4 QA: Update user guide for new features & APIs
> SPARK-25326 ML, Graph 2.4 QA: Programming guide update and migration guide
>
> Other:
>
> SPARK-25346 Document Spark builtin data sources
> SPARK-25347 Document image data source in doc site
> SPARK-12978 Skip unnecessary final group-by when input data already
> clustered with group-by keys
> SPARK-20184 performance regression for complex/long sql when enable
> whole stage codegen
> SPARK-16196 Optimize in-memory scan performance using ColumnarBatches
> SPARK-15693 Write schema definition out for file-based data sources to
> avoid schema inference
> SPARK-23597 Audit Spark SQL code base for non-interpreted expressions
> SPARK-25179 Document the features that require Pyarrow 0.10
> SPARK-25110 make sure Flume streaming connector works with Spark 2.4
> SPARK-21318 The exception message thrown by `lookupFunction` is ambiguous.
> SPARK-24464 Unit tests for MLlib's Instrumentation
> SPARK-23197 Flaky test: spark.streaming.ReceiverSuite."receiver_life_cycle"
> SPARK-22809 pyspark is sensitive to imports with dots
> SPARK-22739 Additional Expression Support for Objects
> SPARK-22231 Support of map, filter, withColumn, dropColumn in nested
> list of structures
> SPARK-21030 extend hint syntax to support any expression for Python and R
> SPARK-22386 Data Source V2 improvements
> SPARK-15117 Generate code that get a value in each compressed column
> from CachedBatch when DataFrame.cache() is called
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to