[
https://issues.apache.org/jira/browse/SPARK-37500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17453294#comment-17453294
]
Rui Wang commented on SPARK-37500:
--
I am thinking this is no longer valid? Running on 3.2 spark-cli:
[
https://issues.apache.org/jira/browse/SPARK-37500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17453215#comment-17453215
]
Rui Wang commented on SPARK-37500:
--
I can take a look on this bug.
> Incorrect scope when using
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
Rui Wang created SPARK-38118:
Summary: MEAN(Boolean) in the HAVING claus should throw data
mismatch error
Key: SPARK-38118
URL: https://issues.apache.org/jira/browse/SPARK-38118
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
Rui Wang created SPARK-38063:
Summary: Support SQL split_part function
Key: SPARK-38063
URL: https://issues.apache.org/jira/browse/SPARK-38063
Project: Spark
Issue Type: Task
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17484029#comment-17484029
]
Rui Wang commented on SPARK-38063:
--
PR is ready: https://github.com/apache/spark/pull/35352
> Support
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
Rui Wang created SPARK-38170:
Summary: Fix
//sql/hive-thriftserver:org.apache.spark.sql.hive.thriftserver.ThriftServerWithSparkContextInHttpSuite-hive-2.3__hadoop-2.7
in ANSI
Key: SPARK-38170
URL:
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-38063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-38063:
-
Description:
`split_part()` is a commonly supported function by other systems such as
Postgres and
[
https://issues.apache.org/jira/browse/SPARK-45012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-45012:
-
Affects Version/s: 4.0.0
(was: 3.5.0)
> CheckAnalysis should throw inlined
Rui Wang created SPARK-45012:
Summary: CheckAnalysis should throw inlined plan in
AnalysisException
Key: SPARK-45012
URL: https://issues.apache.org/jira/browse/SPARK-45012
Project: Spark
Issue
Rui Wang created SPARK-45752:
Summary: Unreferenced CTE should all be checked by CheckAnalysis0
Key: SPARK-45752
URL: https://issues.apache.org/jira/browse/SPARK-45752
Project: Spark
Issue Type:
[
https://issues.apache.org/jira/browse/SPARK-45242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang resolved SPARK-45242.
--
Fix Version/s: 4.0.0
Resolution: Fixed
https://github.com/apache/spark/pull/43010
> Use
Rui Wang created SPARK-45242:
Summary: Use DataFrame ID to semantically validate CollectMetrics
Key: SPARK-45242
URL: https://issues.apache.org/jira/browse/SPARK-45242
Project: Spark
Issue
Rui Wang created SPARK-45421:
Summary: Catch AnalysisException over InlineCTE
Key: SPARK-45421
URL: https://issues.apache.org/jira/browse/SPARK-45421
Project: Spark
Issue Type: Task
Rui Wang created SPARK-44921:
Summary: Remove SqlBaseLexer.tokens from codebase
Key: SPARK-44921
URL: https://issues.apache.org/jira/browse/SPARK-44921
Project: Spark
Issue Type: Task
[
https://issues.apache.org/jira/browse/SPARK-39012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39012:
-
Summary: SparkSQL infer schema does not support all data types (was:
SparkSQL Infer schema path does
Rui Wang created SPARK-39012:
Summary: SparkSQL Infer schema path does not support all data types
Key: SPARK-39012
URL: https://issues.apache.org/jira/browse/SPARK-39012
Project: Spark
Issue
[
https://issues.apache.org/jira/browse/SPARK-39012?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17527767#comment-17527767
]
Rui Wang commented on SPARK-39012:
--
PR is ready to support binary type
[
https://issues.apache.org/jira/browse/SPARK-39012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39012:
-
Description:
When Spark needs to infer schema, it needs to parse string to a type. Not all
data types
[
https://issues.apache.org/jira/browse/SPARK-39012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39012:
-
Description:
When Spark needs to infer schema, it needs to parse string to a type. Not all
data types
[
https://issues.apache.org/jira/browse/SPARK-39144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39144:
-
Description: When we have nested subquery expressions, there is a chance
that deduplicate relations
[
https://issues.apache.org/jira/browse/SPARK-39144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39144:
-
Summary: Nested subquery expressions deduplicate relations should be done
bottom up (was: Spark SQL
Rui Wang created SPARK-39263:
Summary: GetTable, TableExists and DatabaseExists
Key: SPARK-39263
URL: https://issues.apache.org/jira/browse/SPARK-39263
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-39236:
Summary: Make CreateTable API and ListTables API compatible
Key: SPARK-39236
URL: https://issues.apache.org/jira/browse/SPARK-39236
Project: Spark
Issue Type:
[
https://issues.apache.org/jira/browse/SPARK-39235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39235:
-
Component/s: SQL
(was: Spark Core)
> Make Catalog API be compatible with
[
https://issues.apache.org/jira/browse/SPARK-39236?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17539762#comment-17539762
]
Rui Wang commented on SPARK-39236:
--
https://github.com/apache/spark/pull/36586
> Make CreateTable API
Rui Wang created SPARK-39235:
Summary: Make Catalog API be compatible with 3-layer-namespace
Key: SPARK-39235
URL: https://issues.apache.org/jira/browse/SPARK-39235
Project: Spark
Issue Type:
Rui Wang created SPARK-39548:
Summary: CreateView Command with a window clause query hit a wrong
window definition not found issue.
Key: SPARK-39548
URL: https://issues.apache.org/jira/browse/SPARK-39548
[
https://issues.apache.org/jira/browse/SPARK-39548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39548:
-
Summary: CreateView Command with a window clause query hit a wrong window
definition not found issue
Rui Wang created SPARK-39583:
Summary: Make RefreshTable be compatible with 3 layer namespace
Key: SPARK-39583
URL: https://issues.apache.org/jira/browse/SPARK-39583
Project: Spark
Issue Type:
Rui Wang created SPARK-39700:
Summary: Deprecate API that has parameters (DBName,
tableName/FunctionName)
Key: SPARK-39700
URL: https://issues.apache.org/jira/browse/SPARK-39700
Project: Spark
Rui Wang created SPARK-39506:
Summary: CacheTable, isCached, UncacheTable, setCurrentCatalog,
currentCatalog, listCatalogs
Key: SPARK-39506
URL: https://issues.apache.org/jira/browse/SPARK-39506
Project:
Rui Wang created SPARK-39828:
Summary: Catalog.listTables() should respect currentCatalog
Key: SPARK-39828
URL: https://issues.apache.org/jira/browse/SPARK-39828
Project: Spark
Issue Type:
[
https://issues.apache.org/jira/browse/SPARK-39828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17569140#comment-17569140
]
Rui Wang commented on SPARK-39828:
--
however it seems that temp table will not be checked.
>
[
https://issues.apache.org/jira/browse/SPARK-39810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39810:
-
Summary: Catalog.tableExists should handle nested namespace (was:
tableExists can reuse getTable code)
Rui Wang created SPARK-39810:
Summary: tableExists can reuse getTable code
Key: SPARK-39810
URL: https://issues.apache.org/jira/browse/SPARK-39810
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-39012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-39012:
-
Summary: SparkSQL parse partition value does not support all data types
(was: SparkSQL infer schema
[
https://issues.apache.org/jira/browse/SPARK-39144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17534580#comment-17534580
]
Rui Wang commented on SPARK-39144:
--
Testing in https://github.com/apache/spark/pull/36503. Will come up
Rui Wang created SPARK-39144:
Summary: Spark SQL replace wrong attributes for nested subquery
expression in which all tables are the same relation
Key: SPARK-39144
URL:
Rui Wang created SPARK-40296:
Summary: Error Class for DISTINCT function not found
Key: SPARK-40296
URL: https://issues.apache.org/jira/browse/SPARK-40296
Project: Spark
Issue Type: Task
Rui Wang created SPARK-40586:
Summary: Decouple plan transformation and validation on server
side
Key: SPARK-40586
URL: https://issues.apache.org/jira/browse/SPARK-40586
Project: Spark
Issue
Rui Wang created SPARK-40828:
Summary: Drop Python test tables before and after unit tests
Key: SPARK-40828
URL: https://issues.apache.org/jira/browse/SPARK-40828
Project: Spark
Issue Type:
Rui Wang created SPARK-40816:
Summary: Python: rename LogicalPlan.collect to LogicalPlan.to_proto
Key: SPARK-40816
URL: https://issues.apache.org/jira/browse/SPARK-40816
Project: Spark
Issue
Rui Wang created SPARK-40813:
Summary: Add limit and offset to Connect DSL
Key: SPARK-40813
URL: https://issues.apache.org/jira/browse/SPARK-40813
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40809:
Summary: Add as(alias) to connect DSL
Key: SPARK-40809
URL: https://issues.apache.org/jira/browse/SPARK-40809
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-40809?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40809:
-
Summary: Add as(alias: String) to connect DSL (was: Add as(alias) to
connect DSL)
> Add as(alias:
Rui Wang created SPARK-40812:
Summary: Add Deduplicate to Connect proto
Key: SPARK-40812
URL: https://issues.apache.org/jira/browse/SPARK-40812
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40836:
Summary: AnalyzeResult should use struct for schema
Key: SPARK-40836
URL: https://issues.apache.org/jira/browse/SPARK-40836
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40883:
Summary: Support Range in Connect proto
Key: SPARK-40883
URL: https://issues.apache.org/jira/browse/SPARK-40883
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40879:
Summary: Support Join UsingColumns in proto
Key: SPARK-40879
URL: https://issues.apache.org/jira/browse/SPARK-40879
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40915:
Summary: Improve `on` in Join in Python client
Key: SPARK-40915
URL: https://issues.apache.org/jira/browse/SPARK-40915
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40926:
Summary: Refactor server side tests to only use DataFrame API
Key: SPARK-40926
URL: https://issues.apache.org/jira/browse/SPARK-40926
Project: Spark
Issue Type:
Rui Wang created SPARK-40930:
Summary: Support Collect() in Python client
Key: SPARK-40930
URL: https://issues.apache.org/jira/browse/SPARK-40930
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40914:
Summary: Mark private API to be private[connect]
Key: SPARK-40914
URL: https://issues.apache.org/jira/browse/SPARK-40914
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-40914?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40914:
-
Summary: Mark internal API to be private[connect] (was: Mark private API
to be private[connect])
>
[
https://issues.apache.org/jira/browse/SPARK-40970?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40970:
-
Description: Right now Join's on does not support a list of ColumnRef:
[df.age == df2.age, df.name ==
Rui Wang created SPARK-40970:
Summary: Support List[ColumnRef] for Join's on argument.
Key: SPARK-40970
URL: https://issues.apache.org/jira/browse/SPARK-40970
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40971:
Summary: Imports more from connect proto package to avoid calling
`proto.` for Connect DSL
Key: SPARK-40971
URL: https://issues.apache.org/jira/browse/SPARK-40971
Project:
Rui Wang created SPARK-40938:
Summary: Support Alias for every Relation
Key: SPARK-40938
URL: https://issues.apache.org/jira/browse/SPARK-40938
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-40875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17622510#comment-17622510
]
Rui Wang commented on SPARK-40875:
--
I am working on this.
> Add .agg() to Connect DSL
>
Rui Wang created SPARK-40875:
Summary: Add .agg() to Connect DSL
Key: SPARK-40875
URL: https://issues.apache.org/jira/browse/SPARK-40875
Project: Spark
Issue Type: Sub-task
Components:
[
https://issues.apache.org/jira/browse/SPARK-40839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40839:
-
Summary: [Python] Implement `DataFrame.sample` (was: Implement
`DataFrame.sample`)
> [Python]
Rui Wang created SPARK-40823:
Summary: Connect Proto should carry unparsed identifiers
Key: SPARK-40823
URL: https://issues.apache.org/jira/browse/SPARK-40823
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40774:
Summary: Add Sample to proto and DSL
Key: SPARK-40774
URL: https://issues.apache.org/jira/browse/SPARK-40774
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40780:
Summary: Add WHERE to Connect proto and DSL
Key: SPARK-40780
URL: https://issues.apache.org/jira/browse/SPARK-40780
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-40693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40693:
-
Description:
This is one example:
for SparkConnectTestsPlanOnly, those unit tests access the mock
[
https://issues.apache.org/jira/browse/SPARK-40693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40693:
-
Description:
This is one example:
for SparkConnectTestsPlanOnly, those unit tests access the mock
Rui Wang created SPARK-40693:
Summary: mypy complains accessing the variable defined in the
class method
Key: SPARK-40693
URL: https://issues.apache.org/jira/browse/SPARK-40693
Project: Spark
[
https://issues.apache.org/jira/browse/SPARK-40693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-40693:
-
Description:
This is one example:
for SparkConnectTestsPlanOnly, those unit tests access the remote
Rui Wang created SPARK-40707:
Summary: Add groupby to connect DSL and test more than one
grouping expressions
Key: SPARK-40707
URL: https://issues.apache.org/jira/browse/SPARK-40707
Project: Spark
Rui Wang created SPARK-40743:
Summary: StructType should contain a list of StructField and each
field should have a name
Key: SPARK-40743
URL: https://issues.apache.org/jira/browse/SPARK-40743
Project:
Rui Wang created SPARK-40645:
Summary: Throw exception for Collect() and recommend to use
toPandas()
Key: SPARK-40645
URL: https://issues.apache.org/jira/browse/SPARK-40645
Project: Spark
Issue
Rui Wang created SPARK-40713:
Summary: Improve SET operation support in the proto and the server
Key: SPARK-40713
URL: https://issues.apache.org/jira/browse/SPARK-40713
Project: Spark
Issue
Rui Wang created SPARK-40717:
Summary: Support Column Alias in connect DSL
Key: SPARK-40717
URL: https://issues.apache.org/jira/browse/SPARK-40717
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-40587:
Summary: SELECT * shouldn't be empty project list in proto.
Key: SPARK-40587
URL: https://issues.apache.org/jira/browse/SPARK-40587
Project: Spark
Issue Type:
Rui Wang created SPARK-40055:
Summary: listCatalogs should also return spark_catalog even
spark_catalog implementation is defaultSessionCatalog
Key: SPARK-40055
URL: https://issues.apache.org/jira/browse/SPARK-40055
Rui Wang created SPARK-40989:
Summary: Improve `session.sql` testing coverage in Python client
Key: SPARK-40989
URL: https://issues.apache.org/jira/browse/SPARK-40989
Project: Spark
Issue Type:
Rui Wang created SPARK-40993:
Summary: Migrate markdown style README to
python/docs/development/testing.rst
Key: SPARK-40993
URL: https://issues.apache.org/jira/browse/SPARK-40993
Project: Spark
Rui Wang created SPARK-40992:
Summary: Support toDF(columnNames) in Connect DSL
Key: SPARK-40992
URL: https://issues.apache.org/jira/browse/SPARK-40992
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-41691:
Summary: Implement row count based shuffle read for global limit
Key: SPARK-41691
URL: https://issues.apache.org/jira/browse/SPARK-41691
Project: Spark
Issue Type:
[
https://issues.apache.org/jira/browse/SPARK-41911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-41911:
-
Description: We may need this to help maintain compatibility. Depending on
the concrete protocol
Rui Wang created SPARK-41909:
Summary: Update proto fields to use increasing field numbers and
avoid holes
Key: SPARK-41909
URL: https://issues.apache.org/jira/browse/SPARK-41909
Project: Spark
Rui Wang created SPARK-41908:
Summary: Catalog API refactoring
Key: SPARK-41908
URL: https://issues.apache.org/jira/browse/SPARK-41908
Project: Spark
Issue Type: Sub-task
Components:
[
https://issues.apache.org/jira/browse/SPARK-41910?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-41910:
-
Description: Every field in proto3 has a default value. We should revisit
existing proto field to
Rui Wang created SPARK-41912:
Summary: Subquery should not validate CTE
Key: SPARK-41912
URL: https://issues.apache.org/jira/browse/SPARK-41912
Project: Spark
Issue Type: Bug
Rui Wang created SPARK-41911:
Summary: Add version fields to Connect proto
Key: SPARK-41911
URL: https://issues.apache.org/jira/browse/SPARK-41911
Project: Spark
Issue Type: Sub-task
Rui Wang created SPARK-41910:
Summary: Remove `optional` notation in proto
Key: SPARK-41910
URL: https://issues.apache.org/jira/browse/SPARK-41910
Project: Spark
Issue Type: Sub-task
[
https://issues.apache.org/jira/browse/SPARK-41909?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang resolved SPARK-41909.
--
Resolution: Duplicate
https://issues.apache.org/jira/browse/SPARK-41755
> Update proto fields to use
[
https://issues.apache.org/jira/browse/SPARK-41755?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-41755:
-
Summary: Reorder fields to use consecutive field numbers (was: Reorder the
relation IDs)
> Reorder
[
https://issues.apache.org/jira/browse/SPARK-41908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Rui Wang updated SPARK-41908:
-
Description: We may revisit Catalog proto design and refactor it such that
it becomes a breaking
1 - 100 of 309 matches
Mail list logo