Github user asfgit closed the pull request at:
https://github.com/apache/spark/pull/20678
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r172751054
--- Diff: docs/sql-programming-guide.md ---
@@ -1689,6 +1689,10 @@ using the call `toPandas()` and when creating a
Spark DataFrame from a Pandas Da
Github user felixcheung commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r172751164
--- Diff: docs/sql-programming-guide.md ---
@@ -1800,6 +1800,7 @@ working with timestamps in `pandas_udf`s to get the
best performance, see
## Upgr
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r172540555
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3519,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user BryanCutler commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r172267148
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3519,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user ueshin commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r172087998
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,91 @@ def toPandas(self):
timezone = None
if self.sql_ctx.
Github user viirya commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171155800
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala ---
@@ -1518,7 +1525,9 @@ class SQLConf extends Serializable with Logging {
Github user viirya commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171155732
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala ---
@@ -1518,7 +1525,9 @@ class SQLConf extends Serializable with Logging {
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171139748
--- Diff: docs/sql-programming-guide.md ---
@@ -1689,6 +1689,10 @@ using the call `toPandas()` and when creating a
Spark DataFrame from a Pandas Da
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171138898
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql
Github user holdenk commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r17018
--- Diff: docs/sql-programming-guide.md ---
@@ -1689,6 +1689,10 @@ using the call `toPandas()` and when creating a
Spark DataFrame from a Pandas Da
`cr
Github user holdenk commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r17186
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql_ctx
Github user holdenk commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171110887
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql_ctx
Github user holdenk commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171110674
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql_ctx
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r171089207
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3514,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user BryanCutler commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170993760
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3514,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170910402
--- Diff: docs/sql-programming-guide.md ---
@@ -1800,6 +1800,7 @@ working with timestamps in `pandas_udf`s to get the
best performance, see
## Upgr
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170909693
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,87 @@ def toPandas(self):
timezone = None
if self.sql
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170902707
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3514,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user ueshin commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170813278
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3514,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(data_di
Github user ueshin commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170813132
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala ---
@@ -1068,6 +1068,13 @@ object SQLConf {
.booleanConf
Github user viirya commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170810197
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,87 @@ def toPandas(self):
timezone = None
if self.sql_ctx.
Github user viirya commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170809925
--- Diff: docs/sql-programming-guide.md ---
@@ -1800,6 +1800,7 @@ working with timestamps in `pandas_udf`s to get the
best performance, see
## Upgrading
Github user viirya commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170809560
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,87 @@ def toPandas(self):
timezone = None
if self.sql_ctx.
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170808505
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3514,30 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170799255
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170799189
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3495,42 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170796605
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql
Github user ueshin commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170792103
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql_ctx.
Github user ueshin commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170792766
--- Diff: python/pyspark/sql/session.py ---
@@ -666,8 +666,28 @@ def createDataFrame(self, data, schema=None,
samplingRatio=None, verifySchema=Tr
Github user BryanCutler commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170760936
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala ---
@@ -1068,6 +1068,14 @@ object SQLConf {
.booleanConf
Github user BryanCutler commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170763146
--- Diff: python/pyspark/sql/tests.py ---
@@ -3493,19 +3495,42 @@ def create_pandas_data_frame(self):
data_dict["4_float_t"] = np.float32(da
Github user BryanCutler commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170763497
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql
Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/20678#discussion_r170623237
--- Diff: python/pyspark/sql/dataframe.py ---
@@ -1986,55 +1986,89 @@ def toPandas(self):
timezone = None
if self.sql
GitHub user HyukjinKwon opened a pull request:
https://github.com/apache/spark/pull/20678
[SPARK-23380][PYTHON] Adds a conf for Arrow fallback in
toPandas/createDataFrame with Pandas DataFrame
## What changes were proposed in this pull request?
This PR adds a configuration
35 matches
Mail list logo