[jira] [Commented] (SPARK-38963) Make stage navigable from max metrics displayed in UI

2022-04-20 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-38963?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17524761#comment-17524761
 ] 

pavithra ramachandran commented on SPARK-38963:
---

working on it

> Make stage navigable from max metrics displayed in UI
> -
>
> Key: SPARK-38963
> URL: https://issues.apache.org/jira/browse/SPARK-38963
> Project: Spark
>  Issue Type: Improvement
>  Components: Web UI
>Affects Versions: 3.1.1, 3.2.0, 3.3.0
>Reporter: pavithra ramachandran
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-38963) Make stage navigable from max metrics displayed in UI

2022-04-20 Thread pavithra ramachandran (Jira)
pavithra ramachandran created SPARK-38963:
-

 Summary: Make stage navigable from max metrics displayed in UI
 Key: SPARK-38963
 URL: https://issues.apache.org/jira/browse/SPARK-38963
 Project: Spark
  Issue Type: Improvement
  Components: Web UI
Affects Versions: 3.2.0, 3.1.1, 3.3.0
Reporter: pavithra ramachandran






--
This message was sent by Atlassian Jira
(v8.20.7#820007)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-35877) Spark Protobuf jar has CVE issue CVE-2015-5237

2021-06-24 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-35877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17368810#comment-17368810
 ] 

pavithra ramachandran commented on SPARK-35877:
---

i ll raise PR soon

> Spark Protobuf jar has CVE issue CVE-2015-5237
> --
>
> Key: SPARK-35877
> URL: https://issues.apache.org/jira/browse/SPARK-35877
> Project: Spark
>  Issue Type: Bug
>  Components: Security, Spark Core
>Affects Versions: 2.4.5, 3.1.1
>Reporter: jobit mathew
>Priority: Minor
>
> Spark Protobuf jar has CVE issue CVE-2015-5237



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-35835) Select filter query on table with struct complex type fails

2021-06-20 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-35835?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17366386#comment-17366386
 ] 

pavithra ramachandran commented on SPARK-35835:
---

i shall raise a PR soon

> Select filter query on table with struct complex type fails
> ---
>
> Key: SPARK-35835
> URL: https://issues.apache.org/jira/browse/SPARK-35835
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.1
> Environment: Spark 3.1.1
>Reporter: Chetan Bhat
>Priority: Minor
>
> [Steps]:-
> From Spark beeline create a parquet or ORC table having complex type data. 
> Load data in the table and execute select filter query.
> 0: jdbc:hive2://vm2:22550/> create table Struct_com (CUST_ID string, YEAR 
> int, MONTH int, AGE int, GENDER string, EDUCATED string, IS_MARRIED string, 
> STRUCT_INT_DOUBLE_STRING_DATE 
> struct,CARD_COUNT 
> int,DEBIT_COUNT int, CREDIT_COUNT int, DEPOSIT double, HQ_DEPOSIT double) 
> stored as parquet;
> +-+
> | Result |
> +-+
> +-+
> No rows selected (0.161 seconds)
> 0: jdbc:hive2://vm2:22550/> LOAD DATA INPATH 
> 'hdfs://hacluster/chetan/Struct.csv' OVERWRITE INTO TABLE Struct_com;
> +-+
> | Result |
> +-+
> +-+
> No rows selected (1.09 seconds)
> 0: jdbc:hive2://vm2:22550/> SELECT 
> struct_int_double_string_date.COUNTRY,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.CHECK_DATE,struct_in
>  t_double_string_date.Country, SUM(struct_int_double_string_date.id) AS Sum 
> FROM (select * from Struct_com) SUB_QRY WHERE 
> struct_int_double_string_date.id > 5700 GRO UP BY 
> struct_int_double_string_date.COUNTRY,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.Country
>  ORDER BY struct_int_double_string_date.COUNTRY 
> asc,struct_int_double_string_date.CHECK_DATE 
> asc,struct_int_double_string_date.CHECK_DATE asc, struct_int_double_stri 
> ng_date.Country asc;
>  
> [Actual Issue] : - Select filter query on table with struct complex type fails
> 0: jdbc:hive2://vm2:22550/> SELECT 
> struct_int_double_string_date.COUNTRY,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.CHECK_DATE,struct_in
>  t_double_string_date.Country, SUM(struct_int_double_string_date.id) AS Sum 
> FROM (select * from Struct_com) SUB_QRY WHERE 
> struct_int_double_string_date.id > 5700 GRO UP BY 
> struct_int_double_string_date.COUNTRY,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.CHECK_DATE,struct_int_double_string_date.Country
>  ORDER BY struct_int_double_string_date.COUNTRY 
> asc,struct_int_double_string_date.CHECK_DATE 
> asc,struct_int_double_string_date.CHECK_DATE asc, struct_int_double_stri 
> ng_date.Country asc;
> Error: org.apache.hive.service.cli.HiveSQLException: Error running query: 
> org.apache.spark.sql.catalyst.errors.package$TreeNodeException: execute, tree:
> Exchange rangepartitioning(COUNTRY#139896 ASC NULLS FIRST, CHECK_DATE#139897 
> ASC NULLS FIRST, CHECK_DATE#139897 ASC NULLS FIRST, COUNTRY#139896 ASC NULLS 
> FIRST, 200 ), ENSURE_REQUIREMENTS, [id=#17161]
> +- *(2) HashAggregate(keys=[_gen_alias_139928#139928, 
> _gen_alias_139929#139929], functions=[sum(cast(_gen_alias_139931#139931 as 
> bigint))], output=[COUNTRY#139896, CHECK_DATE#139897, CHECK_DATE#139898, 
> Country#139899, Sum#139877L])
> +- Exchange hashpartitioning(_gen_alias_139928#139928, 
> _gen_alias_139929#139929, 200), ENSURE_REQUIREMENTS, [id=#17157]
> +- *(1) HashAggregate(keys=[_gen_alias_139928#139928, 
> _gen_alias_139929#139929], 
> functions=[partial_sum(cast(_gen_alias_139931#139931 as bigint))], output=[_g 
> en_alias_139928#139928, _gen_alias_139929#139929, sum#139934L])
> +- *(1) Project [STRUCT_INT_DOUBLE_STRING_DATE#139885.COUNTRY AS 
> _gen_alias_139928#139928, STRUCT_INT_DOUBLE_STRING_DATE#139885.CHECK_DATE AS 
> _gen_alias_13 9929#139929, STRUCT_INT_DOUBLE_STRING_DATE#139885.COUNTRY AS 
> _gen_alias_139930#139930, STRUCT_INT_DOUBLE_STRING_DATE#139885.ID AS 
> _gen_alias_139931#139931]
> +- *(1) Filter (isnotnull(STRUCT_INT_DOUBLE_STRING_DATE#139885) AND 
> (STRUCT_INT_DOUBLE_STRING_DATE#139885.ID > 5700))
> +- FileScan parquet default.struct_com[STRUCT_INT_DOUBLE_STRING_DATE#139885] 
> Batched: false, DataFilters: [isnotnull(STRUCT_INT_DOUBLE_STRING_DATE#13 
> 9885), (STRUCT_INT_DOUBLE_STRING_DATE#139885.ID > 5700)], Format: Parquet, 
> Location: InMemoryFileIndex[hdfs://hacluster/user/hive/warehouse/struct_com], 
> PartitionFi lters: [], PushedFilters: 
> [IsNotNull(STRUCT_INT_DOUBLE_STRING_DATE), 
> GreaterThan(STRUCT_INT_DOUBLE_STRING_DATE.ID,5700)], ReadSchema: 
> struct G_DATE:struct>
> at 
> 

[jira] [Commented] (SPARK-32351) Partially pushed partition filters are not explained

2020-07-21 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17162477#comment-17162477
 ] 

pavithra ramachandran commented on SPARK-32351:
---

i would like to check this

> Partially pushed partition filters are not explained
> 
>
> Key: SPARK-32351
> URL: https://issues.apache.org/jira/browse/SPARK-32351
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Yuming Wang
>Priority: Major
>
> How to reproduce this issue:
> {code:scala}
> spark.sql(
>   s"""
>  |CREATE TABLE t(i INT, p STRING)
>  |USING parquet
>  |PARTITIONED BY (p)""".stripMargin)
> spark.range(0, 1000).selectExpr("id as col").createOrReplaceTempView("temp")
> for (part <- Seq(1, 2, 3, 4)) {
>   sql(s"""
>  |INSERT OVERWRITE TABLE t PARTITION (p='$part')
>  |SELECT col FROM temp""".stripMargin)
> }
> spark.sql("SELECT * FROM t WHERE  WHERE (p = '1' AND i = 1) OR (p = '2' and i 
> = 2)").explain
> {code}
> We have pushed down {{p = '1' or p = '2'}} since SPARK-28169, but this pushed 
> down filter not in explain
> {noformat}
> == Physical Plan ==
> *(1) Filter (((p#21 = 1) AND (i#20 = 1)) OR ((p#21 = 2) AND (i#20 = 2)))
> +- *(1) ColumnarToRow
>+- FileScan parquet default.t[i#20,p#21] Batched: true, DataFilters: [], 
> Format: Parquet, Location: 
> InMemoryFileIndex[file:/Users/yumwang/spark/SPARK-32289/sql/core/spark-warehouse/org.apache.spark...,
>  PartitionFilters: [], PushedFilters: [], ReadSchema: struct
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-32325) JSON predicate pushdown for nested fields

2020-07-15 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32325?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17158878#comment-17158878
 ] 

pavithra ramachandran commented on SPARK-32325:
---

i would like to work on this

> JSON predicate pushdown for nested fields
> -
>
> Key: SPARK-32325
> URL: https://issues.apache.org/jira/browse/SPARK-32325
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Maxim Gekk
>Priority: Major
>
> SPARK-30648 should support filters pushdown to JSON datasource but it 
> supports only filters that refer to top-level fields. The ticket aims to 
> support nested fields as well. See the needed changes: 
> https://github.com/apache/spark/pull/27366#discussion_r443340603



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-32328) Avro predicate pushdown for nested fields

2020-07-15 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32328?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17158877#comment-17158877
 ] 

pavithra ramachandran commented on SPARK-32328:
---

i would like to work on this

> Avro predicate pushdown for nested fields
> -
>
> Key: SPARK-32328
> URL: https://issues.apache.org/jira/browse/SPARK-32328
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: jobit mathew
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-32322) Pyspark not launching in Spark IPV6 environment

2020-07-15 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17158165#comment-17158165
 ] 

pavithra ramachandran commented on SPARK-32322:
---

i would like to check this

> Pyspark not launching in Spark IPV6 environment
> ---
>
> Key: SPARK-32322
> URL: https://issues.apache.org/jira/browse/SPARK-32322
> Project: Spark
>  Issue Type: Bug
>  Components: PySpark
>Affects Versions: 3.1.0
>Reporter: jobit mathew
>Priority: Minor
>
> pyspark  is not launching in Spark IPV6 environment.
> Initial analysis looks like python is not supporting IPV6.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Issue Comment Deleted] (SPARK-31887) Date casting to string is giving wrong value

2020-06-26 Thread pavithra ramachandran (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-31887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

pavithra ramachandran updated SPARK-31887:
--
Comment: was deleted

(was: i would like to check this.)

> Date casting to string is giving wrong value
> 
>
> Key: SPARK-31887
> URL: https://issues.apache.org/jira/browse/SPARK-31887
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.4.5
> Environment: The spark is running on cluster mode with Mesos.
>  
> Mesos agents are dockerised running on Ubuntu 18.
>  
> Timezone setting of docker instance: UTC
> Timezone of server hosting docker: America/New_York
> Timezone of driver machine: America/New_York
>Reporter: Amit Gupta
>Priority: Major
>
> The code converts the string to date and then write it in csv.
> {code:java}
> val x = Seq(("2020-02-19", "2020-02-19 05:11:00")).toDF("a", 
> "b").select('a.cast("date"), 'b.cast("timestamp"))
> x.show()
> +--+---+
> | a|  b|
> +--+---+
> |2020-02-19|2020-02-19 05:11:00|
> +--+---+
> x.write.mode("overwrite").option("header", true).csv("/tmp/test1.csv")
> {code}
>  
> The date written in CSV file is different:
> {code:java}
> > snakebite cat "/tmp/test1.csv/*.csv"
> a,b
> 2020-02-18,2020-02-19T05:11:00.000Z{code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-32103) Spark support IPV6 in yarn mode

2020-06-26 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17146073#comment-17146073
 ] 

pavithra ramachandran commented on SPARK-32103:
---

i shall raise a pr soon

> Spark support IPV6 in yarn mode
> ---
>
> Key: SPARK-32103
> URL: https://issues.apache.org/jira/browse/SPARK-32103
> Project: Spark
>  Issue Type: Bug
>  Components: YARN
>Affects Versions: 3.1.0
>Reporter: jobit mathew
>Priority: Minor
>
> Spark support IPV6 in yarn mode
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-31908) Spark UI shows wrong driver memory configurations if driver memory is provided at run time

2020-06-05 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17126419#comment-17126419
 ] 

pavithra ramachandran commented on SPARK-31908:
---

i would like to work on this

> Spark UI shows wrong driver memory configurations if driver memory is 
> provided at run time
> --
>
> Key: SPARK-31908
> URL: https://issues.apache.org/jira/browse/SPARK-31908
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.4.3
>Reporter: Rahul Kumar
>Priority: Major
>
> I discovered that in cluster mode , when driver memory is provided via 
> spark.driver.memory configuration at run time after creating spark session, 
> spark doesn't pick this configurations at run time as application master is 
> already launched by that time and picks the default spark driver memory 
> configuration(1GB). 
> However, on spark UI page, in environment tab, it still shows driver memory 
> as the value passed via configurations at run time, which makes identifying 
> and debugging this scenario more difficult. Driver memory should be shown as 
> the value which spark is actually using in job.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-31887) Date casting to string is giving wrong value

2020-06-02 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31887?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17123835#comment-17123835
 ] 

pavithra ramachandran commented on SPARK-31887:
---

i would like to check this.

> Date casting to string is giving wrong value
> 
>
> Key: SPARK-31887
> URL: https://issues.apache.org/jira/browse/SPARK-31887
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.4.5
> Environment: The spark is running on cluster mode with Mesos.
>  
> Mesos agents are dockerised running on Ubuntu 18.
>  
> Timezone setting of docker instance: UTC
> Timezone of server hosting docker: America/New_York
> Timezone of driver machine: America/New_York
>Reporter: Amit Gupta
>Priority: Major
>
> The code converts the string to date and then write it in csv.
> {code:java}
> val x = Seq(("2020-02-19", "2020-02-19 05:11:00")).toDF("a", 
> "b").select('a.cast("date"), 'b.cast("timestamp"))
> x.show()
> +--+---+
> | a|  b|
> +--+---+
> |2020-02-19|2020-02-19 05:11:00|
> +--+---+
> x.write.mode("overwrite").option("header", true).csv("/tmp/test1.csv")
> {code}
>  
> The date written in CSV file is different:
> {code:java}
> > snakebite cat "/tmp/test1.csv/*.csv"
> a,b
> 2020-02-18,2020-02-19T05:11:00.000Z{code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-31634) "show tables like" support for SQL wildcard characters (% and _)

2020-05-11 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17104223#comment-17104223
 ] 

pavithra ramachandran commented on SPARK-31634:
---

[~yumwang] i see that show tables uses catalogue and there is an open Jira in 
hive side. Once that gets fixed, it will work in spark, Or do u want us to 
handle separately handle in spark. 

> "show tables like" support for SQL wildcard characters (% and _)
> 
>
> Key: SPARK-31634
> URL: https://issues.apache.org/jira/browse/SPARK-31634
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Yuming Wang
>Priority: Major
>
> https://docs.snowflake.com/en/sql-reference/sql/show-tables.html
> https://clickhouse.tech/docs/en/sql-reference/statements/show/
> https://www.mysqltutorial.org/mysql-show-tables/
> https://issues.apache.org/jira/browse/HIVE-23359



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30168) Eliminate warnings in Parquet datasource

2020-05-09 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17103181#comment-17103181
 ] 

pavithra ramachandran commented on SPARK-30168:
---

[~maxgekk] I would like to work this, i shall raise the PR soon

> Eliminate warnings in Parquet datasource
> 
>
> Key: SPARK-30168
> URL: https://issues.apache.org/jira/browse/SPARK-30168
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Maxim Gekk
>Priority: Minor
>
> # 
> sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/parquet/ParquetPartitionReaderFactory.scala
> {code}
> Warning:Warning:line (120)class ParquetInputSplit in package hadoop is 
> deprecated: see corresponding Javadoc for more information.
>   Option[TimeZone]) => RecordReader[Void, T]): RecordReader[Void, T] 
> = {
> Warning:Warning:line (125)class ParquetInputSplit in package hadoop is 
> deprecated: see corresponding Javadoc for more information.
>   new org.apache.parquet.hadoop.ParquetInputSplit(
> Warning:Warning:line (134)method readFooter in class ParquetFileReader is 
> deprecated: see corresponding Javadoc for more information.
>   ParquetFileReader.readFooter(conf, filePath, 
> SKIP_ROW_GROUPS).getFileMetaData
> Warning:Warning:line (183)class ParquetInputSplit in package hadoop is 
> deprecated: see corresponding Javadoc for more information.
>   split: ParquetInputSplit,
> Warning:Warning:line (212)class ParquetInputSplit in package hadoop is 
> deprecated: see corresponding Javadoc for more information.
>   split: ParquetInputSplit,
> {code}
> # 
> sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/SpecificParquetRecordReaderBase.java
> {code}
> Warning:Warning:line (55)java: org.apache.parquet.hadoop.ParquetInputSplit in 
> org.apache.parquet.hadoop has been deprecated
> Warning:Warning:line (95)java: 
> org.apache.parquet.hadoop.ParquetInputSplit in org.apache.parquet.hadoop has 
> been deprecated
> Warning:Warning:line (95)java: 
> org.apache.parquet.hadoop.ParquetInputSplit in org.apache.parquet.hadoop has 
> been deprecated
> Warning:Warning:line (97)java: getRowGroupOffsets() in 
> org.apache.parquet.hadoop.ParquetInputSplit has been deprecated
> Warning:Warning:line (105)java: 
> readFooter(org.apache.hadoop.conf.Configuration,org.apache.hadoop.fs.Path,org.apache.parquet.format.converter.ParquetMetadataConverter.MetadataFilter)
>  in org.apache.parquet.hadoop.ParquetFileReader has been deprecated
> Warning:Warning:line (108)java: 
> filterRowGroups(org.apache.parquet.filter2.compat.FilterCompat.Filter,java.util.List,org.apache.parquet.schema.MessageType)
>  in org.apache.parquet.filter2.compat.RowGroupFilter has been deprecated
> Warning:Warning:line (111)java: 
> readFooter(org.apache.hadoop.conf.Configuration,org.apache.hadoop.fs.Path,org.apache.parquet.format.converter.ParquetMetadataConverter.MetadataFilter)
>  in org.apache.parquet.hadoop.ParquetFileReader has been deprecated
> Warning:Warning:line (147)java: 
> ParquetFileReader(org.apache.hadoop.conf.Configuration,org.apache.parquet.hadoop.metadata.FileMetaData,org.apache.hadoop.fs.Path,java.util.List,java.util.List)
>  in org.apache.parquet.hadoop.ParquetFileReader has been deprecated
> Warning:Warning:line (203)java: 
> readFooter(org.apache.hadoop.conf.Configuration,org.apache.hadoop.fs.Path,org.apache.parquet.format.converter.ParquetMetadataConverter.MetadataFilter)
>  in org.apache.parquet.hadoop.ParquetFileReader has been deprecated
> Warning:Warning:line (226)java: 
> ParquetFileReader(org.apache.hadoop.conf.Configuration,org.apache.parquet.hadoop.metadata.FileMetaData,org.apache.hadoop.fs.Path,java.util.List,java.util.List)
>  in org.apache.parquet.hadoop.ParquetFileReader has been deprecated
> {code}
> # 
> sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetCompatibilityTest.scala
> # 
> sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetInteroperabilitySuite.scala
> # 
> sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTest.scala
> # 
> sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveDDLSuite.scala



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-31634) "show tables like" support for SQL wildcard characters (% and _)

2020-05-05 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17099984#comment-17099984
 ] 

pavithra ramachandran commented on SPARK-31634:
---

I would like to work on this

> "show tables like" support for SQL wildcard characters (% and _)
> 
>
> Key: SPARK-31634
> URL: https://issues.apache.org/jira/browse/SPARK-31634
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Yuming Wang
>Priority: Major
>
> https://docs.snowflake.com/en/sql-reference/sql/show-tables.html
> https://clickhouse.tech/docs/en/sql-reference/statements/show/
> https://www.mysqltutorial.org/mysql-show-tables/
> https://issues.apache.org/jira/browse/HIVE-23359



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Comment Edited] (SPARK-31634) "show tables like" support for SQL wildcard characters (% and _)

2020-05-05 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17099984#comment-17099984
 ] 

pavithra ramachandran edited comment on SPARK-31634 at 5/5/20, 3:06 PM:


I would like to work on this. [~yumwang]  can I ?


was (Author: pavithraramachandran):
I would like to work on this

> "show tables like" support for SQL wildcard characters (% and _)
> 
>
> Key: SPARK-31634
> URL: https://issues.apache.org/jira/browse/SPARK-31634
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Yuming Wang
>Priority: Major
>
> https://docs.snowflake.com/en/sql-reference/sql/show-tables.html
> https://clickhouse.tech/docs/en/sql-reference/statements/show/
> https://www.mysqltutorial.org/mysql-show-tables/
> https://issues.apache.org/jira/browse/HIVE-23359



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30687) When reading from a file with pre-defined schema and encountering a single value that is not the same type as that of its column , Spark nullifies the entire row

2020-02-04 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30687?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17029720#comment-17029720
 ] 

pavithra ramachandran commented on SPARK-30687:
---

yes. Issue is present 2.4.x also.

> When reading from a file with pre-defined schema and encountering a single 
> value that is not the same type as that of its column , Spark nullifies the 
> entire row
> -
>
> Key: SPARK-30687
> URL: https://issues.apache.org/jira/browse/SPARK-30687
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.3.0
>Reporter: Bao Nguyen
>Priority: Major
>
> When reading from a file with pre-defined schema and encountering a single 
> value that is not the same type as that of its column , Spark nullifies the 
> entire row instead of setting the value at that cell to be null.
>  
> {code:java}
> case class TestModel(
>   num: Double, test: String, mac: String, value: Double
> )
> val schema = 
> ScalaReflection.schemaFor[TestModel].dataType.asInstanceOf[StructType]
> //here's the content of the file test.data
> //1~test~mac1~2
> //1.0~testdatarow2~mac2~non-numeric
> //2~test1~mac1~3
> val ds = spark
>   .read
>   .schema(schema)
>   .option("delimiter", "~")
>   .csv("/test-data/test.data")
> ds.show();
> //the content of data frame. second row is all null. 
> //  ++-++-+
> //  | num| test| mac|value|
> //  ++-++-+
> //  | 1.0| test|mac1|  2.0|
> //  |null| null|null| null|
> //  | 2.0|test1|mac1|  3.0|
> //  ++-++-+
> //should be
> // ++--++-+ 
> // | num| test | mac|value| 
> // ++--++-+ 
> // | 1.0| test |mac1| 2.0 | 
> // |1.0 |testdatarow2  |mac2| null| 
> // | 2.0|test1 |mac1| 3.0 | 
> // ++--++-+{code}
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30687) When reading from a file with pre-defined schema and encountering a single value that is not the same type as that of its column , Spark nullifies the entire row

2020-01-31 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30687?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17027333#comment-17027333
 ] 

pavithra ramachandran commented on SPARK-30687:
---

l would like to work on this issue.

> When reading from a file with pre-defined schema and encountering a single 
> value that is not the same type as that of its column , Spark nullifies the 
> entire row
> -
>
> Key: SPARK-30687
> URL: https://issues.apache.org/jira/browse/SPARK-30687
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.3.0
>Reporter: Bao Nguyen
>Priority: Major
>
> When reading from a file with pre-defined schema and encountering a single 
> value that is not the same type as that of its column , Spark nullifies the 
> entire row instead of setting the value at that cell to be null.
>  
> {code:java}
> case class TestModel(
>   num: Double, test: String, mac: String, value: Double
> )
> val schema = 
> ScalaReflection.schemaFor[TestModel].dataType.asInstanceOf[StructType]
> //here's the content of the file test.data
> //1~test~mac1~2
> //1.0~testdatarow2~mac2~non-numeric
> //2~test1~mac1~3
> val ds = spark
>   .read
>   .schema(schema)
>   .option("delimiter", "~")
>   .csv("/test-data/test.data")
> ds.show();
> //the content of data frame. second row is all null. 
> //  ++-++-+
> //  | num| test| mac|value|
> //  ++-++-+
> //  | 1.0| test|mac1|  2.0|
> //  |null| null|null| null|
> //  | 2.0|test1|mac1|  3.0|
> //  ++-++-+
> //should be
> // ++--++-+ 
> // | num| test | mac|value| 
> // ++--++-+ 
> // | 1.0| test |mac1| 2.0 | 
> // |1.0 |testdatarow2  |mac2| null| 
> // | 2.0|test1 |mac1| 3.0 | 
> // ++--++-+{code}
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30483) Job History does not show pool properties table

2020-01-16 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17016873#comment-17016873
 ] 

pavithra ramachandran commented on SPARK-30483:
---

Issue is resolved in master and 2.4

 

[https://github.com/apache/spark/commit/6d90298438e627187088a5d8c53d470646d051f4]

> Job History does not show pool properties table
> ---
>
> Key: SPARK-30483
> URL: https://issues.apache.org/jira/browse/SPARK-30483
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> Stage will show the Pool Name column but when user clicks the hyper link Name>  it will not redirect to Pool Properties Table.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30484) Job History Storage Tab does not display RDD Table

2020-01-10 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17012656#comment-17012656
 ] 

pavithra ramachandran commented on SPARK-30484:
---

i shall work on this

> Job History Storage Tab does not display RDD Table
> --
>
> Key: SPARK-30484
> URL: https://issues.apache.org/jira/browse/SPARK-30484
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> scala> import org.apache.spark.storage.StorageLevel._
> import org.apache.spark.storage.StorageLevel._
> scala> val rdd = sc.range(0, 100, 1, 5).setName("rdd")
> rdd: org.apache.spark.rdd.RDD[Long] = rdd MapPartitionsRDD[1] at range at 
> :27
> scala> rdd.persist(MEMORY_ONLY_SER)
> res0: rdd.type = rdd MapPartitionsRDD[1] at range at :27
> scala> rdd.count
> res1: Long = 100  
>   
> scala> val df = Seq((1, "andy"), (2, "bob"), (2, "andy")).toDF("count", 
> "name")
> df: org.apache.spark.sql.DataFrame = [count: int, name: string]
> scala> df.persist(DISK_ONLY)
> res2: df.type = [count: int, name: string]
> scala> df.count
> res3: Long = 3
> Open Storage Tab under Incomplete Jobs in Job History Page
> UI will not display the RDD Table.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30483) Job History does not show pool properties table

2020-01-10 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17012654#comment-17012654
 ] 

pavithra ramachandran commented on SPARK-30483:
---

i shall work on this

> Job History does not show pool properties table
> ---
>
> Key: SPARK-30483
> URL: https://issues.apache.org/jira/browse/SPARK-30483
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> Stage will show the Pool Name column but when user clicks the hyper link Name>  it will not redirect to Pool Properties Table.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-27148) Support CURRENT_TIME and LOCALTIME when ANSI mode enabled

2020-01-10 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-27148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17012620#comment-17012620
 ] 

pavithra ramachandran commented on SPARK-27148:
---

[~maropu] I would like to work on this..

> Support CURRENT_TIME and LOCALTIME when ANSI mode enabled
> -
>
> Key: SPARK-27148
> URL: https://issues.apache.org/jira/browse/SPARK-27148
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: Takeshi Yamamuro
>Priority: Major
>
> CURRENT_TIME and LOCALTIME should be supported in the ANSI standard;
> {code:java}
> postgres=# select CURRENT_TIME;
>        timetz       
> 
> 16:45:43.398109+09
> (1 row)
> postgres=# select LOCALTIME;
>       time      
> 
> 16:45:48.60969
> (1 row){code}
> Before this, we need to support TIME types (java.sql.Time).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30454) Null Dereference in HiveSQLException

2020-01-07 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17010412#comment-17010412
 ] 

pavithra ramachandran commented on SPARK-30454:
---

I shall raise the PR

> Null Dereference in HiveSQLException
> 
>
> Key: SPARK-30454
> URL: https://issues.apache.org/jira/browse/SPARK-30454
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.3.4, 2.4.4, 3.0.0
>Reporter: pavithra ramachandran
>Priority: Major
>
> Null Pointer DeReferencing  found in spark HiveSQLException code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-30454) Null Dereference in HiveSQLException

2020-01-07 Thread pavithra ramachandran (Jira)
pavithra ramachandran created SPARK-30454:
-

 Summary: Null Dereference in HiveSQLException
 Key: SPARK-30454
 URL: https://issues.apache.org/jira/browse/SPARK-30454
 Project: Spark
  Issue Type: Bug
  Components: SQL
Affects Versions: 2.4.4, 2.3.4, 3.0.0
Reporter: pavithra ramachandran


Null Pointer DeReferencing  found in spark HiveSQLException code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28825) Document EXPLAIN Statement in SQL Reference.

2019-12-19 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17000593#comment-17000593
 ] 

pavithra ramachandran commented on SPARK-28825:
---

[~LI,Xiao] and [~dkbiswal] i have the PR ready. Can I handle this?

> Document EXPLAIN Statement in SQL Reference.
> 
>
> Key: SPARK-28825
> URL: https://issues.apache.org/jira/browse/SPARK-28825
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: jobit mathew
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29505) desc extended is case sensitive

2019-12-16 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16997090#comment-16997090
 ] 

pavithra ramachandran commented on SPARK-29505:
---

i will work on this

> desc extended   is case sensitive
> --
>
> Key: SPARK-29505
> URL: https://issues.apache.org/jira/browse/SPARK-29505
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> {code}
> create table customer(id int, name String, *CName String*, address String, 
> city String, pin int, country String);
> insert into customer values(1,'Alfred','Maria','Obere Str 
> 57','Berlin',12209,'Germany');
> insert into customer values(2,'Ana','trujilo','Adva de la','Maxico 
> D.F.',05021,'Maxico');
> insert into customer values(3,'Antonio','Antonio Moreno','Mataderos 
> 2312','Maxico D.F.',05023,'Maxico');
> analyze table customer compute statistics for columns cname; – *Success( 
> Though cname is not as CName)*
> desc extended customer cname; – Failed
> jdbc:hive2://10.18.19.208:23040/default> desc extended customer *cname;*
> +-+-+
> | info_name | info_value |
> +-+-+
> | col_name | cname |
> | data_type | string |
> | comment | NULL |
> | min | NULL |
> | max | NULL |
> | num_nulls | NULL |
> | distinct_count | NULL |
> | avg_col_len | NULL |
> | max_col_len | NULL |
> | histogram | NULL |
> +-+--
> {code}
>  
> But 
> {code}
> desc extended customer CName; – SUCCESS
> 0: jdbc:hive2://10.18.19.208:23040/default> desc extended customer *CName;*
> +-+-+
> | info_name | info_value |
> +-+-+
> | col_name | CName |
> | data_type | string |
> | comment | NULL |
> | min | NULL |
> | max | NULL |
> | num_nulls | 0 |
> | distinct_count | 3 |
> | avg_col_len | 9 |
> | max_col_len | 14 |
> | histogram | NULL |
> +-+-+
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28825) Document EXPLAIN Statement in SQL Reference.

2019-12-13 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16995531#comment-16995531
 ] 

pavithra ramachandran commented on SPARK-28825:
---

[~dkbiswal] are you working on this? If not I would like to work on this.

> Document EXPLAIN Statement in SQL Reference.
> 
>
> Key: SPARK-28825
> URL: https://issues.apache.org/jira/browse/SPARK-28825
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: jobit mathew
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29455) Improve tooltip information for Stages Tab

2019-12-12 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29455?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16994497#comment-16994497
 ] 

pavithra ramachandran commented on SPARK-29455:
---

[~srowen] thank you.. i see sharan has already submitted the patch .

> Improve tooltip information for Stages Tab
> --
>
> Key: SPARK-29455
> URL: https://issues.apache.org/jira/browse/SPARK-29455
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30233) Spark WebUI task table indentation issue

2019-12-12 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16994446#comment-16994446
 ] 

pavithra ramachandran commented on SPARK-30233:
---

its resolved in master using SPARK-28430

[https://github.com/apache/spark/pull/25183]

> Spark WebUI task table indentation  issue
> -
>
> Key: SPARK-30233
> URL: https://issues.apache.org/jira/browse/SPARK-30233
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.3.4
>Reporter: jobit mathew
>Priority: Minor
> Attachments: sparkopensourceissue.PNG
>
>
> !sparkopensourceissue.PNG!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Issue Comment Deleted] (SPARK-30233) Spark WebUI task table indentation issue

2019-12-12 Thread pavithra ramachandran (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-30233?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

pavithra ramachandran updated SPARK-30233:
--
Comment: was deleted

(was: i shall raise PR soon)

> Spark WebUI task table indentation  issue
> -
>
> Key: SPARK-30233
> URL: https://issues.apache.org/jira/browse/SPARK-30233
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 2.3.4
>Reporter: jobit mathew
>Priority: Minor
> Attachments: sparkopensourceissue.PNG
>
>
> !sparkopensourceissue.PNG!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30233) Spark WebUI task table indentation issue

2019-12-12 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30233?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16994426#comment-16994426
 ] 

pavithra ramachandran commented on SPARK-30233:
---

i shall raise PR soon

> Spark WebUI task table indentation  issue
> -
>
> Key: SPARK-30233
> URL: https://issues.apache.org/jira/browse/SPARK-30233
> Project: Spark
>  Issue Type: Bug
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: jobit mathew
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-30171) Eliminate warnings: part2

2019-12-08 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-30171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16991164#comment-16991164
 ] 

pavithra ramachandran commented on SPARK-30171:
---

i will work on this

> Eliminate warnings: part2
> -
>
> Key: SPARK-30171
> URL: https://issues.apache.org/jira/browse/SPARK-30171
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
> AvroFunctionsSuite.scala
> Warning:Warning:line (41)method to_avro in package avro is deprecated (since 
> 3.0.0): Please use 'org.apache.spark.sql.avro.functions.to_avro' instead.
> val avroDF = df.select(to_avro('id).as("a"), to_avro('str).as("b"))
> Warning:Warning:line (41)method to_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.to_avro' 
> instead.
> val avroDF = df.select(to_avro('id).as("a"), to_avro('str).as("b"))
> Warning:Warning:line (54)method from_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.from_avro' 
> instead.
> checkAnswer(avroDF.select(from_avro('a, avroTypeLong), from_avro('b, 
> avroTypeStr)), df)
> Warning:Warning:line (54)method from_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.from_avro' 
> instead.
> checkAnswer(avroDF.select(from_avro('a, avroTypeLong), from_avro('b, 
> avroTypeStr)), df)
> Warning:Warning:line (59)method to_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.to_avro' 
> instead.
> val avroStructDF = df.select(to_avro('struct).as("avro"))
> Warning:Warning:line (70)method from_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.from_avro' 
> instead.
> checkAnswer(avroStructDF.select(from_avro('avro, avroTypeStruct)), df)
> Warning:Warning:line (76)method to_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.to_avro' 
> instead.
> val avroStructDF = df.select(to_avro('struct).as("avro"))
> Warning:Warning:line (118)method to_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.to_avro' 
> instead.
> val readBackOne = dfOne.select(to_avro($"array").as("avro"))
> Warning:Warning:line (119)method from_avro in package avro is deprecated 
> (since 3.0.0): Please use 'org.apache.spark.sql.avro.functions.from_avro' 
> instead.
>   .select(from_avro($"avro", avroTypeArrStruct).as("array"))
> AvroPartitionReaderFactory.scala
> Warning:Warning:line (64)value ignoreExtension in class AvroOptions is 
> deprecated (since 3.0): Use the general data source option pathGlobFilter for 
> filtering file names
> if (parsedOptions.ignoreExtension || 
> partitionedFile.filePath.endsWith(".avro")) {
> AvroFileFormat.scala
> Warning:Warning:line (98)value ignoreExtension in class AvroOptions is 
> deprecated (since 3.0): Use the general data source option pathGlobFilter for 
> filtering file names
>   if (parsedOptions.ignoreExtension || file.filePath.endsWith(".avro")) {
> AvroUtils.scala
> Warning:Warning:line (55)value ignoreExtension in class AvroOptions is 
> deprecated (since 3.0): Use the general data source option pathGlobFilter for 
> filtering file names
> inferAvroSchemaFromFiles(files, conf, parsedOptions.ignoreExtension,



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29841) PostgreSQL dialect: cast to date

2019-11-11 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16971470#comment-16971470
 ] 

pavithra ramachandran commented on SPARK-29841:
---

i will check

> PostgreSQL dialect: cast to date
> 
>
> Key: SPARK-29841
> URL: https://issues.apache.org/jira/browse/SPARK-29841
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: jobit mathew
>Priority: Minor
>
> Make SparkSQL's cast to date behavior be consistent with PostgreSQL when
> spark.sql.dialect is configured as PostgreSQL.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29784) Built in function trim is not compatible in 3.0 with previous version

2019-11-06 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16968949#comment-16968949
 ] 

pavithra ramachandran commented on SPARK-29784:
---

i shall work on this

> Built in function trim is not compatible in 3.0 with previous version
> -
>
> Key: SPARK-29784
> URL: https://issues.apache.org/jira/browse/SPARK-29784
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> SELECT trim('SL', 'SSparkSQLS'); returns empty in Spark 3.0 where as in 2.4 
> and 2.3.2 is returning after leading and trailing character removed.
> Spark 3.0 – Not correct
> jdbc:hive2://10.18.19.208:23040/default> SELECT trim('SL', 'SSparkSQLS');
> +---+
> | trim(SL, SSparkSQLS) |
> +---+
> | |
> +---
> Spark 2.4 – Correct
> jdbc:hive2://10.18.18.214:23040/default> SELECT trim('SL', 'SSparkSQLS');
> +---+--+
> | trim(SSparkSQLS, SL) |
> +---+--+
> | parkSQ |
> +---+--+
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Comment Edited] (SPARK-28794) Document CREATE TABLE in SQL Reference.

2019-11-06 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16966433#comment-16966433
 ] 

pavithra ramachandran edited comment on SPARK-28794 at 11/7/19 3:08 AM:


i shall raise PR by weekend.


was (Author: pavithraramachandran):
i shall raise PR by tomorrow.

> Document CREATE TABLE in SQL Reference.
> ---
>
> Key: SPARK-28794
> URL: https://issues.apache.org/jira/browse/SPARK-28794
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28794) Document CREATE TABLE in SQL Reference.

2019-11-03 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16966433#comment-16966433
 ] 

pavithra ramachandran commented on SPARK-28794:
---

i shall raise PR by tomorrow.

> Document CREATE TABLE in SQL Reference.
> ---
>
> Key: SPARK-28794
> URL: https://issues.apache.org/jira/browse/SPARK-28794
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29601) JDBC ODBC Tab Statement column should be provided ellipsis for bigger SQL statement

2019-10-25 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16959479#comment-16959479
 ] 

pavithra ramachandran commented on SPARK-29601:
---

i shall work on this

> JDBC ODBC Tab Statement column should be provided ellipsis for bigger SQL 
> statement
> ---
>
> Key: SPARK-29601
> URL: https://issues.apache.org/jira/browse/SPARK-29601
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> Statement column in JDBC/ODBC gives whole SQL statement and page size 
> increases.
> Suppose user submit and TPCDS Queries, then Page it display whole Query under 
> statement and User Experience is not good.
> Expected:
> It should display the ... Ellipsis and on clicking the stmt. it should Expand 
> display the whole SQL Statement.
> e.g.
> SELECT * FROM (SELECT count(*) h8_30_to_9 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 8 AND time_dim.t_minute >= 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s1, (SELECT count(*) h9_to_9_30 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 9 AND time_dim.t_minute < 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s2, (SELECT count(*) h9_30_to_10 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 9 AND time_dim.t_minute >= 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s3,(SELECT count(*) h10_to_10_30 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 10 AND time_dim.t_minute < 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s4,(SELECT count(*) h10_30_to_11 FROM 
> store_sales,household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 10 AND time_dim.t_minute >= 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s5,(SELECT count(*) h11_to_11_30 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> time_dim.t_time_sk AND ss_hdemo_sk = household_demographics.hd_demo_sk AND 
> ss_store_sk = s_store_sk AND time_dim.t_hour = 11 AND time_dim.t_minute < 30 
> AND ((household_demographics.hd_dep_count = 4 AND 
> household_demographics.hd_vehicle_count <= 4 + 2) OR 
> (household_demographics.hd_dep_count = 2 AND 
> household_demographics.hd_vehicle_count <= 2 + 2) OR 
> (household_demographics.hd_dep_count = 0 AND 
> household_demographics.hd_vehicle_count <= 0 + 2)) AND store.s_store_name = 
> 'ese') s6,(SELECT count(*) h11_30_to_12 FROM store_sales, 
> household_demographics, time_dim, store WHERE ss_sold_time_sk = 
> 

[jira] [Commented] (SPARK-29584) NOT NULL is not supported in Spark

2019-10-23 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29584?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16958540#comment-16958540
 ] 

pavithra ramachandran commented on SPARK-29584:
---

i shall work on this

> NOT NULL is not supported in Spark
> --
>
> Key: SPARK-29584
> URL: https://issues.apache.org/jira/browse/SPARK-29584
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> Spark while creating table restricting column for NULL value is not supported.
> As below
> PostgreSQL: SUCCESS No Exception
>  CREATE TABLE Persons (ID int *NOT NULL*, LastName varchar(255) *NOT 
> NULL*,FirstName varchar(255) NOT NULL, Age int);
>  insert into Persons values(1,'GUPTA','Abhi',NULL);
>  select * from persons;
>  
> Spark: Parse Exception
> jdbc:hive2://10.18.19.208:23040/default> CREATE TABLE Persons (ID int NOT 
> NULL, LastName varchar(255) NOT NULL,FirstName varchar(255) NOT NULL, Age 
> int);
> Error: org.apache.spark.sql.catalyst.parser.ParseException:
> no viable alternative at input 'CREATE TABLE Persons (ID int NOT'(line 1, pos 
> 29)
>  Parse Exception



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29504) Tooltip not display for Job Description even it shows ellipsed

2019-10-21 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16955807#comment-16955807
 ] 

pavithra ramachandran commented on SPARK-29504:
---

[~sandeep.katta2007]  i tested on the latest on double click its not expanding. 
I tested in 2.3.4 on double click it is expanding. I ll check why in master it 
is broken and fix asap

> Tooltip  not display for Job Description even it shows ellipsed
> ---
>
> Key: SPARK-29504
> URL: https://issues.apache.org/jira/browse/SPARK-29504
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
> Attachments: ToolTip JIRA.png
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29504) Tooltip not display for Job Description even it shows ellipsed

2019-10-17 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16954270#comment-16954270
 ] 

pavithra ramachandran commented on SPARK-29504:
---

i shall work on this

> Tooltip  not display for Job Description even it shows ellipsed
> ---
>
> Key: SPARK-29504
> URL: https://issues.apache.org/jira/browse/SPARK-29504
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29476) Add tooltip information for Thread Dump links and Thread details table columns in Executors Tab

2019-10-15 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16951839#comment-16951839
 ] 

pavithra ramachandran commented on SPARK-29476:
---

i shall work on this

> Add tooltip information for Thread Dump links and Thread details table 
> columns in Executors Tab
> ---
>
> Key: SPARK-29476
> URL: https://issues.apache.org/jira/browse/SPARK-29476
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: jobit mathew
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29460) Improve tooltip for Job Tab

2019-10-14 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950749#comment-16950749
 ] 

pavithra ramachandran commented on SPARK-29460:
---

i shall work on this

> Improve tooltip for Job Tab
> ---
>
> Key: SPARK-29460
> URL: https://issues.apache.org/jira/browse/SPARK-29460
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>
> [~LI,Xiao] I see there is inconsistency in the tool tip added for the column 
> across the tab. like Duration column in Job Tab does not have tooltip but 
> Duration Column in JDBC/ODBC Server tab has tooltip. 
> I submitted this Jira to handle this inconsistency in Job Tab Table Column.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29457) Improve tooltip information for Environment Tab

2019-10-13 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950729#comment-16950729
 ] 

pavithra ramachandran commented on SPARK-29457:
---

i ll work on this

> Improve tooltip information for Environment Tab
> ---
>
> Key: SPARK-29457
> URL: https://issues.apache.org/jira/browse/SPARK-29457
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29456) Add tooltip information for Session Statistics Table column in JDBC/ODBC Server Tab

2019-10-13 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29456?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16950720#comment-16950720
 ] 

pavithra ramachandran commented on SPARK-29456:
---

i will work on this

> Add tooltip information for Session Statistics Table column  in JDBC/ODBC 
> Server Tab
> 
>
> Key: SPARK-29456
> URL: https://issues.apache.org/jira/browse/SPARK-29456
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 3.0.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28794) Document CREATE TABLE in SQL Reference.

2019-09-20 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16934815#comment-16934815
 ] 

pavithra ramachandran commented on SPARK-28794:
---

Yes... I raise a PR soon.. 

> Document CREATE TABLE in SQL Reference.
> ---
>
> Key: SPARK-28794
> URL: https://issues.apache.org/jira/browse/SPARK-28794
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-29108) Add window.sql - Part 2

2019-09-16 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-29108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16931071#comment-16931071
 ] 

pavithra ramachandran commented on SPARK-29108:
---

i shall work on this

> Add window.sql - Part 2
> ---
>
> Key: SPARK-29108
> URL: https://issues.apache.org/jira/browse/SPARK-29108
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL, Tests
>Affects Versions: 3.0.0
>Reporter: Dylan Guedes
>Priority: Major
> Fix For: 3.0.0
>
>
> In this ticket, we plan to add the regression test cases of 
> [https://github.com/postgres/postgres/blob/REL_12_BETA3/src/test/regress/sql/window.sql#L320-L562|https://github.com/postgres/postgres/blob/REL_12_BETA3/src/test/regress/sql/window.sql#L320-L562]



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28972) [Spark] spark.memory.offHeap.size description require to update in document

2019-09-04 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28972?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16922414#comment-16922414
 ] 

pavithra ramachandran commented on SPARK-28972:
---

i will work on this

> [Spark] spark.memory.offHeap.size description require to update in document
> ---
>
> Key: SPARK-28972
> URL: https://issues.apache.org/jira/browse/SPARK-28972
> Project: Spark
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 2.4.3
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
>  
> spark.memory.offHeap.size accept 1G or 1KB also. So User is able to give 
> suffix also but description say only *'absolute amount of memory in bytes'.*
> This require to update like *spark.driver.memory* where it is mentioned is 
> accepts *a size unit suffix ("k", "m", "g" or "t") (e.g. {{512m}}, {{2g}}).* 
>  
> |{{spark.memory.offHeap.size}}|0|The *absolute amount of memory in bytes* 
> which can be used for off-heap allocation. This setting has no impact on heap 
> memory usage, so if your executors' total memory consumption must fit within 
> some hard limit then be sure to shrink your JVM heap size accordingly. This 
> must be set to a positive value when {{spark.memory.offHeap.enabled=true}}.|



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Issue Comment Deleted] (SPARK-28929) Spark Logging level should be INFO instead of Debug in Executor Plugin API[SPARK-24918]

2019-08-30 Thread pavithra ramachandran (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-28929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

pavithra ramachandran updated SPARK-28929:
--
Comment: was deleted

(was: I am working on this

 )

> Spark Logging level should be INFO instead of Debug in Executor Plugin 
> API[SPARK-24918]
> ---
>
> Key: SPARK-28929
> URL: https://issues.apache.org/jira/browse/SPARK-28929
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 2.4.2, 2.4.3
>Reporter: jobit mathew
>Priority: Minor
>
> Spark Logging level should be INFO instead of Debug in Executor Plugin 
> API[SPARK-24918].
> Currently logging level for Executor Plugin API[SPARK-24918] is DEBUG
> logDebug(s"Initializing the following plugins: $\{pluginNames.mkString(", 
> ")}")
> logDebug(s"Successfully loaded plugin " + 
> plugin.getClass().getCanonicalName())
> logDebug("Finished initializing plugins")
> It is better to change to  INFO instead of DEBUG.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28929) Spark Logging level should be INFO instead of Debug in Executor Plugin API[SPARK-24918]

2019-08-30 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16919523#comment-16919523
 ] 

pavithra ramachandran commented on SPARK-28929:
---

I am working on this

 

> Spark Logging level should be INFO instead of Debug in Executor Plugin 
> API[SPARK-24918]
> ---
>
> Key: SPARK-28929
> URL: https://issues.apache.org/jira/browse/SPARK-28929
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 2.4.2, 2.4.3
>Reporter: jobit mathew
>Priority: Minor
>
> Spark Logging level should be INFO instead of Debug in Executor Plugin 
> API[SPARK-24918].
> Currently logging level for Executor Plugin API[SPARK-24918] is DEBUG
> logDebug(s"Initializing the following plugins: $\{pluginNames.mkString(", 
> ")}")
> logDebug(s"Successfully loaded plugin " + 
> plugin.getClass().getCanonicalName())
> logDebug("Finished initializing plugins")
> It is better to change to  INFO instead of DEBUG.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28897) Invalid usage of '*' in expression 'coalesce' error when executing dataframe.na.fill(0)

2019-08-28 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28897?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16917733#comment-16917733
 ] 

pavithra ramachandran commented on SPARK-28897:
---

i will work on this

> Invalid usage of '*' in expression 'coalesce' error when executing 
> dataframe.na.fill(0)
> ---
>
> Key: SPARK-28897
> URL: https://issues.apache.org/jira/browse/SPARK-28897
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.4.0
>Reporter: Saurabh Santhosh
>Priority: Major
>
> Getting the following error when trying to execute the given statements
>  
> {code:java}
> var df = spark.sql(s"select * from default.test_table")
> df.na.fill(0)
> {code}
> This error happens when the following property is set
> {code:java}
> spark.sql("set spark.sql.parser.quotedRegexColumnNames=true")
> {code}
> Error :
> {code:java}
> org.apache.spark.sql.AnalysisException: Invalid usage of '*' in expression 
> 'coalesce';   at 
> org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:42)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:95)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$expandStarExpression$1.applyOrElse(Analyzer.scala:1021)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$expandStarExpression$1.applyOrElse(Analyzer.scala:997)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:278)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:278)
>    at 
> org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:277)  
>  at 
> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$3.apply(TreeNode.scala:275)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$3.apply(TreeNode.scala:275)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
>    at 
> org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)  
>  at 
> org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:275)  
>  at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$.expandStarExpression(Analyzer.scala:997)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveReferences$$buildExpandedProjectList$1.apply(Analyzer.scala:982)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveReferences$$buildExpandedProjectList$1.apply(Analyzer.scala:977)
>    at 
> scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
>    at 
> scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)
>    at 
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>    at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)   at 
> scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)   
> at scala.collection.AbstractTraversable.flatMap(Traversable.scala:104)   at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveReferences$$buildExpandedProjectList(Analyzer.scala:977)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$apply$9.applyOrElse(Analyzer.scala:905)
>    at 
> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveReferences$$anonfun$apply$9.applyOrElse(Analyzer.scala:900)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
>    at 
> org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
>    at 
> org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)

[jira] [Commented] (SPARK-28799) Document TRUNCATE TABLE in SQL Reference.

2019-08-21 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16912054#comment-16912054
 ] 

pavithra ramachandran commented on SPARK-28799:
---

i will work on this

> Document TRUNCATE TABLE in SQL Reference.
> -
>
> Key: SPARK-28799
> URL: https://issues.apache.org/jira/browse/SPARK-28799
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28794) Document CREATE TABLE in SQL Reference.

2019-08-21 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16912050#comment-16912050
 ] 

pavithra ramachandran commented on SPARK-28794:
---

i will work on this

> Document CREATE TABLE in SQL Reference.
> ---
>
> Key: SPARK-28794
> URL: https://issues.apache.org/jira/browse/SPARK-28794
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation, SQL
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28791) Document ALTER TABLE statement in SQL Reference.

2019-08-21 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28791?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16912023#comment-16912023
 ] 

pavithra ramachandran commented on SPARK-28791:
---

i will work on this

> Document ALTER TABLE statement in SQL Reference.
> 
>
> Key: SPARK-28791
> URL: https://issues.apache.org/jira/browse/SPARK-28791
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation
>Affects Versions: 2.4.3
>Reporter: Dilip Biswal
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28672) [UDF] Duplicate function creation should not allow

2019-08-19 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16911003#comment-16911003
 ] 

pavithra ramachandran commented on SPARK-28672:
---

[~maropu] [~viirya] 

The intention of this Jira is to not allow creation of temporary function, when 
a permanent function with the same name exist. 

Could you confirm if its ok to handle this case?

> [UDF] Duplicate function creation should not allow 
> ---
>
> Key: SPARK-28672
> URL: https://issues.apache.org/jira/browse/SPARK-28672
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.4.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> create function addm_3  AS 
> 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.084 seconds)
> {code}
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> create temporary function addm_3  
> AS 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> INFO  : converting to local hdfs://hacluster/user/Multiply.jar
> INFO  : Added 
> [/tmp/8a396308-41f8-4335-9de4-8268ce5c70fe_resources/Multiply.jar] to class 
> path
> INFO  : Added resources: [hdfs://hacluster/user/Multiply.jar]
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.134 seconds)
> {code}
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> show functions like addm_3;
> +-+--+
> |function |
> +-+--+
> | addm_3  |
> | default.addm_3  |
> +-+--+
> 2 rows selected (0.047 seconds)
> {code}
> When show function executed it is listing both the function but what about 
> the db for permanent function when user has not specified.
> Duplicate should not be allowed if user creating temporary one with the same 
> name.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28672) [UDF] Duplicate function creation should not allow

2019-08-19 Thread pavithra ramachandran (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16911002#comment-16911002
 ] 

pavithra ramachandran commented on SPARK-28672:
---

[~abhishek.akg] -  When we execute show function- it is displaying the 
temporary and permanent function that is created, since no database is 
mentioned while creating the permanent it is stored as default.addm3, Temporary 
functions are not specific to any database, So it is displayed without any db 
name. I dont think that is an issue

> [UDF] Duplicate function creation should not allow 
> ---
>
> Key: SPARK-28672
> URL: https://issues.apache.org/jira/browse/SPARK-28672
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.4.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> create function addm_3  AS 
> 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.084 seconds)
> {code}
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> create temporary function addm_3  
> AS 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> INFO  : converting to local hdfs://hacluster/user/Multiply.jar
> INFO  : Added 
> [/tmp/8a396308-41f8-4335-9de4-8268ce5c70fe_resources/Multiply.jar] to class 
> path
> INFO  : Added resources: [hdfs://hacluster/user/Multiply.jar]
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.134 seconds)
> {code}
> {code}
> 0: jdbc:hive2://10.18.18.214:23040/default> show functions like addm_3;
> +-+--+
> |function |
> +-+--+
> | addm_3  |
> | default.addm_3  |
> +-+--+
> 2 rows selected (0.047 seconds)
> {code}
> When show function executed it is listing both the function but what about 
> the db for permanent function when user has not specified.
> Duplicate should not be allowed if user creating temporary one with the same 
> name.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28653) Create table using DDL statement should not auto create the destination folder

2019-08-09 Thread pavithra ramachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-28653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903834#comment-16903834
 ] 

pavithra ramachandran commented on SPARK-28653:
---

i will check this

> Create table using DDL statement should not auto create the destination folder
> --
>
> Key: SPARK-28653
> URL: https://issues.apache.org/jira/browse/SPARK-28653
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.4.3
>Reporter: Thanida
>Priority: Minor
>
> I create external table using this following DDL statement, the destination 
> path was auto-created.
> {code:java}
> CREATE TABLE ${tableName} USING parquet LOCATION ${path}
> {code}
> But, if I specified file format as csv or json, the destination path was not 
> created.
> {code:java}
> CREATE TABLE ${tableName} USING CSV LOCATION ${path}
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28672) [UDF] Duplicate function creation should not allow

2019-08-09 Thread pavithra ramachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-28672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903584#comment-16903584
 ] 

pavithra ramachandran commented on SPARK-28672:
---

i will work on this

> [UDF] Duplicate function creation should not allow 
> ---
>
> Key: SPARK-28672
> URL: https://issues.apache.org/jira/browse/SPARK-28672
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.4.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
> 0: jdbc:hive2://10.18.18.214:23040/default> create function addm_3  AS 
> 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.084 seconds)
> 0: jdbc:hive2://10.18.18.214:23040/default> create temporary function addm_3  
> AS 'com.huawei.bigdata.hive.example.udf.multiply' using jar 
> 'hdfs://hacluster/user/Multiply.jar';
> INFO  : converting to local hdfs://hacluster/user/Multiply.jar
> INFO  : Added 
> [/tmp/8a396308-41f8-4335-9de4-8268ce5c70fe_resources/Multiply.jar] to class 
> path
> INFO  : Added resources: [hdfs://hacluster/user/Multiply.jar]
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> No rows selected (0.134 seconds)
> 0: jdbc:hive2://10.18.18.214:23040/default> show functions like addm_3;
> +-+--+
> |function |
> +-+--+
> | addm_3  |
> | default.addm_3  |
> +-+--+
> 2 rows selected (0.047 seconds)
> 0: jdbc:hive2://10.18.18.214:23040/default>
> When show function executed it is listing both the function but what about 
> the db for permanent function when user has not specified.
> Duplicate should not be allowed if user creating temporary one with the same 
> name.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28671) [UDF] dropping permanent function when a temporary function with the same name already exists giving wrong msg on dropping it again

2019-08-08 Thread pavithra ramachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-28671?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903570#comment-16903570
 ] 

pavithra ramachandran commented on SPARK-28671:
---

i will work on this

> [UDF] dropping permanent function when a temporary function with the same 
> name already exists giving wrong msg on dropping it again
> ---
>
> Key: SPARK-28671
> URL: https://issues.apache.org/jira/browse/SPARK-28671
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.4.0
>Reporter: ABHISHEK KUMAR GUPTA
>Priority: Minor
>
> Created jar and uploaded at hdfs path
> 1../hdfs dfs -put /opt/trash1/AddDoublesUDF.jar /user/user1/
> 2.Launch beeline and created permanent function
> CREATE FUNCTION addDoubles AS 
> 'com.huawei.bigdata.hive.example.udf.AddDoublesUDF' using jar 
> 'hdfs://hacluster/user/user1/AddDoublesUDF.jar';
> 3.Perform select operation
> jdbc:hive2://100.100.208.125:23040/default> SELECT addDoubles(1,2,3);
> +--+--+
> | default.addDoubles(1, 2, 3)  |
> +--+--+
> | 6.0  |
> +--+--+
> 1 row selected (0.111 seconds)
> 4.Created temporary function as below
> jdbc:hive2://100.100.208.125:23040/default> CREATE temporary FUNCTION 
> addDoubles AS 'com.huawei.bigdata.hive.example.udf.AddDoublesUDF' using jar 
> 'hdfs://hacluster/user/user1/AddDoublesUDF.jar';
> 5.jdbc:hive2://100.100.208.125:23040/default> SELECT addDoubles(1,2,3);
> +--+--+
> | addDoubles(1, 2, 3)  |
> +--+--+
> | 6.0  |
> +--+--+
> 1 row selected (0.088 seconds)
> 6.Drop function
> jdbc:hive2://100.100.208.125:23040/default> drop function addDoubles;
> +-+--+
> | Result  |
> +-+--+
> +-+--+
> 7.jdbc:hive2://100.100.208.125:23040/default> SELECT addDoubles(1,2,3); 
> -- It is success
> 8.Drop again Error thrown
> jdbc:hive2://100.100.208.125:23040/default> drop function addDoubles;
> Error: org.apache.spark.sql.catalyst.analysis.NoSuchFunctionException: 
> Undefined function: 'default.addDoubles'. This function is neither a 
> registered temporary function nor a permanent function registered in the 
> database 'default'.; (state=,code=0)
> 9.Perform again select 
> jdbc:hive2://100.100.208.125:23040/default>  SELECT addDoubles(1,2,3);
> +--+--+
> | addDoubles(1, 2, 3)  |
> +--+--+
> | 6.0  |
>   
> Issue is why the Error msg shown is step 8 saying it is neither registered as 
> permanent or temporary function where as it is registered as temporary 
> function in step 4 that is why in step 9 select is returning result.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28338) spark.read.format("csv") treat empty string as null if csv file don't have quotes in data

2019-07-11 Thread pavithra ramachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/SPARK-28338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16882872#comment-16882872
 ] 

pavithra ramachandran commented on SPARK-28338:
---

Could you given more information, as to what is your exact expectation by 
giving an  example ?

> spark.read.format("csv") treat empty string as null if csv file don't have 
> quotes in data
> -
>
> Key: SPARK-28338
> URL: https://issues.apache.org/jira/browse/SPARK-28338
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 2.4.3
>Reporter: Jayadevan M
>Priority: Major
>
> The csv input file
> +cat sample.csv+ 
>  Name,Lastname,Age
>  abc,,32
>  pqr,xxx,30
>  
> +spark-shell+
> spark.read.format("csv").option("header", 
> "true").load("/media/ub_share/projects/*.csv").head(3)
>  res14: Array[org.apache.spark.sql.Row] = Array([abc,null,32], [pqr,xxx,30])
>  
> scala> spark.read.format("csv").option("header", "true").option("nullValue", 
> "?").load("/media/ub_share/projects/*.csv").head(3)
>  res15: Array[org.apache.spark.sql.Row] = Array([abc,null,32], [pqr,xxx,30])
>  
> The empty string get converted to null. Its works fine if the csv file have 
> quotes in columns.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org