Github user navis commented on the pull request:
https://github.com/apache/spark/pull/9107#issuecomment-171484501
It's only happened when using older version of hive jdbc. I'm closing.
---
If your project is set up for it, you can reply to this email and have your
reply appear
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/9107
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/10585#issuecomment-170411652
@srowen moved it to `StatsReportListener`. sorry for delay.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/10585#discussion_r48915612
--- Diff: core/src/main/scala/org/apache/spark/scheduler/StageInfo.scala ---
@@ -61,6 +61,21 @@ class StageInfo(
"ru
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/10585#discussion_r48916174
--- Diff: core/src/main/scala/org/apache/spark/scheduler/StageInfo.scala ---
@@ -61,6 +61,21 @@ class StageInfo(
"ru
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/10585
SPARK-12637 Print stage info of finished stages properly
Currently it prints hashcode of stage info, which seemed not that useful.
```
INFO scheduler.StatsReportListener: Finished stage
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/10572
SPARK-12619 Combine small files in a hadoop directory into single split
When a directory contains too many (small) files, whole spark cluster will
be exhausted scheduling tasks created for each file
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6767#issuecomment-158858838
@chenghao-intel Sorry for long delay. Could see this again when your time
allowed?
---
If your project is set up for it, you can reply to this email and have your
reply
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6767#discussion_r45292473
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -523,6 +531,30 @@ class HiveContext(sc: SparkContext) extends
SQLContext
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/9580#issuecomment-157911907
done
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6767#discussion_r45292499
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -523,6 +531,30 @@ class HiveContext(sc: SparkContext) extends
SQLContext
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6767#discussion_r45292520
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala ---
@@ -875,6 +899,65 @@ private[hive] case class MetastoreRelation
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/6858
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/9074
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/9514
[SPARK-11546] Thrift server makes too many logs about result schema
SparkExecuteStatementOperation logs result schema for each getNextRowSet()
calls which is by default every 1000 rows, overwhelming
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8456#discussion_r43711653
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/generators.scala
---
@@ -53,7 +53,7 @@ trait Generator extends Expression
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8456#discussion_r43590635
--- Diff: sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala ---
@@ -1119,7 +1119,8 @@ class DataFrame private[sql](
def explode[A <: Prod
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/8355
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8456#discussion_r42953994
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala
---
@@ -148,7 +148,7 @@ class Analyzer(
child match
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8456#issuecomment-150997853
Rebased to trunk
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/9107#issuecomment-150998671
@chenghao-intel It will not happen with `columnar` format which just calls
`String.valueOf(field)` whatever the field is. But in row-based set which was
used older
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/6790
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8899#issuecomment-149788617
Then, the problem would be which one is better. #9142 looks more intuitive
and this provides a chance to handle the situation(invalid path) for input
formatter
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8354#discussion_r42444143
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -547,10 +548,15 @@ class HiveContext(sc: SparkContext) extends
SQLContext
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8354#discussion_r42443991
--- Diff: sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala
---
@@ -118,8 +118,8 @@ private[hive] object HiveShim {
*
* @param
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8354#discussion_r42444696
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -547,10 +548,15 @@ class HiveContext(sc: SparkContext) extends
SQLContext
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/9074#issuecomment-149426130
This was for fixing NPE from operation log and I didn't verified it worked.
I'll check.
---
If your project is set up for it, you can reply to this email and have your
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8354#discussion_r42450109
--- Diff: sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala
---
@@ -196,14 +194,8 @@ private[hive] object HiveShim {
if (instance
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8354#discussion_r42455359
--- Diff: sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveShim.scala
---
@@ -196,14 +194,8 @@ private[hive] object HiveShim {
if (instance
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/6488
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/9130#discussion_r42212370
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/jsonExpressions.scala
---
@@ -251,16 +254,18 @@ case class GetJsonObject(json
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/9130#discussion_r42200517
--- Diff: core/src/main/scala/org/apache/spark/util/Utils.scala ---
@@ -2153,6 +2153,9 @@ private[spark] object Utils extends Logging {
conf.getInt
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/9130
[SPARK-11124] JsonParser/Generator should be closed for resource recycle
Some json parsers are not closed. parser in JacksonParser#parseJson, for
example.
You can merge this pull request into a Git
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/9107#issuecomment-148281071
@davies It'll be reproduced only when using older version of hive jdbc,
which does not support columnar wire format (RowBasedSet is older format). Idea?
---
If your
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/9107
[SPARK-11067] Spark SQL thrift server fails to handle decimal value
When executing the following query through beeline connecting to Spark sql
thrift server, it errors out for decimal column
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/9074
[SPARK-11062] Thrift server does not support operationLog
Currently, SparkExecuteStatementOperation is skipping beforeRun/afterRun
method.
You can merge this pull request into a Git repository
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/7118
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8909#issuecomment-146104351
@davies I love the idea removing thread-local, which caused headache for me
doing #7118. One thing I like to see also fixed in here is that jdbc meta calls
like
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8899#discussion_r41105674
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JSONRelation.scala
---
@@ -79,12 +76,16 @@ private[sql] class JSONRelation
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8899#issuecomment-145407288
@srowen Sorry, I've been on vacation. Addressed comment.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/8893
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8893#discussion_r40290254
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/mathExpressions.scala
---
@@ -72,6 +72,18 @@ abstract class
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8893#discussion_r40388170
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/mathExpressions.scala
---
@@ -52,11 +52,12 @@ abstract class
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8893
[SPARK-10724] [SQL] SQL's floor() returns DOUBLE
This is a change in behavior from 1.4.1 where {{floor}} returns a BIGINT.
{code}
scala> sql("select floor(1)").printSc
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8899
[SPARK-10709] [SQL] When loading a json dataset as a data frame, if the
input path is wrong, the error message is very confusing
If you do something like {{sqlContext.read.json("a wrong
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8899#issuecomment-142811330
Now it throws,
```
org.apache.hadoop.mapred.InvalidInputException: Input path does not exist:
hdfs://localhost:9000/user/navis/a wrong path
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8804#issuecomment-141905344
@rxin Added comments
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8808#issuecomment-142143125
@rxin It's just a table with 100+ string columns partitioned by a string
key. It happened by a simple query just like select <100+> from where
.
---
If your p
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8354#issuecomment-141379396
@JoshRosen It isn't because spark does not support "create macro". I've
tried to support that in new patch.
---
If your project is set up for it, you
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/8804#discussion_r39837163
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/TableReader.scala ---
@@ -295,13 +298,25 @@ private[hive] object HadoopTableReader extends
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8808
[SPARK-10684] [SQL] StructType.interpretedOrdering need not to be serialized
Kryo fails with buffer overflow even with max value (2G).
{noformat}
org.apache.spark.SparkException: Kryo
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8804
[SPARK-10679] [CORE] javax.jdo.JDOFatalUserException in executor
HadoopRDD throws exception in executor, something like below.
{noformat}
5/09/17 18:51:21 INFO metastore.HiveMetaStore: 0
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/7118#issuecomment-139934174
@WangTaoTheTonic I've fixed test itself, which seemed invalid. "add jar"
should be applied to the session which called it but in test, it expects added
jar to b
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/7118#issuecomment-139177871
@WangTaoTheTonic I've also seen that rebasing this to spark-1.5.0. Seemed
done something wrong in rebase process. I'll update soon.
---
If your project is set up
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8456
[SPARK-9034] [SQL] Reflect field names defined in GenericUDTF
Hive GenericUDTF#initialize() defines field names in a returned schema
though,
the current HiveGenericUDTF drops these names.
We
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/8390
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/8390#issuecomment-134843288
@liancheng Sorry, char type cannot be properly compared to other text
values if once converted to string. I'll close PR.
---
If your project is set up for it, you can
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8390
[SPARK-9685] [SQL] Unspported dataType: char(X) in Hive
I'm getting the following error when I try to read a Hive table with
char(X) fields:
{code}
15/08/06 11:38:51 INFO
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8355
[SPARK-10152] [SQL] Support Init script for hive-thriftserver
If some queries can be executed on thrift server in initialization
stage(mostly for registering functions or macros), things are done
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/8354
[SPARK-10151] [SQL] Support invocation of hive macro
Macro in hive (which is GenericUDFMacro) contains real function inside of
it but it's not conveyed to tasks, resulting null-pointer exception
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/7225
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/7225#discussion_r33933969
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Projection.scala
---
@@ -22,9 +22,15 @@ package
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/7225
[SPARK-8826] [SQL] Fix ClassCastException in GeneratedAggregate
When codegen is disabled, ClassCastException is thrown in some cases.
You can merge this pull request into a Git repository by running
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6810#issuecomment-118584139
Thanks, @JoshRosen. Actually, it's two bugs which is
* memory leak on empty input
* CCE in some cases (codeGen=false (groupbyException.isEmpty || unsafe =
false
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/7127#issuecomment-118571972
Adressed comments. Thanks, @kmadhugit
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/7118
[SPARK-8552] [THRIFTSERVER] Using incorrect database in multiple sessions
Currently, metaHive in HiveContext shares single SessionState instance with
all execution threads, which makes problems
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/7127
[SPARK-8707] RDD#toDebugString fails if any cached RDD has invalid
partitions
Added numPartitions(evaluate: Boolean) to RDD. With evaluate=true the
method is same with partitions.length
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6931#issuecomment-116902528
@andrewor14 Currently, timestamp == string converts string into
timestamp, but for timestamp string, string is converted into timestamp.
And this inconsistency makes me
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/6931
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33533643
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,59 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33535187
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,59 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33535200
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/SparkPlan.scala ---
@@ -153,13 +153,14 @@ abstract class SparkPlan extends QueryPlan[SparkPlan
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33533948
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,59 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r5666
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Projection.scala
---
@@ -22,9 +22,11 @@ package
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r5692
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Projection.scala
---
@@ -22,9 +22,11 @@ package
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33016547
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/parquet/ParquetTableOperations.scala
---
@@ -49,7 +49,7 @@ import org.apache.spark.sql.SQLConf
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r33016397
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,42 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r32998276
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,43 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6810#discussion_r32998221
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/AggregateSuite.scala ---
@@ -0,0 +1,43 @@
+/*
+ * Licensed to the Apache Software
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6810#issuecomment-114017606
@JoshRosen Sure, I'll review your patch first. Thanks.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6931
[SPARK-8420] [SQL] Inconsistent behavior with Dataframe Timestamp between
1.3.1 and 1.4.0
Havn't seen that the issue was fixed already. But It seemed it still has a
inconsistency problem on equals
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6931#issuecomment-114016842
@marmbrus Could you review this, plz?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6810#issuecomment-114038518
@JoshRosen Done.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6858
[SPARK-8333] [SQL] Spark failed to delete temp directory created by
HiveContext
Shutdown underlying megastore and delete temporary directory when the hive
context is closed.
You can merge
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6840
[SPARK-8333] [SQL] Spark failed to delete temp directory created by
HiveContext
Added onContextStop(SparkListenerContextStop) method to SparkListener and
registered the location of temporary
Github user navis closed the pull request at:
https://github.com/apache/spark/pull/6840
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6840#issuecomment-112617257
Ignore this, for now.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6696#issuecomment-111796747
done in SPARK-8357
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6696#issuecomment-111774230
@JoshRosen ok, sure.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6696#issuecomment-111765959
Strange.. cannot reproduce the fail in local env. I'll check it again.
---
If your project is set up for it, you can reply to this email and have your
reply appear
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6810
[SPARK-8357] [SQL] Memory leakage on unsafe aggregation path with empty
input
Currently, unsafe-based hash is released on 'next' call but if input is
empty, it would not be called ever.
You can
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6696#issuecomment-111772515
The memory leakage caused the test fail was a existing bug in master
branch. Currently, unsafe-based hash is released on 'next' call but if input is
empty, it would
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6696#discussion_r32368950
--- Diff: core/src/main/scala/org/apache/spark/util/collection/BitSet.scala
---
@@ -32,6 +34,10 @@ class BitSet(numBits: Int) extends Serializable
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6696#discussion_r32369543
--- Diff: core/src/main/scala/org/apache/spark/util/collection/BitSet.scala
---
@@ -32,6 +34,10 @@ class BitSet(numBits: Int) extends Serializable
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6790
[SPARK-8334] [SQL] Binary logical plan should provide more realistic
statistics
Currently, spark-sql multiplies size of two children for output size, which
makes following join hash the other side
Github user navis commented on the pull request:
https://github.com/apache/spark/pull/6696#issuecomment-111671743
Test fail was just caused by appearance order. Added order-by for
deterministic result
---
If your project is set up for it, you can reply to this email and have your
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6767
[SPARK-8312] [SQL] Populate statistics info of hive tables if it's needed
to be
Currently, spark-sql uses stats in metastore for estimating size of hive
table, which means analyze command should
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6736
[SPARK-8285] [SQL] CombineSum should be calculated as unlimited decimal
first
case cs @ CombineSum(expr) =
val calcType = expr.dataType
expr.dataType match
Github user navis commented on a diff in the pull request:
https://github.com/apache/spark/pull/6699#discussion_r31972106
--- Diff:
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala
---
@@ -81,7 +84,7 @@ trait Catalog {
}
class
GitHub user navis opened a pull request:
https://github.com/apache/spark/pull/6699
[SPARK-7792] [SQL] HiveContext registerTempTable not thread safe
Just replaced mutable.HashMap to ConcurrentHashMap
You can merge this pull request into a Git repository by running:
$ git pull
1 - 100 of 103 matches
Mail list logo