Re: 【Hive Alter Table Add column at specified position】

2020-07-20 Thread Rui Li
Yeah, according to our DDL doc, we don't support this use case at the
moment. Perhaps you can use REPLACE COLUMNS as a workaround.

On Sat, Jun 27, 2020 at 5:32 PM 忝忝向仧 <153488...@qq.com> wrote:

> Hi,all:
>
>
> It seems that Hive can not alter table to add column atspecified
> position.
> For instance,the Table A has c1,c2,c3 columns,and i want to add column c4
> after c1,therefore,the table would be like c1,c4,c2,c3 instead of
> c1,c2,c3,c4.
>
>
> Thanks.



-- 
Best regards!
Rui Li


Hit HIVE-13023 with 2.0.1 maven artifacts

2019-09-04 Thread Rui Li
Hello guys,

I hit HIVE-13023 <https://issues.apache.org/jira/browse/HIVE-13023> when I
programmatically executed some queries with Hive-2.0.1. I did some
investigation and there seemed to be some issues with the 2.0.1 artifacts
we published.
I compared the hive-exec artifact from maven central
<https://repo1.maven.org/maven2/org/apache/hive/hive-exec/2.0.1/hive-exec-2.0.1.jar>
with
the jar in our binary distribution
<https://archive.apache.org/dist/hive/hive-2.0.1/apache-hive-2.0.1-bin.tar.gz>,
and the two hive-exec jars are of different sizes.
I also decompiled these two jars to check the offending method
*StorageFormat::fillStorageFormat*. It turned out the jar from maven got
the ordinals of some tokens wrong (the ordinal of TOK_FILEFORMAT_GENERIC
should be 715):
[image: 屏幕快照 2019-09-04 下午5.08.59.png]

And the jar in our distribution has the correct ordinals:
[image: 屏幕快照 2019-09-04 下午5.14.00.png]

I wonder whether anybody could help verify the issue, and whether it's
possible to update the published jars if the issue is valid?

-- 
Best regards!
Rui Li


[jira] [Created] (HIVE-22053) Function name is not normalized when creating function

2019-07-26 Thread Rui Li (JIRA)
Rui Li created HIVE-22053:
-

 Summary: Function name is not normalized when creating function
 Key: HIVE-22053
 URL: https://issues.apache.org/jira/browse/HIVE-22053
 Project: Hive
  Issue Type: Bug
  Components: Standalone Metastore
Reporter: Rui Li
Assignee: Rui Li


If a function is created with a name containing upper case characters, we get 
NoSuchObjectException when trying to get that function.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


Re: Is Github PR mandatory?

2019-03-17 Thread Rui Li
Got it. Thanks!

On Fri, Mar 15, 2019 at 1:47 PM Mani M  wrote:

> It's used as substitution for review board.
>
>
> With Regards
> M.Mani
> +61 432 461 087
>
> On Fri, 15 Mar 2019, 13:59 Rui Li,  wrote:
>
> > Hi,
> >
> > I believe we still need to upload patch to JIRA for precommit testing. So
> > just want to make sure whether opening a github PR is mandatory? Or is it
> > just a substitution for the review board?
> >
> > --
> > Best regards!
> > Rui Li
> >
>


-- 
Best regards!
Rui Li


Is Github PR mandatory?

2019-03-14 Thread Rui Li
Hi,

I believe we still need to upload patch to JIRA for precommit testing. So
just want to make sure whether opening a github PR is mandatory? Or is it
just a substitution for the review board?

-- 
Best regards!
Rui Li


[jira] [Created] (HIVE-19895) The unique ID in SparkPartitionPruningSinkOperator is no longer needed

2018-06-14 Thread Rui Li (JIRA)
Rui Li created HIVE-19895:
-

 Summary: The unique ID in SparkPartitionPruningSinkOperator is no 
longer needed
 Key: HIVE-19895
 URL: https://issues.apache.org/jira/browse/HIVE-19895
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19671) Distribute by rand() can lead to data inconsistency

2018-05-23 Thread Rui Li (JIRA)
Rui Li created HIVE-19671:
-

 Summary: Distribute by rand() can lead to data inconsistency
 Key: HIVE-19671
 URL: https://issues.apache.org/jira/browse/HIVE-19671
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li


Noticed the following queries can give different results:
{code}
select count(*) from tbl;
select count(*) from (select * from tbl distribute by rand());
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [VOTE] Stricter commit guidelines

2018-05-15 Thread Rui Li
+1

On Tue, May 15, 2018 at 2:24 PM, Prasanth Jayachandran <
pjayachand...@hortonworks.com> wrote:

> +1
>
>
>
> Thanks
> Prasanth
>
>
>
> On Mon, May 14, 2018 at 10:44 PM -0700, "Jesus Camacho Rodriguez" <
> jcama...@apache.org<mailto:jcama...@apache.org>> wrote:
>
>
> After work has been done to ignore most of the tests that were failing
> consistently/intermittently [1], I wanted to start this vote to gather
> support from the community to be stricter wrt committing patches to Hive.
> The committers guide [2] already specifies that a +1 should be obtained
> before committing, but there is another clause that allows committing under
> the presence of flaky tests (clause 4). Flaky tests are as good as having
> no tests, hence I propose to remove clause 4 and enforce the +1 from
> testing infra before committing.
>
>
>
> As I see it, by enforcing that we always get a +1 from the testing infra
> before committing, 1) we will have a more stable project, and 2) we will
> have another incentive as a community to create a more robust testing
> infra, e.g., replacing flaky tests for similar unit tests that are not
> flaky, trying to decrease running time for tests, etc.
>
>
>
> Please, share your thoughts about this.
>
>
>
> Here is my +1.
>
>
>
> Thanks,
>
> Jes?s
>
>
>
> [1] http://mail-archives.apache.org/mod_mbox/hive-dev/201805.
> mbox/%3C63023673-AEE5-41A9-BA52-5A5DFB2078B6%40apache.org%3E
>
> [2] https://cwiki.apache.org/confluence/display/Hive/
> HowToCommit#HowToCommit-PreCommitruns,andcommittingpatches
>
>
>
>
>


-- 
Best regards!
Rui Li


Re: [DISCUSS] Unsustainable situation with ptests

2018-05-14 Thread Rui Li
cally. May be create a profile which runs a subset of
> > important tests which are consistent. We can apply some label that
> > pre-checkin-local tests are runs successful and only then we submit
> > for the
> > full-suite.
> >
> > More thoughts are welcome. Thanks for starting this conversation.
> >
> > On Fri, May 11, 2018 at 1:27 PM, Jesus Camacho Rodriguez <
> > jcama...@apache.org<mailto:jcama...@apache.org>> wrote:
> >
> > I believe we have reached a state (maybe we did reach it a while ago)
> > that
> > is not sustainable anymore, as there are so many tests failing /
> > timing out
> > that it is not possible to verify whether a patch is breaking some
> > critical
> > parts of the system or not. It also seems to me that due to the
> > timeouts
> > (maybe due to infra, maybe not), ptest runs are taking even longer
> than
> > usual, which in turn creates even longer queue of patches.
> >
> > There is an ongoing effort to improve ptests usability (
> > https://issues.apache.org/jira/browse/HIVE-19425), but apart from
> > that,
> > we need to make an effort to stabilize existing tests and bring that
> > failure count to zero.
> >
> > Hence, I am suggesting *we stop committing any patch before we get a
> > green
> > run*. If someone thinks this proposal is too radical, please come up
> > with
> > an alternative, because I do not think it is OK to have the ptest
> runs
> > in
> > their current state. Other projects of certain size (e.g., Hadoop,
> > Spark)
> > are always green, we should be able to do the same.
> >
> > Finally, once we get to zero failures, I suggest we are less tolerant
> > with
> > committing without getting a clean ptests run. If there is a failure,
> > we
> > need to fix it or revert the patch that caused it, then we continue
> > developing.
> >
> > Please, let’s all work together as a community to fix this issue,
> that
> > is
> > the only way to get to zero quickly.
> >
> > Thanks,
> > Jesús
> >
> > PS. I assume the flaky tests will come into the discussion. Let´s see
> > first how many of those we have, then we can work to find a fix.
> >
> >
> >
> >
> >
> >
> >
> >
>



-- 
Best regards!
Rui Li


[jira] [Created] (HIVE-19439) MapWork shouldn't be reused when Spark task fails during initialization

2018-05-07 Thread Rui Li (JIRA)
Rui Li created HIVE-19439:
-

 Summary: MapWork shouldn't be reused when Spark task fails during 
initialization
 Key: HIVE-19439
 URL: https://issues.apache.org/jira/browse/HIVE-19439
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li


Issue identified in HIVE-19388. When a Spark task fails during initializing the 
map operator, the task is retried with the same MapWork retrieved from cache. 
This can be problematic because the MapWork may be partially initialized, e.g. 
some operators are already in INIT state.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-19316) StatsTask fails due to ClassCastException

2018-04-26 Thread Rui Li (JIRA)
Rui Li created HIVE-19316:
-

 Summary: StatsTask fails due to ClassCastException
 Key: HIVE-19316
 URL: https://issues.apache.org/jira/browse/HIVE-19316
 Project: Hive
  Issue Type: Bug
  Components: Statistics
Reporter: Rui Li


The stack trace:
{noformat}
2018-04-26T20:17:37,674 ERROR [pool-7-thread-11] metastore.RetryingHMSHandler: 
java.lang.ClassCastException: 
org.apache.hadoop.hive.metastore.api.LongColumnStatsData cannot be cast to 
org.apache.hadoop.hive.metastore.columnstats.cache.LongColumnStatsDataInspector
at 
org.apache.hadoop.hive.metastore.columnstats.merge.LongColumnStatsMerger.merge(LongColumnStatsMerger.java:30)
at 
org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.mergeColStats(MetaStoreUtils.java:1052)
at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.set_aggr_stats_for(HiveMetaStore.java:7202)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:147)
at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:108)
at com.sun.proxy.$Proxy26.set_aggr_stats_for(Unknown Source)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$set_aggr_stats_for.getResult(ThriftHiveMetastore.java:16795)
at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$set_aggr_stats_for.getResult(ThriftHiveMetastore.java:16779)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at 
org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:111)
at 
org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:107)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
at 
org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:119)
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-18955) HoS: Unable to create Channel from class NioServerSocketChannel

2018-03-14 Thread Rui Li (JIRA)
Rui Li created HIVE-18955:
-

 Summary: HoS: Unable to create Channel from class 
NioServerSocketChannel
 Key: HIVE-18955
 URL: https://issues.apache.org/jira/browse/HIVE-18955
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li


Hit the issue when trying launch spark job. Stack trace:
{noformat}
Caused by: java.lang.NoSuchMethodError: 
io.netty.channel.DefaultChannelId.newInstance()Lio/netty/channel/DefaultChannelId;
at io.netty.channel.AbstractChannel.newId(AbstractChannel.java:111) 
~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at io.netty.channel.AbstractChannel.(AbstractChannel.java:83) 
~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at 
io.netty.channel.nio.AbstractNioChannel.(AbstractNioChannel.java:84) 
~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at 
io.netty.channel.nio.AbstractNioMessageChannel.(AbstractNioMessageChannel.java:42)
 ~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at 
io.netty.channel.socket.nio.NioServerSocketChannel.(NioServerSocketChannel.java:86)
 ~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at 
io.netty.channel.socket.nio.NioServerSocketChannel.(NioServerSocketChannel.java:72)
 ~[netty-all-4.1.17.Final.jar:4.1.17.Final]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
Method) ~[?:1.8.0_151]
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
 ~[?:1.8.0_151]
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
 ~[?:1.8.0_151]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423) 
~[?:1.8.0_151]
at 
io.netty.channel.ReflectiveChannelFactory.newChannel(ReflectiveChannelFactory.java:38)
 ~[netty-all-4.1.17.Final.jar:4.1.17.Final]
... 32 more
{noformat}

It seems we have conflicts versions of class 
{{io.netty.channel.DefaultChannelId}} from async-http-client.jar and 
netty-all.jar



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-18647) Cannot create table: Unknown column 'CREATION_METADATA_MV_CREATION_METADATA_ID_OID'

2018-02-07 Thread Rui Li (JIRA)
Rui Li created HIVE-18647:
-

 Summary: Cannot create table: Unknown column 
'CREATION_METADATA_MV_CREATION_METADATA_ID_OID'
 Key: HIVE-18647
 URL: https://issues.apache.org/jira/browse/HIVE-18647
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (HIVE-18442) HoS: No FileSystem for scheme: nullscan

2018-01-11 Thread Rui Li (JIRA)
Rui Li created HIVE-18442:
-

 Summary: HoS: No FileSystem for scheme: nullscan
 Key: HIVE-18442
 URL: https://issues.apache.org/jira/browse/HIVE-18442
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-18282) Spark tar is downloaded every time for itest

2017-12-14 Thread Rui Li (JIRA)
Rui Li created HIVE-18282:
-

 Summary: Spark tar is downloaded every time for itest
 Key: HIVE-18282
 URL: https://issues.apache.org/jira/browse/HIVE-18282
 Project: Hive
  Issue Type: Test
Reporter: Rui Li


Seems we missed the md5 file for spark-2.2.0?
cc [~kellyzly], [~stakiar]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: checkstyle changes

2017-12-07 Thread Rui Li
I also believe 140 is a little too long.

BTW do we use 2 or 4 chars for continuation indent? I personally prefer 4,
but I do find both cases in out code.

On Fri, Dec 8, 2017 at 6:20 AM, Alexander Kolbasov <ak...@cloudera.com>
wrote:

> Problem with 140-wide code isn't just laptops - in many cases we need to do
> side-by-side diffs (e.g. for code reviews) and this doubles the required
> size.
>
> - Alex.
>
> On Thu, Dec 7, 2017 at 1:38 PM, Sergey Shelukhin <ser...@hortonworks.com>
> wrote:
>
> > I think the 140-character change will make the code hard to use on a
> > laptop without a monitor.
> >
> >
> > On 17/12/7, 02:43, "Peter Vary" <pv...@cloudera.com> wrote:
> >
> > >Disclaimer: I did not have time to test it out, but according to
> > >http://checkstyle.sourceforge.net/config_misc.html#Indentation
> > ><http://checkstyle.sourceforge.net/config_misc.html#Indentation>
> > >Maybe the indentation could be solved by:
> > >lineWrappingIndentation=2 (default 4)
> > >forceStrictCondition=false (default false)
> > >
> > >http://checkstyle.sourceforge.net/config_misc.html#TrailingComment
> > ><http://checkstyle.sourceforge.net/config_misc.html#TrailingComment>
> > >might help with the comments
> > >
> > >Sorry for not being more helpful. Maybe sometime later I will have time
> > >to check these out.
> > >
> > >Thanks,
> > >Peter
> > >
> > >> On Dec 7, 2017, at 10:26 AM, Zoltan Haindrich
> > >><zhaindr...@hortonworks.com> wrote:
> > >>
> > >> Hello Eugene!
> > >>
> > >> I've looked into doing something with these; but I was not able to
> > >>relieve the warnings you've mentioned:
> > >>
> > >> * the ;// is seems to be not configurable
> > >>   It seems like its handled by the whitespaceafter module; I'm not
> sure
> > >>how to allow / after ;
> > >> * I think that indentation of 4 for many method arguments makes it
> more
> > >>readable; so I think it would be the best to just drop this check...but
> > >>I've not seen any way to do this(w/o disabling the whole indentation
> > >>module...)
> > >>
> > >> maybe someone else should take a look at itI find it pretty hard
> to
> > >>get docs about specific chechkstyle configurations; since the search
> > >>keywords mostly contain keywords like: semicolon, whitespace,
> > >>comment...which tends to pull in all kind of garbage results :)
> > >>
> > >> cheers,
> > >> Zoltan
> > >>
> > >> On 6 Dec 2017 8:53 p.m., Eugene Koifman <ekoif...@hortonworks.com>
> > >>wrote:
> > >> It currently complains about no space between ; and // as in
> “…);//foo”
> > >>
> > >> And also about indentation when a single method call is split into
> > >>multiple lines.
> > >> It insists on 4 chars in this case, though we use 2 in (all?) other
> > >>cases.
> > >>
> > >> Could this be dialed down as well?
> > >>
> > >>
> > >> On 12/5/17, 7:26 AM, "Peter Vary" <pv...@cloudera.com> wrote:
> > >>
> > >>+1 for the changes
> > >>
> > >>> On Dec 5, 2017, at 1:02 PM, Zoltan Haindrich <k...@rxd.hu> wrote:
> > >>>
> > >>> Hello,
> > >>>
> > >>> I've filed a ticket to make the checkstyle warnings less noisy
> > >>>(https://issues.apache.org/jira/browse/HIVE-18222)
> > >>>
> > >>> * set maxlinelength to 140
> > >>>   I think everyone is working with big-enough displays to handle this
> > >>>:)
> > >>>   There are many methods which have complicated names / arguments /
> > >>>etc ; breaking the lines more frequently hurts readability...
> > >>> * disabled some restrictions like: declaration via get/set
> > >>>methods for protected/package fields are not mandatory
> > >>>
> > >>> If you don't feel comfortable with these changes, please share your
> > >>>point of view.
> > >>>
> > >>> cheers,
> > >>> Zoltan
> > >>>
> > >>>
> > >>
> > >>
> > >>
> > >>
> > >
> >
> >
>



-- 
Best regards!
Rui Li


[jira] [Created] (HIVE-18242) VectorizedRowBatch cast exception when analyzing partitioned table

2017-12-06 Thread Rui Li (JIRA)
Rui Li created HIVE-18242:
-

 Summary: VectorizedRowBatch cast exception when analyzing 
partitioned table
 Key: HIVE-18242
 URL: https://issues.apache.org/jira/browse/HIVE-18242
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li


Happens when I run the following (vectorization enabled):
{code}
ANALYZE TABLE srcpart PARTITION(ds, hr) COMPUTE STATISTICS;
{code}
The stack trace is:
{noformat}
Caused by: java.lang.ClassCastException: 
org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch cannot be cast to 
org.apache.hadoop.io.Text
at 
org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableStringObjectInspector.copyObject(WritableStringObjectInspector.java:36)
at 
org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.copyToStandardObject(ObjectInspectorUtils.java:425)
at 
org.apache.hadoop.hive.serde2.objectinspector.ObjectInspectorUtils.partialCopyToStandardObject(ObjectInspectorUtils.java:314)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.gatherStats(TableScanOperator.java:191)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:138)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.setupPartitionContextVars(VectorMapOperator.java:682)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.cleanUpInputFileChangedOp(VectorMapOperator.java:607)
at 
org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1187)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorMapOperator.process(VectorMapOperator.java:784)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-18148) NPE in SparkDynamicPartitionPruningResolver

2017-11-26 Thread Rui Li (JIRA)
Rui Li created HIVE-18148:
-

 Summary: NPE in SparkDynamicPartitionPruningResolver
 Key: HIVE-18148
 URL: https://issues.apache.org/jira/browse/HIVE-18148
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li


The stack trace is:
{noformat}
2017-11-27T10:32:38,752 ERROR [e6c8aab5-ddd2-461d-b185-a7597c3e7519 main] 
ql.Driver: FAILED: NullPointerException null
java.lang.NullPointerException
at 
org.apache.hadoop.hive.ql.optimizer.physical.SparkDynamicPartitionPruningResolver$SparkDynamicPartitionPruningDispatcher.dispatch(SparkDynamicPartitionPruningResolver.java:100)
at 
org.apache.hadoop.hive.ql.lib.TaskGraphWalker.dispatch(TaskGraphWalker.java:111)
at 
org.apache.hadoop.hive.ql.lib.TaskGraphWalker.walk(TaskGraphWalker.java:180)
at 
org.apache.hadoop.hive.ql.lib.TaskGraphWalker.startWalking(TaskGraphWalker.java:125)
at 
org.apache.hadoop.hive.ql.optimizer.physical.SparkDynamicPartitionPruningResolver.resolve(SparkDynamicPartitionPruningResolver.java:74)
at 
org.apache.hadoop.hive.ql.parse.spark.SparkCompiler.optimizeTaskPlan(SparkCompiler.java:568)
{noformat}
At this stage, there shouldn't be a DPP sink whose target map work is null. The 
root cause seems to be a malformed operator tree generated by SplitOpTreeForDPP.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-18129) The ConditionalResolverMergeFiles doesn't merge empty files

2017-11-21 Thread Rui Li (JIRA)
Rui Li created HIVE-18129:
-

 Summary: The ConditionalResolverMergeFiles doesn't merge empty 
files
 Key: HIVE-18129
 URL: https://issues.apache.org/jira/browse/HIVE-18129
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


If a query produces lots of empty files, these files won't be merged by the 
merge-small-file feature.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-18111) Fix temp path for Spark DPP sink

2017-11-20 Thread Rui Li (JIRA)
Rui Li created HIVE-18111:
-

 Summary: Fix temp path for Spark DPP sink
 Key: HIVE-18111
 URL: https://issues.apache.org/jira/browse/HIVE-18111
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-18041) Add SORT_QUERY_RESULTS to subquery_multi

2017-11-09 Thread Rui Li (JIRA)
Rui Li created HIVE-18041:
-

 Summary: Add SORT_QUERY_RESULTS to subquery_multi
 Key: HIVE-18041
 URL: https://issues.apache.org/jira/browse/HIVE-18041
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Priority: Trivial






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17976) HoS: don't set output collector if there's no data to process

2017-11-03 Thread Rui Li (JIRA)
Rui Li created HIVE-17976:
-

 Summary: HoS: don't set output collector if there's no data to 
process
 Key: HIVE-17976
 URL: https://issues.apache.org/jira/browse/HIVE-17976
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


MR doesn't set an output collector if no row is processed, i.e. 
{{ExecMapper::map}} is never called. Let's investigate whether Spark should do 
the same.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17964) HoS: some spark configs doesn't require re-creating a session

2017-11-02 Thread Rui Li (JIRA)
Rui Li created HIVE-17964:
-

 Summary: HoS: some spark configs doesn't require re-creating a 
session
 Key: HIVE-17964
 URL: https://issues.apache.org/jira/browse/HIVE-17964
 Project: Hive
  Issue Type: Improvement
Reporter: Rui Li
Priority: Minor


I guess the {{hive.spark.}} configs were initially intended for the RSC. 
Therefore when they're changed, we'll re-create the session for them to take 
effect. There're some configs not related to RSC that also start with 
{{hive.spark.}}. We'd better rename them so that we don't unnecessarily 
re-create sessions, which is usually time consuming.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17877) HoS: combine equivalent DPP sink works

2017-10-23 Thread Rui Li (JIRA)
Rui Li created HIVE-17877:
-

 Summary: HoS: combine equivalent DPP sink works
 Key: HIVE-17877
 URL: https://issues.apache.org/jira/browse/HIVE-17877
 Project: Hive
  Issue Type: Improvement
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17383) ArrayIndexOutOfBoundsException in VectorGroupByOperator

2017-08-24 Thread Rui Li (JIRA)
Rui Li created HIVE-17383:
-

 Summary: ArrayIndexOutOfBoundsException in VectorGroupByOperator
 Key: HIVE-17383
 URL: https://issues.apache.org/jira/browse/HIVE-17383
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li


Query to reproduce:
{noformat}
set hive.cbo.enable=false;
select count(*) from (select key from src group by key) s where s.key='98';
{noformat}
The stack trace is:
{noformat}
Caused by: java.lang.ArrayIndexOutOfBoundsException: 1
at 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupKeyHelper.copyGroupKey(VectorGroupKeyHelper.java:107)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator$ProcessingModeReduceMergePartial.doProcessBatch(VectorGroupByOperator.java:831)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator$ProcessingModeBase.processBatch(VectorGroupByOperator.java:174)
at 
org.apache.hadoop.hive.ql.exec.vector.VectorGroupByOperator.process(VectorGroupByOperator.java:1046)
at 
org.apache.hadoop.hive.ql.exec.tez.ReduceRecordSource.processVectorGroup(ReduceRecordSource.java:462)
... 18 more
{noformat}
More details can be found in HIVE-16823



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17321) HoS: analyze ORC table doesn't compute raw data size when noscan/partialscan is not specified

2017-08-15 Thread Rui Li (JIRA)
Rui Li created HIVE-17321:
-

 Summary: HoS: analyze ORC table doesn't compute raw data size when 
noscan/partialscan is not specified
 Key: HIVE-17321
 URL: https://issues.apache.org/jira/browse/HIVE-17321
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17193) HoS: don't combine map works that are targets of different DPPs

2017-07-27 Thread Rui Li (JIRA)
Rui Li created HIVE-17193:
-

 Summary: HoS: don't combine map works that are targets of 
different DPPs
 Key: HIVE-17193
 URL: https://issues.apache.org/jira/browse/HIVE-17193
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17133) NoSuchMethodError in Hadoop FileStatus.compareTo

2017-07-20 Thread Rui Li (JIRA)
Rui Li created HIVE-17133:
-

 Summary: NoSuchMethodError in Hadoop FileStatus.compareTo
 Key: HIVE-17133
 URL: https://issues.apache.org/jira/browse/HIVE-17133
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li


The stack trace is:
{noformat}
Caused by: java.lang.NoSuchMethodError: 
org.apache.hadoop.fs.FileStatus.compareTo(Lorg/apache/hadoop/fs/FileStatus;)I
at 
org.apache.hadoop.hive.ql.io.AcidUtils.lambda$getAcidState$0(AcidUtils.java:931)
at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
at java.util.TimSort.sort(TimSort.java:234)
at java.util.Arrays.sort(Arrays.java:1512)
at java.util.ArrayList.sort(ArrayList.java:1454)
at java.util.Collections.sort(Collections.java:175)
at 
org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:929)
{noformat}

I'm on Hive master and using Hadoop 2.7.2. The method signature in Hadoop 2.7.2 
is:
https://github.com/apache/hadoop/blob/release-2.7.2-RC2/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileStatus.java#L336
In Hadoop 2.8.0 it becomes:
https://github.com/apache/hadoop/blob/release-2.8.0-RC3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileStatus.java#L332
I think that breaks binary compatibility.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17114) HoS: Possible skew in shuffling when data is not really skewed

2017-07-17 Thread Rui Li (JIRA)
Rui Li created HIVE-17114:
-

 Summary: HoS: Possible skew in shuffling when data is not really 
skewed
 Key: HIVE-17114
 URL: https://issues.apache.org/jira/browse/HIVE-17114
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


Observed in HoS and may apply to other engines as well.
When we join 2 tables on a single int key, we use the key itself as hash code 
in {{ObjectInspectorUtils.hashCode}}:
{code}
  case INT:
return ((IntObjectInspector) poi).get(o);
{code}
Suppose the keys are different but are all some multiples of 10. And if we 
choose 10 as #reducers, the shuffle will be skewed.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17034) The spark tar for itests is downloaded every time if md5sum is not installed

2017-07-05 Thread Rui Li (JIRA)
Rui Li created HIVE-17034:
-

 Summary: The spark tar for itests is downloaded every time if 
md5sum is not installed
 Key: HIVE-17034
 URL: https://issues.apache.org/jira/browse/HIVE-17034
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li


I think we should either skip verifying md5, or fail the build to let developer 
know md5sum is required.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: Review Request 60632: HIVE-16659: Query plan should reflect hive.spark.use.groupby.shuffle

2017-07-04 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/60632/#review179595
---


Ship it!




Ship It!

- Rui Li


On July 5, 2017, 4:07 a.m., Bing Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/60632/
> ---
> 
> (Updated July 5, 2017, 4:07 a.m.)
> 
> 
> Review request for hive.
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> HIVE-16659: Query plan should reflect hive.spark.use.groupby.shuffle
> 
> 
> Diffs
> -
> 
>   itests/src/test/resources/testconfiguration.properties 19ff316 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RepartitionShuffler.java 
> d0c708c 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java 
> 5f85f9e 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java 
> b9901da 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkWork.java afbeccb 
>   ql/src/test/queries/clientpositive/spark_explain_groupbyshuffle.q 
> PRE-CREATION 
>   ql/src/test/results/clientpositive/spark/spark_explain_groupbyshuffle.q.out 
> PRE-CREATION 
> 
> 
> Diff: https://reviews.apache.org/r/60632/diff/2/
> 
> 
> Testing
> ---
> 
> set hive.spark.use.groupby.shuffle=true;
> explain select key, count(val) from t1 group by key;
> 
> STAGE DEPENDENCIES:
>   Stage-1 is a root stage
>   Stage-0 depends on stages: Stage-1
> 
> STAGE PLANS:
>   Stage: Stage-1
> Spark
>   Edges:
> Reducer 2 <- Map 1 (GROUP, 2)
>   DagName: root_20170625202742_58335619-7107-4026-9911-43d2ec449088:2
>   Vertices:
> Map 1
> Map Operator Tree:
> TableScan
>   alias: t1
>   Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
>   Select Operator
> expressions: key (type: int), val (type: string)
> outputColumnNames: key, val
> Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
> Group By Operator
>   aggregations: count(val)
>   keys: key (type: int)
>   mode: hash
>   outputColumnNames: _col0, _col1
>   Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> key expressions: _col0 (type: int)
> sort order: +
> Map-reduce partition columns: _col0 (type: int)
> Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col1 (type: bigint)
> Reducer 2
> Reduce Operator Tree:
>   Group By Operator
> aggregations: count(VALUE._col0)
> keys: KEY._col0 (type: int)
> mode: mergepartial
> outputColumnNames: _col0, _col1
> Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
> Column stats: NONE
> File Output Operator
>   compressed: false
>   Statistics: Num rows: 10 Data size: 70 Basic stats: 
> COMPLETE Column stats: NONE
>   table:
>   input format: 
> org.apache.hadoop.mapred.SequenceFileInputFormat
>   output format: 
> org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
>   serde: 
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
> 
>   Stage: Stage-0
> Fetch Operator
>   limit: -1
>   Processor Tree:
> ListSink
> 
> 
> set hive.spark.use.groupby.shuffle=false;
> explain select key, count(val) from t1 group by key;
> 
> STAGE DEPENDENCIES:
>   Stage-1 is a root stage
>   Stage-0 depends on stages: Stage-1
> 
> STAGE PLANS:
>   Stage: Stage-1
> Spark
>   Edges:
> Reducer 2 <- Map 1 (GROUP, 2)
>   DagName: root_20170625203122_3afe01dd-41cc-477e-9098-ddd58b37ad4e:3
>   Vertices:
> Map 1
> Map Operator Tree:
> TableScan
>   alias: t1
>   Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Colum

[jira] [Created] (HIVE-17020) Aggressive RS dedup can incorrectly remove OP tree branch

2017-07-04 Thread Rui Li (JIRA)
Rui Li created HIVE-17020:
-

 Summary: Aggressive RS dedup can incorrectly remove OP tree branch
 Key: HIVE-17020
 URL: https://issues.apache.org/jira/browse/HIVE-17020
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


Suppose we have an OP tree like this:
{noformat}
 ...
  |
 RS[1]
  |
SEL[2]
/\
SEL[3]   SEL[4]
  | |
RS[5] FS[6]
  |
 ... 
{noformat}
When doing aggressive RS dedup, we'll remove all the operators between RS5 and 
RS1, and thus the branch containing FS6 is lost.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: Review Request 60632: HIVE-16659: Query plan should reflect hive.spark.use.groupby.shuffle

2017-07-04 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/60632/#review179554
---




ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
Lines 68 (patched)
<https://reviews.apache.org/r/60632/#comment254315>

Please avoid * import



ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
Lines 432 (patched)
<https://reviews.apache.org/r/60632/#comment254316>

it's preferable to use HiveConf::getBoolVar



ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
Line 438 (original), 441 (patched)
<https://reviews.apache.org/r/60632/#comment254317>

nit: extra space before !useSparkGroupBy



ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java
Line 471 (original), 477 (patched)
<https://reviews.apache.org/r/60632/#comment254319>

let's delete this comment


- Rui Li


On July 4, 2017, 8:48 a.m., Bing Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/60632/
> ---
> 
> (Updated July 4, 2017, 8:48 a.m.)
> 
> 
> Review request for hive.
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> HIVE-16659: Query plan should reflect hive.spark.use.groupby.shuffle
> 
> 
> Diffs
> -
> 
>   itests/src/test/resources/testconfiguration.properties 19ff316 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RepartitionShuffler.java 
> d0c708c 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkPlanGenerator.java 
> 5f85f9e 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkUtils.java 
> b9901da 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/spark/GenSparkWork.java afbeccb 
>   ql/src/test/queries/clientpositive/spark_explain_groupbyshuffle.q 
> PRE-CREATION 
>   ql/src/test/results/clientpositive/spark/spark_explain_groupbyshuffle.q.out 
> PRE-CREATION 
> 
> 
> Diff: https://reviews.apache.org/r/60632/diff/1/
> 
> 
> Testing
> ---
> 
> set hive.spark.use.groupby.shuffle=true;
> explain select key, count(val) from t1 group by key;
> 
> STAGE DEPENDENCIES:
>   Stage-1 is a root stage
>   Stage-0 depends on stages: Stage-1
> 
> STAGE PLANS:
>   Stage: Stage-1
> Spark
>   Edges:
> Reducer 2 <- Map 1 (GROUP, 2)
>   DagName: root_20170625202742_58335619-7107-4026-9911-43d2ec449088:2
>   Vertices:
> Map 1
> Map Operator Tree:
> TableScan
>   alias: t1
>   Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
>   Select Operator
> expressions: key (type: int), val (type: string)
> outputColumnNames: key, val
> Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
> Group By Operator
>   aggregations: count(val)
>   keys: key (type: int)
>   mode: hash
>   outputColumnNames: _col0, _col1
>   Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> key expressions: _col0 (type: int)
> sort order: +
> Map-reduce partition columns: _col0 (type: int)
> Statistics: Num rows: 20 Data size: 140 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col1 (type: bigint)
> Reducer 2
> Reduce Operator Tree:
>   Group By Operator
> aggregations: count(VALUE._col0)
> keys: KEY._col0 (type: int)
> mode: mergepartial
> outputColumnNames: _col0, _col1
> Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
> Column stats: NONE
> File Output Operator
>   compressed: false
>   Statistics: Num rows: 10 Data size: 70 Basic stats: 
> COMPLETE Column stats: NONE
>   table:
>   input format: 
> org.apache.hadoop.mapred.SequenceFileInputFormat
>   output format: 
> org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
>   serde: 
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
> 
>   

[jira] [Created] (HIVE-16876) RpcServer should be re-created when Rpc configs change

2017-06-10 Thread Rui Li (JIRA)
Rui Li created HIVE-16876:
-

 Summary: RpcServer should be re-created when Rpc configs change
 Key: HIVE-16876
 URL: https://issues.apache.org/jira/browse/HIVE-16876
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (HIVE-16767) Update people website with recent changes

2017-05-26 Thread Rui Li (JIRA)
Rui Li created HIVE-16767:
-

 Summary: Update people website with recent changes
 Key: HIVE-16767
 URL: https://issues.apache.org/jira/browse/HIVE-16767
 Project: Hive
  Issue Type: Task
  Components: Documentation
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Welcome Rui Li to Hive PMC

2017-05-25 Thread Rui Li
Thank you guys :)

On Thu, May 25, 2017 at 3:29 PM, Peter Vary <pv...@cloudera.com> wrote:

> Congratulations Rui!
>
> > On May 25, 2017, at 6:19 AM, Xuefu Zhang <xu...@apache.org> wrote:
> >
> > Hi all,
> >
> > It's an honer to announce that Apache Hive PMC has recently voted to
> invite
> > Rui Li as a new Hive PMC member. Rui is a long time Hive contributor and
> > committer, and has made significant contribution in Hive especially in
> Hive
> > on Spark. Please join me in congratulating him and looking forward to a
> > bigger role that he will play in Apache Hive project.
> >
> > Thanks,
> > Xuefu
>
>


-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


[jira] [Created] (HIVE-16739) HoS DPP generates malformed plan when hive.tez.dynamic.semijoin.reduction is on

2017-05-23 Thread Rui Li (JIRA)
Rui Li created HIVE-16739:
-

 Summary: HoS DPP generates malformed plan when 
hive.tez.dynamic.semijoin.reduction is on
 Key: HIVE-16739
 URL: https://issues.apache.org/jira/browse/HIVE-16739
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li


HoS DPP currently can't handle dynamic semi join and will result in 
{{ClassCastException org.apache.hadoop.hive.ql.plan.ReduceWork cannot be cast 
to org.apache.hadoop.hive.ql.plan.MapWork}}.
We should either disable or implement it for HoS.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (HIVE-16659) Query plan should reflect hive.spark.use.groupby.shuffle

2017-05-11 Thread Rui Li (JIRA)
Rui Li created HIVE-16659:
-

 Summary: Query plan should reflect hive.spark.use.groupby.shuffle
 Key: HIVE-16659
 URL: https://issues.apache.org/jira/browse/HIVE-16659
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


It's useful to show the shuffle type used in the query plan. Currently it shows 
"GROUP" no matter what we set for hive.spark.use.groupby.shuffle.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: [Announce] New committer: Vineet Garg

2017-05-10 Thread Rui Li
Congrats :)

On Wed, May 10, 2017 at 2:06 PM, Zoltan Haindrich <
zhaindr...@hortonworks.com> wrote:

> Congratulations!
>
>
> On 10 May 2017 7:57 a.m., Prasanth Jayachandran <
> pjayachand...@hortonworks.com> wrote:
> Congratulations Vineeth!!
>
> Thanks
> Prasanth
>
>
>
> On Tue, May 9, 2017 at 10:52 PM -0700, "Jesus Camacho Rodriguez" <
> jcama...@apache.org<mailto:jcama...@apache.org>> wrote:
>
>
> Congrats Vineet! Well deserved!
>
> --
> Jesús
>
>
>
>
>
> On 5/10/17, 6:45 AM, "Peter Vary"  wrote:
>
> >Congratulations Vineet! :)
> >
> >2017. máj. 9. 22:25 ezt írta ("Ashutosh Chauhan" ):
> >
> >> The Project Management Committee (PMC) for Apache Hive has invited
> Vineet
> >> Garg to become a committer and we are pleased to announce that he has
> >> accepted.
> >>
> >> Welcome, Vineet!
> >>
> >> Thanks,
> >> Ashutosh
> >>
>
>
>
>
>


-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-09 Thread Rui Li
/BinarySortableSerDe.java
 89e15c3 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
56b4ca3 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
8237b64 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 ba44bae 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 9642a7e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampTZWritable.java 
PRE-CREATION 
  service-rpc/if/TCLIService.thrift 824b049 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h 8accf66 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp b6995c4 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 18a7825 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php ee5acd2 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py e9faa2a 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 714309c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 


Diff: https://reviews.apache.org/r/50787/diff/9/

Changes: https://reviews.apache.org/r/50787/diff/8-9/


Testing
---


Thanks,

Rui Li



Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-09 Thread Rui Li


> On May 9, 2017, 11:05 p.m., Ashutosh Chauhan wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/TypeConverter.java
> > Lines 204 (patched)
> > <https://reviews.apache.org/r/50787/diff/8/?file=1710538#file1710538line204>
> >
> > Can you file a bug in Calcite that it should have sql type to represent 
> > TS w TZ?

Filed CALCITE-1784 for it


> On May 9, 2017, 11:05 p.m., Ashutosh Chauhan wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToString.java
> > Lines 160 (patched)
> > <https://reviews.apache.org/r/50787/diff/8/?file=1710546#file1710546line160>
> >
> > Add a comment that string reperesentation will return TS in UTC zone 
> > and not in original TZ.

Here we convert timestamptz to string, which means TZ is already in UTC. I will 
add the comment when we convert string to timestamptz.


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/50787/#review174383
-------


On May 8, 2017, 3:17 p.m., Rui Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/50787/
> ---
> 
> (Updated May 8, 2017, 3:17 p.m.)
> 
> 
> Review request for hive, pengcheng xiong and Xuefu Zhang.
> 
> 
> Bugs: HIVE-14412
> https://issues.apache.org/jira/browse/HIVE-14412
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> The 1st patch to add timezone-aware timestamp.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
> PRE-CREATION 
>   common/src/test/org/apache/hadoop/hive/common/type/TestTimestampTZ.java 
> PRE-CREATION 
>   contrib/src/test/queries/clientnegative/serde_regex.q a676338 
>   contrib/src/test/queries/clientpositive/serde_regex.q d75d607 
>   contrib/src/test/results/clientnegative/serde_regex.q.out 58b1c02 
>   contrib/src/test/results/clientpositive/serde_regex.q.out 2984293 
>   hbase-handler/src/test/queries/positive/hbase_timestamp.q 0350afe 
>   hbase-handler/src/test/results/positive/hbase_timestamp.q.out 3918121 
>   itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q 
> 358eccd 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
>  c462538 
>   itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q 
> c563d3a 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
>  d17c281 
>   itests/hive-blobstore/src/test/results/clientpositive/orc_format_part.q.out 
> 5d1319f 
>   
> itests/hive-blobstore/src/test/results/clientpositive/orc_nonstd_partitions_loc.q.out
>  70e72f7 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_format_part.q.out
>  bed10ab 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_nonstd_partitions_loc.q.out
>  c6442f9 
>   jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java ade1900 
>   jdbc/src/java/org/apache/hive/jdbc/JdbcColumn.java 38918f0 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 1b556ac 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/GroupByOperator.java f8b55da 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/SerializationUtilities.java 
> 01a652d 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/TypeConverter.java
>  38308c9 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
> 0cf9205 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 190b66b 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g ca639d3 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 645ced9 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java 
> c3227c9 
>   ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java bda2050 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java 7cdf2c3 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToString.java 5cacd59 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java 68d98f5 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFDate.java 
> 5a31e61 
>   
> ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToTimestampTZ.java
>  PRE-CREATION 
>   
> ql/src/test/org/apache/hadoop/hive/ql/parse/TestSQL11ReservedKeyWordsNegative.java
>  0dc6b19 
>   ql/src/test/queries/clientnegative/serde_regex.q c9cfc7d 
>   ql/src/test/queries/clientnegative/serde_regex2.q a

[jira] [Created] (HIVE-16613) SaslClientHandler.sendHello is eating exceptions

2017-05-09 Thread Rui Li (JIRA)
Rui Li created HIVE-16613:
-

 Summary: SaslClientHandler.sendHello is eating exceptions
 Key: HIVE-16613
 URL: https://issues.apache.org/jira/browse/HIVE-16613
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-08 Thread Rui Li
/BinarySortableSerDe.java
 89e15c3 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
56b4ca3 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
8237b64 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 ba44bae 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 9642a7e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampTZWritable.java 
PRE-CREATION 
  service-rpc/if/TCLIService.thrift 824b049 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h 8accf66 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp b6995c4 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 18a7825 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php ee5acd2 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py e9faa2a 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 714309c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 


Diff: https://reviews.apache.org/r/50787/diff/8/

Changes: https://reviews.apache.org/r/50787/diff/7-8/


Testing
---


Thanks,

Rui Li



Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-08 Thread Rui Li


> On May 7, 2017, 11:22 p.m., Xuefu Zhang wrote:
> > common/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java
> > Lines 138 (patched)
> > <https://reviews.apache.org/r/50787/diff/7/?file=1706764#file1706764line138>
> >
> > Not sure if I understand this, but why cannot we get seconds/nanos from 
> > date/timestamp and then convert to TimestapTZ? I assume this is a faster 
> > way.

Hi Xuefu, the reason why I did this:

1. As Ashutosh suggested, we will use LocalDate and LocalDateTime for Date and 
Timestamp in the future. When that happens, date/timestamp won't have 
seconds/nanos part, instead they're only descriptions of time. So the 
conversion should be done based on text format.
2. At the moment, the seconds/nanos of date/timestamp is computed using system 
timezone. So the conversion can have different results in different systems.

I noted Carter also suggested that SQL standard requires session zone should be 
taken into consideration in the conversion.
Consolidating your suggestions with Carter's, I think we can: make the 
conversion text-wise, and append the system zone (Hive currently doesn't have 
session zone). For example, a date of '2017-01-01' in LA will be converted to 
timestamptz as '2017-01-01 00:00:00 America/Los_Angeles', which in turn 
converted to '2017-01-01 08:00:00.0 Z'. Does this make sense?


> On May 7, 2017, 11:22 p.m., Xuefu Zhang wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/TypeConverter.java
> > Lines 204 (patched)
> > <https://reviews.apache.org/r/50787/diff/7/?file=1706785#file1706785line204>
> >
> > What does this imply?

The method converts our primitive type to a SqlTypeName in calcite. But 
SqlTypeName currently doesn't have timestamp with time zone. This will have 
some impact when calcite does the optimization, e.g. computing average value 
sizes. But I think we have to live with it untile SqlTypeName supports 
timestamp with time zone.


> On May 7, 2017, 11:22 p.m., Xuefu Zhang wrote:
> > serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java
> > Lines 32 (patched)
> > <https://reviews.apache.org/r/50787/diff/7/?file=1706827#file1706827line32>
> >
> > Can you also make a note about the source of the code, like 
> > TimeStampTZWritable?

sure, will do


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/50787/#review174136
---


On May 8, 2017, 6:51 a.m., Rui Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/50787/
> ---
> 
> (Updated May 8, 2017, 6:51 a.m.)
> 
> 
> Review request for hive, pengcheng xiong and Xuefu Zhang.
> 
> 
> Bugs: HIVE-14412
> https://issues.apache.org/jira/browse/HIVE-14412
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> The 1st patch to add timezone-aware timestamp.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
> PRE-CREATION 
>   common/src/test/org/apache/hadoop/hive/common/type/TestTimestampTZ.java 
> PRE-CREATION 
>   contrib/src/test/queries/clientnegative/serde_regex.q a676338 
>   contrib/src/test/queries/clientpositive/serde_regex.q d75d607 
>   contrib/src/test/results/clientnegative/serde_regex.q.out 58b1c02 
>   contrib/src/test/results/clientpositive/serde_regex.q.out 2984293 
>   hbase-handler/src/test/queries/positive/hbase_timestamp.q 0350afe 
>   hbase-handler/src/test/results/positive/hbase_timestamp.q.out 3918121 
>   itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q 
> 358eccd 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
>  c462538 
>   itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q 
> c563d3a 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
>  d17c281 
>   itests/hive-blobstore/src/test/results/clientpositive/orc_format_part.q.out 
> 5d1319f 
>   
> itests/hive-blobstore/src/test/results/clientpositive/orc_nonstd_partitions_loc.q.out
>  70e72f7 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_format_part.q.out
>  bed10ab 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_nonstd_partitions_loc.q.out
>  c6442f9 
>   jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java ade1900 
>   jdbc/src/java/org/apache/hive/jdbc/JdbcColumn.java 38918f0 
>   ql/src/ja

Re: Welcome new Hive committer, Zhihai Xu

2017-05-07 Thread Rui Li
Congrats Zhihai :)

On Sun, May 7, 2017 at 3:41 PM, Lefty Leverenz <leftylever...@gmail.com>
wrote:

> Congratulations Zhihai!
>
> -- Lefty
>
>
> On Sat, May 6, 2017 at 2:06 AM, Zoltan Haindrich <
> zhaindr...@hortonworks.com
> > wrote:
>
> > Congratulations Zhihai!
> >
> > On 6 May 2017 9:52 a.m., Mohammad Islam <misla...@yahoo.com.INVALID>
> > wrote:
> > Congrats Zhihai!!
> >
> > On Friday, May 5, 2017 9:52 AM, Xuefu Zhang <xu...@apache.org>
> wrote:
> >
> >
> >  Hi all,
> >
> > I'm very please to announce that Hive PMC has recently voted to offer
> > Zhihai a committership which he accepted. Please join me in
> congratulating
> > on this recognition and thanking him for his contributions to Hive.
> >
> > Regards,
> > Xuefu
> >
> >
> >
> >
> >
>



-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


[jira] [Created] (HIVE-16593) SparkClientFactory.stop may prevent JVM from exiting

2017-05-05 Thread Rui Li (JIRA)
Rui Li created HIVE-16593:
-

 Summary: SparkClientFactory.stop may prevent JVM from exiting
 Key: HIVE-16593
 URL: https://issues.apache.org/jira/browse/HIVE-16593
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-03 Thread Rui Li


> On May 3, 2017, 9:57 p.m., pengcheng xiong wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g
> > Lines 132 (patched)
> > <https://reviews.apache.org/r/50787/diff/7/?file=1706787#file1706787line132>
> >
> > I think Identifier["timestamptz"] and Identifier["zone"] may be 
> > sufficient. It is not necessary to make them as key words and then add them 
> > back as identifiers. You can have a try and see if it works. Thanks..

Hi Pengcheng, sorry I'm quite ignorant about antlr. Could you please be more 
specific how to add the Identifiers? Let me explain what I intend to do. The 
new data type is named "timestamp with time zone", and "timestamptz" is added 
as a type alias. I thought it's required to add key words for type names. And 
according to the PostgreSQL doc we referenced 
(https://www.postgresql.org/docs/9.5/static/sql-keywords-appendix.html), "zone" 
is a non-reserved SQL key word and "timestamptz" is not a key word. So I added 
them in IdentifierParser.g as nonReserved.


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/50787/#review173833
---


On May 3, 2017, 6:34 a.m., Rui Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/50787/
> ---
> 
> (Updated May 3, 2017, 6:34 a.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-14412
> https://issues.apache.org/jira/browse/HIVE-14412
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> The 1st patch to add timezone-aware timestamp.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
> PRE-CREATION 
>   common/src/test/org/apache/hadoop/hive/common/type/TestTimestampTZ.java 
> PRE-CREATION 
>   contrib/src/test/queries/clientnegative/serde_regex.q a676338 
>   contrib/src/test/queries/clientpositive/serde_regex.q d75d607 
>   contrib/src/test/results/clientnegative/serde_regex.q.out 58b1c02 
>   contrib/src/test/results/clientpositive/serde_regex.q.out 2984293 
>   hbase-handler/src/test/queries/positive/hbase_timestamp.q 0350afe 
>   hbase-handler/src/test/results/positive/hbase_timestamp.q.out 3918121 
>   itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q 
> 358eccd 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
>  c462538 
>   itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q 
> c563d3a 
>   
> itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
>  d17c281 
>   itests/hive-blobstore/src/test/results/clientpositive/orc_format_part.q.out 
> 5d1319f 
>   
> itests/hive-blobstore/src/test/results/clientpositive/orc_nonstd_partitions_loc.q.out
>  70e72f7 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_format_part.q.out
>  bed10ab 
>   
> itests/hive-blobstore/src/test/results/clientpositive/rcfile_nonstd_partitions_loc.q.out
>  c6442f9 
>   jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java ade1900 
>   jdbc/src/java/org/apache/hive/jdbc/JdbcColumn.java 38918f0 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 8dc5f2e 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/GroupByOperator.java f8b55da 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/SerializationUtilities.java 
> 01a652d 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/TypeConverter.java
>  38308c9 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
> 0cf9205 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 0721b92 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g d98a663 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 8598fae 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java 
> 8f8eab0 
>   ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java bda2050 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java 7cdf2c3 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToString.java 5cacd59 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java 68d98f5 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFDate.java 
> 5a31e61 
>   
> ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToTimestampTZ.java
>  PRE-CREATION 
>   
> ql/src/test/org/apache/hadoop/hive/ql/parse/TestSQL11Res

[jira] [Created] (HIVE-16573) In-place update for HoS can't be disabled

2017-05-03 Thread Rui Li (JIRA)
Rui Li created HIVE-16573:
-

 Summary: In-place update for HoS can't be disabled
 Key: HIVE-16573
 URL: https://issues.apache.org/jira/browse/HIVE-16573
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


{{hive.spark.exec.inplace.progress}} has no effect



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Review Request 58865: HIVE-16552: Limit the number of tasks a Spark job may contain

2017-05-03 Thread Rui Li


> On May 3, 2017, 3:35 a.m., Rui Li wrote:
> >

Xuefu, the patch looks good to me overall. Thanks for the work. Do you think we 
should add some negative test case for it?


> On May 3, 2017, 3:35 a.m., Rui Li wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java
> > Lines 132 (patched)
> > <https://reviews.apache.org/r/58865/diff/3/?file=1705971#file1705971line132>
> >
> > I think the log is unnecessary because the failure should already be 
> > logged in the monitor
> 
> Xuefu Zhang wrote:
> This is not new code.

Do you mean "LOG.info("Failed to submit Spark job " + sparkJobID);" is not new 
code? I don't find it in the current SparkTask.java.


> On May 3, 2017, 3:35 a.m., Rui Li wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java
> > Lines 135 (patched)
> > <https://reviews.apache.org/r/58865/diff/3/?file=1705971#file1705971line135>
> >
> > Same as above. Can we consolidate the logs a bit?
> 
> Xuefu Zhang wrote:
> Jobmonitor prints it on console, while the log here is written to 
> hive.log.

The console.printInfo method does both printing and logging:

public void printInfo(String info, String detail, boolean isSilent) {
  if (!isSilent) {
getInfoStream().println(info);
  }
  LOG.info(info + StringUtils.defaultString(detail));
}


> On May 3, 2017, 3:35 a.m., Rui Li wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
> > Lines 104 (patched)
> > <https://reviews.apache.org/r/58865/diff/3/?file=1705972#file1705972line104>
> >
> > Maybe I was being misleading. I mean we can compute the total task only 
> > once when the job first reaches RUNNING state, i.e. in the "if (!running)". 
> > At this point, the total count is determined and won't change.
> 
> Xuefu Zhang wrote:
> Yeah. However, I'd like to keep the state transition to running first 
> before breaking up and returning rc=4. In fact, if we lose the transition, 
> Hive actually goes into an instable state. What you said was what I tried in 
> first place.

I see. Thanks for the explanation.


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/58865/#review173689
---


On May 2, 2017, 6:49 p.m., Xuefu Zhang wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/58865/
> ---
> 
> (Updated May 2, 2017, 6:49 p.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-16552
> https://issues.apache.org/jira/browse/HIVE-16552
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See JIRA description
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 84398c6 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java 32a7730 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
>  dd73f3e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/SparkJobMonitor.java 
> 0b224f2 
> 
> 
> Diff: https://reviews.apache.org/r/58865/diff/3/
> 
> 
> Testing
> ---
> 
> Test locally
> 
> 
> Thanks,
> 
> Xuefu Zhang
> 
>



Re: Review Request 50787: Add a timezone-aware timestamp

2017-05-03 Thread Rui Li
/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
56b4ca3 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
8237b64 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 ba44bae 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 9642a7e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampTZWritable.java 
PRE-CREATION 
  service-rpc/if/TCLIService.thrift 824b049 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h 8accf66 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp b6995c4 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 18a7825 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php ee5acd2 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py e9faa2a 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 714309c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 


Diff: https://reviews.apache.org/r/50787/diff/7/

Changes: https://reviews.apache.org/r/50787/diff/6-7/


Testing
---


Thanks,

Rui Li



Re: [VOTE] Apache Hive 2.3.0 Release Candidate 0

2017-05-02 Thread Rui Li
The patch has been reverted in master and branch-2.3

On Wed, May 3, 2017 at 3:01 AM, Sergio Pena <sergio.p...@cloudera.com>
wrote:

> Hi Pengcheng,
>
> There is a request from the HDFS team to revert the patch committed on
> HIVE-16047 from
> our code because it might cause problems when future Hadoop versions are
> released due to being a
> private API on Hadoop. This API method signature has been changed between
> releases, and
> we don't want to have additional shims to support future Hadoop versions
> just for this method.
>
> I'd like to revert it from 2.3.0 release before doing the release. It is
> marked as being fixed on 2.2 but it is not cherry-picked on branch-2.2 but
> branch-2.3.
>
> Do you agree?
>
> - Sergio
>
> On Fri, Apr 28, 2017 at 1:40 PM, Pengcheng Xiong <pxi...@apache.org>
> wrote:
>
> > Withdraw the VOTE on candidate 0. Will propose candidate 1 soon. Thanks.
> >
> > On Thu, Apr 27, 2017 at 8:10 PM, Owen O'Malley <owen.omal...@gmail.com>
> > wrote:
> >
> > > -1 you need a release of storage-API first.
> > >
> > > .. Owen
> > >
> > > > On Apr 27, 2017, at 17:43, Pengcheng Xiong <pxi...@apache.org>
> wrote:
> > > >
> > > > Apache Hive 2.3.0 Release Candidate 0 is available here:
> > > > http://home.apache.org/~pxiong/apache-hive-2.3.0-rc0/
> > > >
> > > >
> > > > Maven artifacts are available here:
> > > > https://repository.apache.org/content/repositories/
> orgapachehive-1073/
> > > >
> > > >
> > > > Source tag for RC0 is at:
> > > >
> > > > https://github.com/apache/hive/releases/tag/release-2.3.0-rc0
> > > >
> > > > Voting will conclude in 72 hours.
> > > >
> > > > Hive PMC Members: Please test and vote.
> > > >
> > > > Thanks.
> > >
> >
>



-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Re: Review Request 58865: HIVE-16552: Limit the number of tasks a Spark job may contain

2017-05-02 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/58865/#review173689
---




ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java
Lines 132 (patched)
<https://reviews.apache.org/r/58865/#comment246728>

I think the log is unnecessary because the failure should already be logged 
in the monitor



ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java
Lines 135 (patched)
<https://reviews.apache.org/r/58865/#comment246729>

Same as above. Can we consolidate the logs a bit?



ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
Lines 104 (patched)
<https://reviews.apache.org/r/58865/#comment246731>

Maybe I was being misleading. I mean we can compute the total task only 
once when the job first reaches RUNNING state, i.e. in the "if (!running)". At 
this point, the total count is determined and won't change.


- Rui Li


On May 2, 2017, 6:49 p.m., Xuefu Zhang wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/58865/
> ---
> 
> (Updated May 2, 2017, 6:49 p.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-16552
> https://issues.apache.org/jira/browse/HIVE-16552
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See JIRA description
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 84398c6 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java 32a7730 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
>  dd73f3e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/SparkJobMonitor.java 
> 0b224f2 
> 
> 
> Diff: https://reviews.apache.org/r/58865/diff/3/
> 
> 
> Testing
> ---
> 
> Test locally
> 
> 
> Thanks,
> 
> Xuefu Zhang
> 
>



Re: Review Request 58865: HIVE-16552: Limit the number of tasks a Spark job may contain

2017-05-01 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/58865/#review173556
---




ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java
Lines 135 (patched)
<https://reviews.apache.org/r/58865/#comment246543>

The log is incorrect because cancelling the job doesn't mean killing the 
application.



ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
Lines 106 (patched)
<https://reviews.apache.org/r/58865/#comment246544>

I think the total task count needs only be computed once. It shouldn't 
change during the execution of the job, assuming we don't count failed/retried 
tasks.


- Rui Li


On May 1, 2017, 5:13 p.m., Xuefu Zhang wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/58865/
> ---
> 
> (Updated May 1, 2017, 5:13 p.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-16552
> https://issues.apache.org/jira/browse/HIVE-16552
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See JIRA description
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java d3ea824 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkTask.java 32a7730 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/RemoteSparkJobMonitor.java
>  dd73f3e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/SparkJobMonitor.java 
> 0b224f2 
> 
> 
> Diff: https://reviews.apache.org/r/58865/diff/2/
> 
> 
> Testing
> ---
> 
> Test locally
> 
> 
> Thanks,
> 
> Xuefu Zhang
> 
>



Re: Review Request 50787: Add a timezone-aware timestamp

2017-04-26 Thread Rui Li
/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
56b4ca3 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
8237b64 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 ba44bae 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 9642a7e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampTZWritable.java 
PRE-CREATION 
  service-rpc/if/TCLIService.thrift 824b049 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h 8accf66 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp b6995c4 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 18a7825 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php ee5acd2 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py e9faa2a 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 714309c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 


Diff: https://reviews.apache.org/r/50787/diff/6/

Changes: https://reviews.apache.org/r/50787/diff/5-6/


Testing
---


Thanks,

Rui Li



[jira] [Created] (HIVE-16459) Cancel outstanding RPCs when channel closes

2017-04-17 Thread Rui Li (JIRA)
Rui Li created HIVE-16459:
-

 Summary: Cancel outstanding RPCs when channel closes
 Key: HIVE-16459
 URL: https://issues.apache.org/jira/browse/HIVE-16459
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (HIVE-16418) Allow HiveKey to skip some bytes for comparison

2017-04-11 Thread Rui Li (JIRA)
Rui Li created HIVE-16418:
-

 Summary: Allow HiveKey to skip some bytes for comparison
 Key: HIVE-16418
 URL: https://issues.apache.org/jira/browse/HIVE-16418
 Project: Hive
  Issue Type: New Feature
Reporter: Rui Li
Assignee: Rui Li


The feature is required when we have to serialize some fields and prevent them 
from being used in comparison, e.g. HIVE-14412.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (HIVE-16315) Describe table doesn't show num of partitions

2017-03-28 Thread Rui Li (JIRA)
Rui Li created HIVE-16315:
-

 Summary: Describe table doesn't show num of partitions
 Key: HIVE-16315
 URL: https://issues.apache.org/jira/browse/HIVE-16315
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


This doesn't comply with our wiki: 
https://cwiki.apache.org/confluence/display/Hive/StatsDev#StatsDev-Examples



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: [ANNOUNCE] New PMC Member : Eugene Koifman

2017-03-16 Thread Rui Li
Congratulations :)

On Thu, Mar 16, 2017 at 4:26 PM, Lefty Leverenz <leftylever...@gmail.com>
wrote:

> More congratulations!
>
> -- Lefty
>
> On Wed, Mar 15, 2017 at 1:27 PM, Eugene Koifman <ekoif...@hortonworks.com>
> wrote:
>
> > Thank you everyone!
> >
> > On 3/15/17, 12:21 PM, "Gunther Hagleitner" <ghagleit...@hortonworks.com>
> > wrote:
> >
> > Congratulations!
> > 
> > From: Sergey Shelukhin <ser...@hortonworks.com>
> > Sent: Wednesday, March 15, 2017 11:18 AM
> > To: dev@hive.apache.org
> > Subject: Re: [ANNOUNCE] New PMC Member : Eugene Koifman
> >
> > Congrats!
> >
> > On 17/3/15, 01:02, "Zoltan Haindrich" <zhaindr...@hortonworks.com>
> > wrote:
> >
> > >Congrats Eugene!!
> > >
> > >On 15 Mar 2017 07:50, Peter Vary <pv...@cloudera.com> wrote:
> > >Congratulations! :)
> > >
> > >2017. márc. 15. 7:05 ezt írta ("Vaibhav Gumashta"
> > ><vgumas...@hortonworks.com
> > >>):
> > >
> > >> Congrats Eugene!
> > >>
> > >>
> > >> On 3/14/17, 11:03 PM, "Rajesh Balamohan" <rbalamo...@apache.org>
> > wrote:
> > >>
> > >> >Congrats Eugene!! :)
> > >> >
> > >> >~Rajesh.B
> > >> >
> > >> >On Wed, Mar 15, 2017 at 11:21 AM, Pengcheng Xiong <
> > pxi...@apache.org>
> > >> >wrote:
> > >> >
> > >> >> Congrats! Well deserved!
> > >> >>
> > >> >> Thanks.
> > >> >> Pengcheng
> > >> >>
> > >> >> On Tue, Mar 14, 2017 at 10:39 PM, Ashutosh Chauhan
> > >> >><hashut...@apache.org>
> > >> >> wrote:
> > >> >>
> > >> >> > On behalf of the Hive PMC I am delighted to announce Eugene
> > >>Koifman is
> > >> >> > joining Hive PMC.
> > >> >> > Eugene is a long time contributor in Hive and is focusing on
> > ACID
> > >> >>support
> > >> >> > areas these days.
> > >> >> >
> > >> >> > Welcome, Eugene!
> > >> >> >
> > >> >> > Thanks,
> > >> >> > Ashutosh
> > >> >> >
> > >> >>
> > >>
> > >>
> > >
> >
> >
> >
> >
> >
>



-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Re: Review Request 57586: HIVE-16183: Fix potential thread safety issues with static variables

2017-03-15 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/57586/#review169094
---


Ship it!




Ship It!

- Rui Li


On March 15, 2017, 10:03 p.m., Xuefu Zhang wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/57586/
> ---
> 
> (Updated March 15, 2017, 10:03 p.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-16183
> https://issues.apache.org/jira/browse/HIVE-16183
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Please see JIRA description
> 
> 
> Diffs
> -
> 
>   beeline/src/java/org/apache/hive/beeline/BeeLineOpts.java 7e6846d 
>   beeline/src/java/org/apache/hive/beeline/HiveSchemaHelper.java 181f0d2 
>   beeline/src/java/org/apache/hive/beeline/HiveSchemaTool.java 2c088c9 
>   cli/src/java/org/apache/hadoop/hive/cli/RCFileCat.java f1806a0 
>   cli/src/test/org/apache/hadoop/hive/cli/TestRCFileCat.java 11ceb31 
>   common/src/java/org/apache/hadoop/hive/common/LogUtils.java c2a0d9a 
>   common/src/java/org/apache/hadoop/hive/common/StatsSetupConst.java 926b4a6 
>   
> metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreSchemaInfo.java 
> 9c30ee7 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/ArchiveUtils.java 6381a21 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 4ac25c2 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 6693134 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java 
> 5b0c2bf 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CuckooSetBytes.java
>  6383e8a 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/fast/VectorMapJoinFastHashTable.java
>  9030e5f 
>   ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java 6582cdd 
>   ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java a1408e9 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java 7727114 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java 4995bdf 
>   ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java d391164 
>   ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java 59682db 
>   ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java 
> 90b1dff 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/VirtualColumn.java 044d64c 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 0e67ea6 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunerUtils.java
>  4d3e74e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
>  93202c3 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java 
> 50eda15 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/VectorizerReason.java
>  e0a6198 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java 
> f762fee 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
> f175663 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/WindowingSpec.java 01b5559 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/AbstractVectorDesc.java e85a418 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/GroupByDesc.java 0b49294 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MapJoinDesc.java ca69697 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/ReduceSinkDesc.java 9ae30ab 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorAppMasterEventDesc.java 
> 2e11321 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorFileSinkDesc.java 325ac91 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorFilterDesc.java 6feed84 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorGroupByDesc.java f8554e2 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorLimitDesc.java c9bc45a 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorMapJoinDesc.java 3aa65d3 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorMapJoinInfo.java 9429785 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorPartitionDesc.java 4078c7d 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorReduceSinkDesc.java 
> 2eb44b8 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorReduceSinkInfo.java 
> 8c35415 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorSMBJoinDesc.java 031f11e 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorSelectDesc.java c2c9450 
>   
> ql/src/java/org/apache/hadoop/hive/ql/plan/VectorSparkHashTableSinkDesc.java 
> 7fb59db 
>   
> ql/src/java/org

Re: Review Request 57586: HIVE-16183: Fix potential thread safety issues with static variables

2017-03-14 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/57586/#review168965
---




metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreSchemaInfo.java
Line 58 (original), 56 (patched)
<https://reviews.apache.org/r/57586/#comment241305>

shall we remove the hiveConf parameter as it's not needed?



ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
Line 502 (original), 480 (patched)
<https://reviews.apache.org/r/57586/#comment241306>

do we still need this method?



ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
Line 545 (original), 520 (patched)
<https://reviews.apache.org/r/57586/#comment241307>

same as above


- Rui Li


On March 14, 2017, 4:32 a.m., Xuefu Zhang wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/57586/
> ---
> 
> (Updated March 14, 2017, 4:32 a.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-16183
> https://issues.apache.org/jira/browse/HIVE-16183
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> Please see JIRA description
> 
> 
> Diffs
> -
> 
>   beeline/src/java/org/apache/hive/beeline/BeeLineOpts.java 7e6846d 
>   beeline/src/java/org/apache/hive/beeline/HiveSchemaHelper.java 181f0d2 
>   cli/src/java/org/apache/hadoop/hive/cli/RCFileCat.java f1806a0 
>   cli/src/test/org/apache/hadoop/hive/cli/TestRCFileCat.java 11ceb31 
>   common/src/java/org/apache/hadoop/hive/common/LogUtils.java c2a0d9a 
>   common/src/java/org/apache/hadoop/hive/common/StatsSetupConst.java 926b4a6 
>   
> metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreSchemaInfo.java 
> 9c30ee7 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/ArchiveUtils.java 6381a21 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 4ac25c2 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 6693134 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizationContext.java 
> 5b0c2bf 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CuckooSetBytes.java
>  6383e8a 
>   
> ql/src/java/org/apache/hadoop/hive/ql/exec/vector/mapjoin/fast/VectorMapJoinFastHashTable.java
>  9030e5f 
>   ql/src/java/org/apache/hadoop/hive/ql/history/HiveHistoryImpl.java 6582cdd 
>   ql/src/java/org/apache/hadoop/hive/ql/index/HiveIndex.java a1408e9 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java 7727114 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java 4995bdf 
>   ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java d391164 
>   ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java 369584b 
>   ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/stats/PartialScanTask.java 
> 90b1dff 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/VirtualColumn.java 044d64c 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 0e67ea6 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/listbucketingpruner/ListBucketingPrunerUtils.java
>  4d3e74e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
>  93202c3 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java 
> 50eda15 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/VectorizerReason.java
>  e0a6198 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java 
> 36009bf 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
> f175663 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/WindowingSpec.java 01b5559 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/AbstractVectorDesc.java e85a418 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/GroupByDesc.java 0b49294 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MapJoinDesc.java ca69697 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/ReduceSinkDesc.java 9ae30ab 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorAppMasterEventDesc.java 
> 2e11321 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorFileSinkDesc.java 325ac91 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorFilterDesc.java 6feed84 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorGroupByDesc.java f8554e2 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorLimitDesc.java c9bc45a 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorMapJoinDesc.java 3aa65d3 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/VectorMapJoinInfo.java 9429785 
>   ql/src/java/or

[jira] [Created] (HIVE-16155) No need for ConditionalTask if no conditional map join is created

2017-03-09 Thread Rui Li (JIRA)
Rui Li created HIVE-16155:
-

 Summary: No need for ConditionalTask if no conditional map join is 
created
 Key: HIVE-16155
 URL: https://issues.apache.org/jira/browse/HIVE-16155
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Review Request 56687: Intern strings in various critical places to reduce memory consumption.

2017-02-27 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/56687/#review166991
---




ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java (line 3178)
<https://reviews.apache.org/r/56687/#comment239095>

do we still need this? I think createEmptyFile will intern the strings for 
us?



ql/src/java/org/apache/hadoop/hive/ql/plan/MapWork.java (line 173)
<https://reviews.apache.org/r/56687/#comment239097>

instead of creating a new map, can we use the pathToAliases map and intern 
the paths in-place?


- Rui Li


On Feb. 27, 2017, 7:42 p.m., Misha Dmitriev wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/56687/
> ---
> 
> (Updated Feb. 27, 2017, 7:42 p.m.)
> 
> 
> Review request for hive, Chaoyu Tang, Mohit Sabharwal, and Sergio Pena.
> 
> 
> Bugs: https://issues.apache.org/jira/browse/HIVE-15882
> 
> https://issues.apache.org/jira/browse/https://issues.apache.org/jira/browse/HIVE-15882
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See the description of the problem in 
> https://issues.apache.org/jira/browse/HIVE-15882 Interning strings per this 
> review removes most of the overhead due to duplicate strings.
> 
> Also, where maps in several places are created from other maps, use the 
> original map's size for the new map. This is to avoid the situation when a 
> map with default capacity (typically 16) is created to hold just 2-3 entries, 
> and the rest of the internal 16-entry array is wasted.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/StringInternUtils.java 
> PRE-CREATION 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 
> e81cbce3e333d44a4088c10491f399e92a505293 
>   ql/src/java/org/apache/hadoop/hive/ql/hooks/Entity.java 
> 08420664d59f28f75872c25c9f8ee42577b23451 
>   ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java 
> e91064b9c75e8adb2b36f21ff19ec0c1539b03b9 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java 
> 51530ac16c92cc75d501bfcb573557754ba0c964 
>   ql/src/java/org/apache/hadoop/hive/ql/io/SymbolicInputFormat.java 
> 55b3b551a1dac92583b6e03b10beb8172ca93d45 
>   ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java 
> 82dc89803be9cf9e0018720eeceb90ff450bfdc8 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java 
> c0edde9e92314d86482b5c46178987e79fae57fe 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java 
> c6ae6f290857cfd10f1023058ede99bf4a10f057 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 
> 24d16812515bdfa90b4be7a295c0388fcdfe95ef 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
>  ede4fcbe342052ad86dadebcc49da2c0f515ea98 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
>  0882ae2c6205b1636cbc92e76ef66bb70faadc76 
>   
> ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java 
> 68b0ad9ea63f051f16fec3652d8525f7ab07eb3f 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MapWork.java 
> d4bdd96eaf8d179bed43b8a8c3be0d338940154a 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MsckDesc.java 
> b7a7e4b7a5f8941b080c7805d224d3885885f444 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java 
> 73981e826870139a42ad881103fdb0a2ef8433a2 
> 
> Diff: https://reviews.apache.org/r/56687/diff/
> 
> 
> Testing
> ---
> 
> I've measured how much memory this change plus another one (interning 
> Properties in PartitionDesc) save in my HS2 benchmark - the result is 37%. 
> See the details in HIVE-15882.
> 
> 
> Thanks,
> 
> Misha Dmitriev
> 
>



Re: Review Request 56687: Intern strings in various critical places to reduce memory consumption.

2017-02-27 Thread Rui Li


> On Feb. 24, 2017, 7:38 a.m., Rui Li wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java,
> >  line 322
> > <https://reviews.apache.org/r/56687/diff/2/?file=1643011#file1643011line322>
> >
> > will this cause the hash map to resize since the default load factor is 
> > 0.75? and several similar concerns below
> 
> Misha Dmitriev wrote:
> You are probably right, in that this constructor's parameter is the 
> initial capacity of this table (more or less the size of the internal array) 
> - not how many elements the table is expected to hold. However, if you check 
> the code of HashMap, the things are more interesting. The actual capacity of 
> the table is always a power of two, so unless this parameter is also a power 
> of two, the capacity will be chosen as the nearest higher power of two, i.e. 
> it will be higher than the parameter and closer to what we actually need. 
> Also, if we create a table with the default size (16) here and then will put 
> many more elements into it, it will be resized several times, whereas with 
> the current code it will be resized at most once. Trying to "factor in" the 
> load factor will likely add more confusion/complexity. All in all, given that 
> choosing capacity in HashMap internally is non-trivial, I think it's 
> easier/safer to just call 'new HashMap(oldMap.size())' as we do now.

Then could you explain why we need to change the current code? The JavaDoc of 
LinkedHashMap(Map m) indicates it will create an 
instance "with a default load factor (0.75) and an initial capacity sufficient 
to hold the mappings in the specified map". Looking at the code, it computes 
the initial cap like "m.size()/loadFactor + 1", rounds it to next power of two, 
and it avoids re-hashing. Won't that be good enough for us?


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/56687/#review166649
---


On Feb. 24, 2017, 9:27 p.m., Misha Dmitriev wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/56687/
> ---
> 
> (Updated Feb. 24, 2017, 9:27 p.m.)
> 
> 
> Review request for hive, Chaoyu Tang, Mohit Sabharwal, and Sergio Pena.
> 
> 
> Bugs: https://issues.apache.org/jira/browse/HIVE-15882
> 
> https://issues.apache.org/jira/browse/https://issues.apache.org/jira/browse/HIVE-15882
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See the description of the problem in 
> https://issues.apache.org/jira/browse/HIVE-15882 Interning strings per this 
> review removes most of the overhead due to duplicate strings.
> 
> Also, where maps in several places are created from other maps, use the 
> original map's size for the new map. This is to avoid the situation when a 
> map with default capacity (typically 16) is created to hold just 2-3 entries, 
> and the rest of the internal 16-entry array is wasted.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/StringInternUtils.java 
> PRE-CREATION 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 
> e81cbce3e333d44a4088c10491f399e92a505293 
>   ql/src/java/org/apache/hadoop/hive/ql/hooks/Entity.java 
> 08420664d59f28f75872c25c9f8ee42577b23451 
>   ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java 
> e91064b9c75e8adb2b36f21ff19ec0c1539b03b9 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java 
> 51530ac16c92cc75d501bfcb573557754ba0c964 
>   ql/src/java/org/apache/hadoop/hive/ql/io/SymbolicInputFormat.java 
> 55b3b551a1dac92583b6e03b10beb8172ca93d45 
>   ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java 
> 82dc89803be9cf9e0018720eeceb90ff450bfdc8 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java 
> c0edde9e92314d86482b5c46178987e79fae57fe 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java 
> c6ae6f290857cfd10f1023058ede99bf4a10f057 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 
> 24d16812515bdfa90b4be7a295c0388fcdfe95ef 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
>  ede4fcbe342052ad86dadebcc49da2c0f515ea98 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
>  0882ae2c6205b1636cbc92e76ef66bb70faadc76 
>   
> ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java 
> 68b

[jira] [Created] (HIVE-16047) Shouldn't try to get KeyProvider unless encryption is enabled

2017-02-27 Thread Rui Li (JIRA)
Rui Li created HIVE-16047:
-

 Summary: Shouldn't try to get KeyProvider unless encryption is 
enabled
 Key: HIVE-16047
 URL: https://issues.apache.org/jira/browse/HIVE-16047
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


Found lots of following errors in HS2 log:
{noformat}
hdfs.KeyProviderCache: Could not find uri with key 
[dfs.encryption.key.provider.uri] to create a keyProvider !!
{noformat}

Similar to HDFS-7931



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: Review Request 56687: Intern strings in various critical places to reduce memory consumption.

2017-02-23 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/56687/#review166649
---




common/src/java/org/apache/hadoop/hive/common/StringInternUtils.java (line 65)
<https://reviews.apache.org/r/56687/#comment238671>

do we need to check whether uri is null?



common/src/java/org/apache/hadoop/hive/common/StringInternUtils.java (line 67)
<https://reviews.apache.org/r/56687/#comment238672>

why the stringField doesn't need the null check like other fields?



ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java (line 3147)
<https://reviews.apache.org/r/56687/#comment238673>

How about intern the path in the createEmptyFile method?



ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java (line 183)
<https://reviews.apache.org/r/56687/#comment238676>

can we call the util method?



ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java (line 188)
<https://reviews.apache.org/r/56687/#comment238677>

guess we can also add a util method for this



ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java (line 253)
<https://reviews.apache.org/r/56687/#comment238681>

since we'll intern strings in the new path, do we have to intern taskTmpDir 
here?



ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java 
(line 322)
<https://reviews.apache.org/r/56687/#comment238682>

will this cause the hash map to resize since the default load factor is 
0.75? and several similar concerns below


- Rui Li


On Feb. 23, 2017, 9:01 p.m., Misha Dmitriev wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/56687/
> ---
> 
> (Updated Feb. 23, 2017, 9:01 p.m.)
> 
> 
> Review request for hive, Chaoyu Tang, Mohit Sabharwal, and Sergio Pena.
> 
> 
> Bugs: https://issues.apache.org/jira/browse/HIVE-15882
> 
> https://issues.apache.org/jira/browse/https://issues.apache.org/jira/browse/HIVE-15882
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See the description of the problem in 
> https://issues.apache.org/jira/browse/HIVE-15882 Interning strings per this 
> review removes most of the overhead due to duplicate strings.
> 
> Also, where maps in several places are created from other maps, use the 
> original map's size for the new map. This is to avoid the situation when a 
> map with default capacity (typically 16) is created to hold just 2-3 entries, 
> and the rest of the internal 16-entry array is wasted.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/StringInternUtils.java 
> PRE-CREATION 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 
> e81cbce3e333d44a4088c10491f399e92a505293 
>   ql/src/java/org/apache/hadoop/hive/ql/hooks/Entity.java 
> 08420664d59f28f75872c25c9f8ee42577b23451 
>   ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java 
> e91064b9c75e8adb2b36f21ff19ec0c1539b03b9 
>   ql/src/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java 
> 51530ac16c92cc75d501bfcb573557754ba0c964 
>   ql/src/java/org/apache/hadoop/hive/ql/io/SymbolicInputFormat.java 
> 55b3b551a1dac92583b6e03b10beb8172ca93d45 
>   ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java 
> 82dc89803be9cf9e0018720eeceb90ff450bfdc8 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java 
> c0edde9e92314d86482b5c46178987e79fae57fe 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java 
> c6ae6f290857cfd10f1023058ede99bf4a10f057 
>   ql/src/java/org/apache/hadoop/hive/ql/optimizer/GenMapRedUtils.java 
> 24d16812515bdfa90b4be7a295c0388fcdfe95ef 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java
>  ede4fcbe342052ad86dadebcc49da2c0f515ea98 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
>  0882ae2c6205b1636cbc92e76ef66bb70faadc76 
>   
> ql/src/java/org/apache/hadoop/hive/ql/plan/ConditionalResolverMergeFiles.java 
> 68b0ad9ea63f051f16fec3652d8525f7ab07eb3f 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MapWork.java 
> d4bdd96eaf8d179bed43b8a8c3be0d338940154a 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/MsckDesc.java 
> b7a7e4b7a5f8941b080c7805d224d3885885f444 
>   ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java 
> 73981e826870139a42ad881103fdb0a2ef8433a2 
> 
> Diff: https://reviews.apache.org/r/56687/diff/
> 
> 
> Testing
> ---
> 
> I've measured how much memory this change plus another one (interning 
> Properties in PartitionDesc) save in my HS2 benchmark - the result is 37%. 
> See the details in HIVE-15882.
> 
> 
> Thanks,
> 
> Misha Dmitriev
> 
>



Re: [ANNOUNCE] New committer: Zoltan Haindrich

2017-02-21 Thread Rui Li
Congratulations!

On Wed, Feb 22, 2017 at 8:47 AM, Sergey Shelukhin <ser...@hortonworks.com>
wrote:

> Congratulations!
>
> On 17/2/21, 16:43, "Prasanth Jayachandran" <pjayachand...@hortonworks.com>
> wrote:
>
> >Congratulations Zoltan!!
> >
> >Thanks
> >Prasanth
> >
> >
> >
> >
> >On Tue, Feb 21, 2017 at 4:35 PM -0800, "Eugene Koifman"
> ><ekoif...@hortonworks.com<mailto:ekoif...@hortonworks.com>> wrote:
> >
> >
> >Congratulations!
> >
> >On 2/21/17, 4:17 PM, "Vihang Karajgaonkar"  wrote:
> >
> >Congrats Zoltan!
> >
> >On Tue, Feb 21, 2017 at 4:16 PM, Vaibhav Gumashta  wrote:
> >
> >> Congrats Zoltan!
> >>
> >> On 2/21/17, 4:16 PM, "Jimmy Xiang"  wrote:
> >>
> >> >Congrats, Zoltan!!
> >> >
> >> >On Tue, Feb 21, 2017 at 4:15 PM, Sushanth Sowmyan
> >> >wrote:
> >> >> Congrats, Zoltan!
> >> >>
> >> >> Welcome aboard. :)
> >> >>
> >> >> On Feb 21, 2017 15:42, "Rajesh Balamohan"
> >> wrote:
> >> >>
> >> >>> Congrats Zoltan. :)
> >> >>>
> >> >>> ~Rajesh.B
> >> >>>
> >> >>> On Wed, Feb 22, 2017 at 4:43 AM, Wei Zheng
> >> >>>wrote:
> >> >>>
> >> >>> > Congrats Zoltan!
> >> >>> >
> >> >>> > Thanks,
> >> >>> > Wei
> >> >>> >
> >> >>> > On 2/21/17, 13:09, "Alan Gates"  wrote:
> >> >>> >
> >> >>> > On behalf of the Hive PMC I am happy to announce Zoltan
> >> >>>Haindrich is
> >> >>> > our newest committer.  He has been contributing to Hive for
> >several
> >> >>> months
> >> >>> > across a number of areas, including the parser, HiveServer2,
> >and
> >> >>>cleaning
> >> >>> > up unit tests and documentation.  Please join me in welcoming
> >Zoltan
> >> >>>to
> >> >>> > Hive.
> >> >>> >
> >> >>> > Zoltan, feel free to say a few words introducing yourself
> >if you
> >> >>> would
> >> >>> > like to.
> >> >>> >
> >> >>> > Alan.
> >> >>> >
> >> >>> >
> >> >>> >
> >> >>>
> >> >
> >>
> >>
> >
> >
> >
>
>


-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


[jira] [Created] (HIVE-15860) RemoteSparkJobMonitor may hang when RemoteDriver exists abnormally

2017-02-09 Thread Rui Li (JIRA)
Rui Li created HIVE-15860:
-

 Summary: RemoteSparkJobMonitor may hang when RemoteDriver exists 
abnormally
 Key: HIVE-15860
 URL: https://issues.apache.org/jira/browse/HIVE-15860
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (HIVE-15526) Some tests need SORT_QUERY_RESULTS

2016-12-29 Thread Rui Li (JIRA)
Rui Li created HIVE-15526:
-

 Summary: Some tests need SORT_QUERY_RESULTS
 Key: HIVE-15526
 URL: https://issues.apache.org/jira/browse/HIVE-15526
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Invitation for Hive committers to become ORC committers

2016-12-15 Thread Rui Li
I'm interested. Thanks!

On Fri, Dec 16, 2016 at 1:18 PM, Chinna Rao Lalam <
lalamchinnara...@gmail.com> wrote:

> I would be interested. Thanks.
>
> Chinna Rao Lalam
>
> On Fri, Dec 16, 2016 at 6:43 AM, Owen O'Malley <omal...@apache.org> wrote:
>
> > Ok, I've added the people who have responded so far and updated the ORC
> > website.
> >
> > http://orc.apache.org/news/2016/12/15/new-committers/
> > http://orc.apache.org/develop/
> >
> > Please make sure that I didn't typo your names.
> >
> > .. Owen
> >
> > On Thu, Dec 15, 2016 at 4:44 PM, Chaoyu Tang <ctang...@gmail.com> wrote:
> >
> > > I am interested in. Thanks
> > >
> > > Chaoyu
> > >
> > > On Thu, Dec 15, 2016 at 5:13 PM, Rajesh Balamohan <
> rbalamo...@apache.org
> > >
> > > wrote:
> > >
> > > > I would be interested. Thanks.
> > > >
> > > > ~Rajesh.B
> > > >
> > > > On Fri, Dec 16, 2016 at 3:31 AM, Mithun Radhakrishnan <
> > > > mithun.radhakrish...@yahoo.com.invalid> wrote:
> > > >
> > > > > I'd be keen.
> > > > > Thanks,Mithun
> > > > > On Thursday, December 15, 2016, 1:37:36 PM PST, Wei Zheng <
> > > > > wzh...@hortonworks.com> wrote:I’m interested. Thanks.
> > > > >
> > > > > Thanks,
> > > > > Wei
> > > > >
> > > > > On 12/15/16, 13:21, "Vaibhav Gumashta" <vgumas...@hortonworks.com>
> > > > wrote:
> > > > >
> > > > > I¹d be interested.
> > > > >
> > > > > Thanks,
> > > > > ‹Vaibhav
> > > > >
> > > > > On 12/15/16, 1:12 PM, "Owen O'Malley" <omal...@apache.org>
> > wrote:
> > > > >
> > > > > >All,
> > > > > >  As you are aware, we are in the last stages of removing the
> > > forked
> > > > > ORC
> > > > > >code out of Hive. The goal of moving ORC out of Hive was to
> > > increase
> > > > > its
> > > > > >community and we want to be very deliberately inclusive of the
> > > Hive
> > > > > >development community. Towards that end, the ORC PMC wants to
> > > > welcome
> > > > > >anyone who is already a Hive committer to become a committer
> on
> > > ORC.
> > > > > >
> > > > > >  Please respond on this thread to let us know if you are
> > > > interested.
> > > > > >
> > > > > >Thanks,
> > > > > >  Owen on behalf of the ORC PMC
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > >
> > >
> >
>
>
>
> --
> Hope It Helps,
> Chinna
>



-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Re: [ANNOUNCE] New Hive Committer - Rajesh Balamohan

2016-12-14 Thread Rui Li
Congratulations :)

On Thu, Dec 15, 2016 at 6:50 AM, Gunther Hagleitner <
ghagleit...@hortonworks.com> wrote:

> Congrats Rajesh!
> 
> From: Jimmy Xiang <jxc...@gmail.com>
> Sent: Wednesday, December 14, 2016 11:38 AM
> To: u...@hive.apache.org
> Cc: dev@hive.apache.org; rbalamo...@apache.org
> Subject: Re: [ANNOUNCE] New Hive Committer - Rajesh Balamohan
>
> Congrats, Rajesh!!
>
> On Wed, Dec 14, 2016 at 11:32 AM, Sergey Shelukhin
> <ser...@hortonworks.com> wrote:
> > Congratulations!
> >
> > From: Chao Sun <sunc...@apache.org>
> > Reply-To: "u...@hive.apache.org" <u...@hive.apache.org>
> > Date: Wednesday, December 14, 2016 at 10:52
> > To: "dev@hive.apache.org" <dev@hive.apache.org>
> > Cc: "u...@hive.apache.org" <u...@hive.apache.org>, "
> rbalamo...@apache.org"
> > <rbalamo...@apache.org>
> > Subject: Re: [ANNOUNCE] New Hive Committer - Rajesh Balamohan
> >
> > Congrats Rajesh!
> >
> > On Wed, Dec 14, 2016 at 9:26 AM, Vihang Karajgaonkar <
> vih...@cloudera.com>
> > wrote:
> >>
> >> Congrats Rajesh!
> >>
> >> On Wed, Dec 14, 2016 at 1:54 AM, Jesus Camacho Rodriguez <
> >> jcamachorodrig...@hortonworks.com> wrote:
> >>
> >> > Congrats Rajesh, well deserved! :)
> >> >
> >> > --
> >> > Jesús
> >> >
> >> >
> >> >
> >> >
> >> > On 12/14/16, 8:41 AM, "Lefty Leverenz" <leftylever...@gmail.com>
> wrote:
> >> >
> >> > >Congratulations Rajesh!
> >> > >
> >> > >-- Lefty
> >> > >
> >> > >
> >> > >On Tue, Dec 13, 2016 at 11:58 PM, Rajesh Balamohan
> >> > > <rbalamo...@apache.org
> >> > >
> >> > >wrote:
> >> > >
> >> > >> Thanks a lot for providing this opportunity and to all for their
> >> > messages.
> >> > >> :)
> >> > >>
> >> > >> ~Rajesh.B
> >> > >>
> >> > >> On Wed, Dec 14, 2016 at 11:33 AM, Dharmesh Kakadia
> >> > >> <dhkaka...@gmail.com
> >> > >
> >> > >> wrote:
> >> > >>
> >> > >> > Congrats Rajesh !
> >> > >> >
> >> > >> > Thanks,
> >> > >> > Dharmesh
> >> > >> >
> >> > >> > On Tue, Dec 13, 2016 at 7:37 PM, Vikram Dixit K <
> >> > vikram.di...@gmail.com>
> >> > >> > wrote:
> >> > >> >
> >> > >> >> Congrats Rajesh! :)
> >> > >> >>
> >> > >> >> On Tue, Dec 13, 2016 at 9:36 PM, Pengcheng Xiong
> >> > >> >> <pxi...@apache.org>
> >> > >> >> wrote:
> >> > >> >>
> >> > >> >>> Congrats Rajesh! :)
> >> > >> >>>
> >> > >> >>> On Tue, Dec 13, 2016 at 6:51 PM, Prasanth Jayachandran <
> >> > >> >>> prasan...@apache.org
> >> > >> >>> > wrote:
> >> > >> >>>
> >> > >> >>> > The Apache Hive PMC has voted to make Rajesh Balamohan a
> >> > committer on
> >> > >> >>> the
> >> > >> >>> > Apache Hive Project. Please join me in congratulating Rajesh.
> >> > >> >>> >
> >> > >> >>> > Congratulations Rajesh!
> >> > >> >>> >
> >> > >> >>> > Thanks
> >> > >> >>> > Prasanth
> >> > >> >>>
> >> > >> >>
> >> > >> >>
> >> > >> >>
> >> > >> >> --
> >> > >> >> Nothing better than when appreciated for hard work.
> >> > >> >> -Mark
> >> > >> >>
> >> > >> >
> >> > >> >
> >> > >>
> >> >
> >
> >
>
>


-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


[jira] [Created] (HIVE-15428) HoS DPP doesn't remove cyclic dependency

2016-12-13 Thread Rui Li (JIRA)
Rui Li created HIVE-15428:
-

 Summary: HoS DPP doesn't remove cyclic dependency
 Key: HIVE-15428
 URL: https://issues.apache.org/jira/browse/HIVE-15428
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15357) Fix and re-enable the spark-only tests

2016-12-05 Thread Rui Li (JIRA)
Rui Li created HIVE-15357:
-

 Summary: Fix and re-enable the spark-only tests
 Key: HIVE-15357
 URL: https://issues.apache.org/jira/browse/HIVE-15357
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15302) Relax the requirement that HoS needs Spark built w/o Hive

2016-11-28 Thread Rui Li (JIRA)
Rui Li created HIVE-15302:
-

 Summary: Relax the requirement that HoS needs Spark built w/o Hive
 Key: HIVE-15302
 URL: https://issues.apache.org/jira/browse/HIVE-15302
 Project: Hive
  Issue Type: Improvement
Reporter: Rui Li
Assignee: Rui Li


This requirement becomes more and more unacceptable as SparkSQL becomes widely 
adopted. Let's use this JIRA to find out how we can relax the limitation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15299) Yarn-cluster and yarn-client deprecated in Spark 2.0

2016-11-28 Thread Rui Li (JIRA)
Rui Li created HIVE-15299:
-

 Summary: Yarn-cluster and yarn-client deprecated in Spark 2.0
 Key: HIVE-15299
 URL: https://issues.apache.org/jira/browse/HIVE-15299
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


Need to use master "yarn" with specified deploy mode instead.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15202) Concurrent compactions for the same partition may generate malformed folder structure

2016-11-14 Thread Rui Li (JIRA)
Rui Li created HIVE-15202:
-

 Summary: Concurrent compactions for the same partition may 
generate malformed folder structure
 Key: HIVE-15202
 URL: https://issues.apache.org/jira/browse/HIVE-15202
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li


If two compactions run concurrently on a single partition, it may generate 
folder structure like this: (nested base dir)
{noformat}
drwxr-xr-x   - root supergroup  0 2016-11-14 22:23 
/user/hive/warehouse/test/z=1/base_007/base_007
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_0
-rw-r--r--   3 root supergroup611 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_1
-rw-r--r--   3 root supergroup614 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_2
-rw-r--r--   3 root supergroup621 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_3
-rw-r--r--   3 root supergroup621 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_4
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_5
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_6
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_7
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_8
-rw-r--r--   3 root supergroup201 2016-11-14 21:46 
/user/hive/warehouse/test/z=1/base_007/bucket_9
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15139) HoS local mode fails with NumberFormatException

2016-11-07 Thread Rui Li (JIRA)
Rui Li created HIVE-15139:
-

 Summary: HoS local mode fails with NumberFormatException
 Key: HIVE-15139
 URL: https://issues.apache.org/jira/browse/HIVE-15139
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


It's because we store {{stageId_attemptNum}} in JobMetricsListener but expect 
only {{stageId}} in LocalSparkJobStatus.
{noformat}
java.lang.NumberFormatException: For input string: "0_0"
at 
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Integer.parseInt(Integer.java:580)
at java.lang.Integer.parseInt(Integer.java:615)
at 
org.apache.hadoop.hive.ql.exec.spark.status.impl.LocalSparkJobStatus.getSparkStatistics(LocalSparkJobStatus.java:146)
at 
org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:104)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199)

{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15081) RetryingMetaStoreClient.getProxy(HiveConf, Boolean) doesn't match constructor of HiveMetaStoreClient

2016-10-26 Thread Rui Li (JIRA)
Rui Li created HIVE-15081:
-

 Summary: RetryingMetaStoreClient.getProxy(HiveConf, Boolean) 
doesn't match constructor of HiveMetaStoreClient
 Key: HIVE-15081
 URL: https://issues.apache.org/jira/browse/HIVE-15081
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


Calling RetryingMetaStoreClient.getProxy(HiveConf, Boolean) will result in error
{noformat}
Exception in thread "main" java.lang.RuntimeException: Unable to instantiate 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient
at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1661)
at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:81)
at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:131)
at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:87)
Caused by: java.lang.NoSuchMethodException: 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(org.apache.hadoop.hive.conf.HiveConf,
 java.lang.Boolean)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-15039) A better job monitor console output for HoS

2016-10-23 Thread Rui Li (JIRA)
Rui Li created HIVE-15039:
-

 Summary: A better job monitor console output for HoS
 Key: HIVE-15039
 URL: https://issues.apache.org/jira/browse/HIVE-15039
 Project: Hive
  Issue Type: Improvement
Reporter: Rui Li
Assignee: Rui Li


When there're many stages, it's very difficult to read the console output of 
job progress of HoS. Attached screenshot is an example.
We may learn from HoT as it does a much better than HoS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 50787: Add a timezone-aware timestamp

2016-09-23 Thread Rui Li


> On Sept. 22, 2016, 11:20 a.m., Jason Dere wrote:
> > - How about compatbility with the various date functions 
> > (year()/month()/day()/etc)?

For most of the functions, TIMESTAMPTZ is implicitly converted to text. 
Therefore I think we can get correct results. I added some special handle in 
HOUR because some hour may be unavailable due to DST.
So far I've verified the following funcsions work:

to_date
year
quarter
month
day
dayofmonth
hour
minute
second
weekofyear

Is it OK we leave others in follow-on tasks? I'd like to keep the patch small.


> On Sept. 22, 2016, 11:20 a.m., Jason Dere wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToTimestampTZ.java,
> >  line 58
> > <https://reviews.apache.org/r/50787/diff/4/?file=1507773#file1507773line58>
> >
> > No conversions to/from DATE/TIMESTAMP?

Added conversion from date/timestamp to timestamptz. Default timezone is used 
for the converted timestamptz.
We can add convertion from numeric types in follow-on task.


> On Sept. 22, 2016, 11:20 a.m., Jason Dere wrote:
> > serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java,
> >  line 1109
> > <https://reviews.apache.org/r/50787/diff/4/?file=1507829#file1507829line1109>
> >
> > If the local timezone is different from the timezone in the 
> > TimestampTZ, is it possible that the year/month/day of the DATE might be 
> > different from the year/month/day of the TimestampTZ?

Good catch! It makes more sense to convert from the text representation than 
the time/nanos. So I convert the timestamptz to string first, and use that 
string to create the date. Same applies when converting to timestamp.


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/50787/#review149983
---


On Sept. 22, 2016, 4:05 a.m., Rui Li wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/50787/
> ---
> 
> (Updated Sept. 22, 2016, 4:05 a.m.)
> 
> 
> Review request for hive.
> 
> 
> Bugs: HIVE-14412
> https://issues.apache.org/jira/browse/HIVE-14412
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> The 1st patch to add timezone-aware timestamp.
> 
> 
> Diffs
> -
> 
>   common/src/test/org/apache/hadoop/hive/common/type/TestTimestampTZ.java 
> PRE-CREATION 
>   contrib/src/test/queries/clientnegative/serde_regex.q a676338 
>   contrib/src/test/queries/clientpositive/serde_regex.q d75d607 
>   contrib/src/test/results/clientnegative/serde_regex.q.out 0f9b036 
>   contrib/src/test/results/clientpositive/serde_regex.q.out 2984293 
>   hbase-handler/src/test/queries/positive/hbase_timestamp.q 0350afe 
>   hbase-handler/src/test/results/positive/hbase_timestamp.q.out 3918121 
>   jdbc/src/java/org/apache/hive/jdbc/HiveBaseResultSet.java 93f093f 
>   jdbc/src/java/org/apache/hive/jdbc/JdbcColumn.java 38918f0 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java de74c3e 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/GroupByOperator.java f28d33e 
>   ql/src/java/org/apache/hadoop/hive/ql/exec/SerializationUtilities.java 
> 7be628e 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/TypeConverter.java
>  ba41518 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java 
> 8b0db4a 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g 7ceb005 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g 62bbcc6 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/IdentifiersParser.g 9ba1865 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/TypeCheckProcFactory.java 
> 82080eb 
>   ql/src/java/org/apache/hadoop/hive/ql/stats/StatsUtils.java a718264 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToBoolean.java 17b892c 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToByte.java efae82d 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToDouble.java 9cbc114 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToFloat.java 5808c90 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToInteger.java a7551cb 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToLong.java c961d14 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToShort.java 570408a 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/UDFToString.java 5cacd59 
>   ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDF.java 259fde8 
>   
> ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFToTimestampT

Re: Review Request 50787: Add a timezone-aware timestamp

2016-09-23 Thread Rui Li
/src/gen/thrift/gen-rb/serde_constants.rb 0ce9f27 
  serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java 7ffc964 
  
serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/BinarySortableSerDe.java
 5e119d7 
  
serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/fast/BinarySortableDeserializeRead.java
 a7785b2 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritable.java 
bbccc7f 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritableBase.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java 56945d1 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
54bfd2d 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
f8a110d 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 1ac72c6 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampObjectInspector.java
 509189e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 51b529e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectorInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampWritable.java 
3c483cc 
  service-rpc/if/TCLIService.thrift a4fa7b0 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h b249544 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp 2f460e8 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 bce2a0c 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php 786c773 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py fdf6b1f 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 4b1854c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 
  storage-api/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
PRE-CREATION 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java 
4a745e4 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/TimestampUtils.java 
41db9ca 

Diff: https://reviews.apache.org/r/50787/diff/


Testing
---


Thanks,

Rui Li



Re: Review Request 50787: Add a timezone-aware timestamp

2016-09-21 Thread Rui Li
/fast/BinarySortableDeserializeRead.java
 a7785b2 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritable.java 
bbccc7f 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritableBase.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java 56945d1 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
54bfd2d 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
f8a110d 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 1ac72c6 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampObjectInspector.java
 509189e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 51b529e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectorInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampWritable.java 
3c483cc 
  service-rpc/if/TCLIService.thrift a4fa7b0 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h b249544 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp 2f460e8 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 bce2a0c 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php 786c773 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py fdf6b1f 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 4b1854c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 
  storage-api/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
PRE-CREATION 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java 
4a745e4 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/TimestampUtils.java 
41db9ca 

Diff: https://reviews.apache.org/r/50787/diff/


Testing
---


Thanks,

Rui Li



Re: Review Request 50787: Add a timezone-aware timestamp

2016-09-21 Thread Rui Li
/fast/BinarySortableDeserializeRead.java
 a7785b2 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampTZWritable.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritable.java 
bbccc7f 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritableBase.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyFactory.java 23dbe6a 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestamp.java 56945d1 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyTimestampTZ.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 73c72e1 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyPrimitiveObjectInspectorFactory.java
 5601734 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazy/objectinspector/primitive/LazyTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryFactory.java 
52f3527 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinarySerDe.java 
54bfd2d 
  
serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryTimestampTZ.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/lazybinary/LazyBinaryUtils.java 
f8a110d 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 1ac72c6 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampObjectInspector.java
 509189e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 51b529e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/TimestampTZObjectorInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantTimestampTZObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableTimestampTZObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampWritable.java 
3c483cc 
  service-rpc/if/TCLIService.thrift a4fa7b0 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h b249544 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp 2f460e8 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 bce2a0c 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php 786c773 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py fdf6b1f 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 4b1854c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 
  storage-api/src/java/org/apache/hadoop/hive/common/type/TimestampTZ.java 
PRE-CREATION 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java 
4a745e4 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/TimestampUtils.java 
41db9ca 

Diff: https://reviews.apache.org/r/50787/diff/


Testing
---


Thanks,

Rui Li



Re: Review Request 50787: Add a timezone-aware timestamp

2016-09-11 Thread Rui Li
/hive/serde2/lazybinary/LazyBinaryUtils.java 
f8a110d 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorConverters.java
 24b3d4e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java
 1ac72c6 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/PrimitiveObjectInspector.java
 70633f3 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/HiveTimestampObjectorInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaHiveTimestampObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/JavaTimestampObjectInspector.java
 509189e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorConverter.java
 e08ad43 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorFactory.java
 2ed0843 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/PrimitiveObjectInspectorUtils.java
 51b529e 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/SettableHiveTimestampObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableConstantHiveTimestampObjectInspector.java
 PRE-CREATION 
  
serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/primitive/WritableHiveTimestampObjectInspector.java
 PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/thrift/Type.java 0ad8c02 
  serde/src/java/org/apache/hadoop/hive/serde2/typeinfo/TypeInfoFactory.java 
43c4819 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampWritable.java 
3c483cc 
  service-rpc/if/TCLIService.thrift a4fa7b0 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_constants.cpp 991cb2e 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.h b249544 
  service-rpc/src/gen/thrift/gen-cpp/TCLIService_types.cpp 2f460e8 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TCLIServiceConstants.java
 930bed7 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TProtocolVersion.java
 bce2a0c 
  
service-rpc/src/gen/thrift/gen-javabean/org/apache/hive/service/rpc/thrift/TTypeId.java
 a3735eb 
  service-rpc/src/gen/thrift/gen-php/Types.php 786c773 
  service-rpc/src/gen/thrift/gen-py/TCLIService/constants.py c8d4f8f 
  service-rpc/src/gen/thrift/gen-py/TCLIService/ttypes.py fdf6b1f 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_constants.rb 25adbb4 
  service-rpc/src/gen/thrift/gen-rb/t_c_l_i_service_types.rb 4b1854c 
  service/src/java/org/apache/hive/service/cli/ColumnValue.java 76e8c03 
  service/src/java/org/apache/hive/service/cli/TypeDescriptor.java d634bef 
  storage-api/src/java/org/apache/hadoop/hive/common/type/HiveTimestamp.java 
PRE-CREATION 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java 
4a745e4 
  storage-api/src/java/org/apache/hadoop/hive/ql/util/TimestampUtils.java 
41db9ca 

Diff: https://reviews.apache.org/r/50787/diff/


Testing
---


Thanks,

Rui Li



[jira] [Created] (HIVE-14728) Redundant orig files

2016-09-09 Thread Rui Li (JIRA)
Rui Li created HIVE-14728:
-

 Summary: Redundant orig files
 Key: HIVE-14728
 URL: https://issues.apache.org/jira/browse/HIVE-14728
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Priority: Minor


I find some orig files in master, e.g. SemanticAnalyzer.java.orig. Wondering if 
they are added by mistake?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-14719) ASTNode rootNode is not maintained properly when changing child/parent relation

2016-09-08 Thread Rui Li (JIRA)
Rui Li created HIVE-14719:
-

 Summary: ASTNode rootNode is not maintained properly when changing 
child/parent relation
 Key: HIVE-14719
 URL: https://issues.apache.org/jira/browse/HIVE-14719
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


When I run some query like:
{code}
set hive.cbo.enable=false;
select * from A where exists (select * from B where B.k1=A.k1 and B.k2=A.k2);
{code}
It gets error like:
{noformat}
FAILED: SemanticException Line 0:-1 Invalid table alias or column reference 
'sq_1': (possible column names are: _table_or_col b) k2) sq_corr_1)) (tok, (. 
(tok_table_or_col sq_1) sq_corr_1))
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-14595) TimestampWritable::setTimestamp gives wrong result when 2nd VInt exists

2016-08-21 Thread Rui Li (JIRA)
Rui Li created HIVE-14595:
-

 Summary: TimestampWritable::setTimestamp gives wrong result when 
2nd VInt exists
 Key: HIVE-14595
 URL: https://issues.apache.org/jira/browse/HIVE-14595
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: YourKit open source license

2016-08-16 Thread Rui Li
Our wiki doesn't mention it's only for committers. Anyway I suggest you
contact YourKit sales to figure out.

On Tue, Aug 16, 2016 at 8:38 PM, calvin hung <calvinh...@wasaitech.com>
wrote:

>
>
> Thanks for your response, Rui.
>
> I don't have an apache email account.
>
> It looks like only committer can get an email account according to this
> page http://www.apache.org/dev/committers.html
>
> Does it mean that only Hive committers can get YourKit free licenses for
> Hive performance profiling?
>
>
>
>
>
>  On Tue, 16 Aug 2016 13:33:34 +0800 Rui Li lirui.fu...@gmail.com
> wrote 
>
>
>
>
> If I remember correctly, I just contacted the sales of Yourkit and they
>
> sent me the license by email. You'd better send your email using your
>
> apache email account, in order to convince them you're a developer of Hive.
>
>
>
> On Tue, Aug 16, 2016 at 2:51 AM, calvin hung calvinh...@wasaitech.com&
> gt;
>
> wrote:
>
>
>
>  Hi Rui and Alan,
>
> 
>
>  Could you or any nice guy share more detail steps of getting a Yourkit
>
>  license for Hive?
>
>  I've searched the full Hive dev mail archive but got no exact steps
> to get
>
>  one.
>
>  Thanks!
>
> 
>
>  Calvin
>
>  From: "Li, Rui"rui...@intel.com
>
>  Date: Tue, 31 Mar 2015 01:22:51 +
>
>  To: "dev@hive.apache.org"dev@hive.apache.org
>
> 
>
>  - Contents -
>
> 
>
>  Thanks Alan! But I don’t see Hive in the sponsored open source project
>
>  list. I’ll contact them anyway.
>
> 
>
> 
>
> 
>
>  Cheers,
>
> 
>
>  Rui Li
>
> 
>
> 
>
> 
>
>  From: Alan Gates [mailto:alanfga...@gmail.com]
>
>  Sent: Tuesday, March 31, 2015 1:02 AM
>
>  To: dev@hive.apache.org
>
>  Subject: Re: YourKit open source license
>
> 
>
> 
>
> 
>
>  Seehttps://www.yourkit.com/customers/.
>
> 
>
>  Alan.
>
> 
>
> 
>
> 
>
> 
>
> 
>
>  Li, Rui
>
> 
>
>  March 30, 2015 at 0:54
>
> 
>
>  Hi guys,
>
> 
>
>  I want to use YourKit to profile hive performance. According to the
> wiki
>
>  https://cwiki.apache.org/confluence/display/Hive/Performance;
> hive has
>
>  been granted open source license. Could anybody tell me how I can get
> the
>
>  license? Thanks!
>
> 
>
>  Cheers,
>
>  Rui Li
>
>
>
>
>
>
>
>
>
> --
>
> Best regards!
>
> Rui Li
>
> Cell: (+86) 13564950210
>
>
>
>
>
>
>


-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Re: YourKit open source license

2016-08-15 Thread Rui Li
If I remember correctly, I just contacted the sales of Yourkit and they
sent me the license by email. You'd better send your email using your
apache email account, in order to convince them you're a developer of Hive.

On Tue, Aug 16, 2016 at 2:51 AM, calvin hung <calvinh...@wasaitech.com>
wrote:

> Hi Rui and Alan,
>
> Could you or any nice guy share more detail steps of getting a Yourkit
> license for Hive?
> I've searched the full Hive dev mail archive but got no exact steps to get
> one.
> Thanks!
>
> Calvin
> From: "Li, Rui"<rui...@intel.com>
> Date: Tue, 31 Mar 2015 01:22:51 +
> To: "dev@hive.apache.org"<dev@hive.apache.org>
>
> - Contents -
>
> Thanks Alan! But I don’t see Hive in the sponsored open source project
> list. I’ll contact them anyway.
>
>
>
> Cheers,
>
> Rui Li
>
>
>
> From: Alan Gates [mailto:alanfga...@gmail.com]
> Sent: Tuesday, March 31, 2015 1:02 AM
> To: dev@hive.apache.org
> Subject: Re: YourKit open source license
>
>
>
> Seehttps://www.yourkit.com/customers/.
>
> Alan.
>
>
>
>
>
> Li, Rui
>
> March 30, 2015 at 0:54
>
> Hi guys,
>
> I want to use YourKit to profile hive performance. According to the wiki<
> https://cwiki.apache.org/confluence/display/Hive/Performance> hive has
> been granted open source license. Could anybody tell me how I can get the
> license? Thanks!
>
> Cheers,
> Rui Li




-- 
Best regards!
Rui Li
Cell: (+86) 13564950210


Review Request 50787: Add a timezone-aware timestamp

2016-08-04 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/50787/
---

Review request for hive.


Bugs: HIVE-14412
https://issues.apache.org/jira/browse/HIVE-14412


Repository: hive-git


Description
---

The 1st patch to add timezone-aware timestamp.


Diffs
-

  common/src/test/org/apache/hadoop/hive/common/type/TestHiveTimestamp.java 
PRE-CREATION 
  serde/src/java/org/apache/hadoop/hive/serde2/io/TimestampWritable.java 
b90e576 
  serde/src/test/org/apache/hadoop/hive/serde2/io/TestTimestampWritable.java 
7619efa 
  storage-api/src/java/org/apache/hadoop/hive/common/type/HiveTimestamp.java 
PRE-CREATION 

Diff: https://reviews.apache.org/r/50787/diff/


Testing
---


Thanks,

Rui Li



[jira] [Created] (HIVE-14412) Add a timezone-aware timestamp

2016-08-03 Thread Rui Li (JIRA)
Rui Li created HIVE-14412:
-

 Summary: Add a timezone-aware timestamp
 Key: HIVE-14412
 URL: https://issues.apache.org/jira/browse/HIVE-14412
 Project: Hive
  Issue Type: Sub-task
Reporter: Rui Li
Assignee: Rui Li


Java's Timestamp stores the time elapsed since the epoch. While it's by itself 
unambiguous, ambiguity comes when we parse a string into timestamp, or convert 
a timestamp to string, causing problems like HIVE-14305.
To solve the issue, I think we should make timestamp aware of timezone.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-14305) To/From UTC timestamp may return incorrect result because of DST

2016-07-21 Thread Rui Li (JIRA)
Rui Li created HIVE-14305:
-

 Summary: To/From UTC timestamp may return incorrect result because 
of DST
 Key: HIVE-14305
 URL: https://issues.apache.org/jira/browse/HIVE-14305
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-14238) Ownership shouldn't be checked if external table location doesn't exist

2016-07-14 Thread Rui Li (JIRA)
Rui Li created HIVE-14238:
-

 Summary: Ownership shouldn't be checked if external table location 
doesn't exist
 Key: HIVE-14238
 URL: https://issues.apache.org/jira/browse/HIVE-14238
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


When creating external table with SQL authorization, we require RWX permission 
+ ownership of the table location. If the location doesn't exist, we check on 
parent dir (recursively), which means we require the user owns everything under 
parent dir. I think this is not necessary - we don't have to check ownership of 
parent dir, or we just check non-recursively.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-14139) NPE dropping permanent function

2016-06-30 Thread Rui Li (JIRA)
Rui Li created HIVE-14139:
-

 Summary: NPE dropping permanent function
 Key: HIVE-14139
 URL: https://issues.apache.org/jira/browse/HIVE-14139
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


To reproduce:
1. Start a CLI session and create a permanent function.
2. Exit current CLI session.
3. Start a new CLI session and drop the function.

Stack trace:
{noformat}
FAILED: error during drop function: java.lang.NullPointerException
at 
org.apache.hadoop.hive.ql.exec.Registry.removePersistentFunctionUnderLock(Registry.java:513)
at 
org.apache.hadoop.hive.ql.exec.Registry.unregisterFunction(Registry.java:501)
at 
org.apache.hadoop.hive.ql.exec.FunctionRegistry.unregisterPermanentFunction(FunctionRegistry.java:1532)
at 
org.apache.hadoop.hive.ql.exec.FunctionTask.dropPermanentFunction(FunctionTask.java:228)
at 
org.apache.hadoop.hive.ql.exec.FunctionTask.execute(FunctionTask.java:95)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:197)
at 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1860)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1564)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1316)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1085)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1073)
at 
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:232)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-13997) Insert overwrite directory doesn't overwrite existing files

2016-06-11 Thread Rui Li (JIRA)
Rui Li created HIVE-13997:
-

 Summary: Insert overwrite directory doesn't overwrite existing 
files
 Key: HIVE-13997
 URL: https://issues.apache.org/jira/browse/HIVE-13997
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


Can be easily reproduced by running {{INSERT OVERWRITE DIRECTORY}} to the same 
dir twice.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-13921) Fix spark on yarn tests for HoS

2016-06-01 Thread Rui Li (JIRA)
Rui Li created HIVE-13921:
-

 Summary: Fix spark on yarn tests for HoS
 Key: HIVE-13921
 URL: https://issues.apache.org/jira/browse/HIVE-13921
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li


{{index_bitmap3}} and {{constprog_partitioner}} have been failing. Let's fix 
them here.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-13895) HoS start-up overhead in yarn-client mode

2016-05-29 Thread Rui Li (JIRA)
Rui Li created HIVE-13895:
-

 Summary: HoS start-up overhead in yarn-client mode
 Key: HIVE-13895
 URL: https://issues.apache.org/jira/browse/HIVE-13895
 Project: Hive
  Issue Type: Bug
Reporter: Rui Li
Assignee: Rui Li


To avoid the too verbose app state report, HIVE-13376 increases the state check 
interval to a default 60s. However, bigger interval brings considerable 
start-up wait time for yarn-client mode.
Since the state report only exists in yarn-cluster mode, we can disable it 
using {{spark.yarn.submit.waitAppCompletion}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-13843) Re-enable the HoS tests disabled in HIVE-13402

2016-05-25 Thread Rui Li (JIRA)
Rui Li created HIVE-13843:
-

 Summary: Re-enable the HoS tests disabled in HIVE-13402
 Key: HIVE-13843
 URL: https://issues.apache.org/jira/browse/HIVE-13843
 Project: Hive
  Issue Type: Test
Reporter: Rui Li
Assignee: Rui Li


With HIVE-13525, we can now fix and re-enable the tests for Spark.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-13789) Repeatedly checking configuration in TextRecordWriter/Reader hurts performance

2016-05-19 Thread Rui Li (JIRA)
Rui Li created HIVE-13789:
-

 Summary: Repeatedly checking configuration in 
TextRecordWriter/Reader hurts performance
 Key: HIVE-13789
 URL: https://issues.apache.org/jira/browse/HIVE-13789
 Project: Hive
  Issue Type: Improvement
Reporter: Rui Li
Assignee: Rui Li
Priority: Minor


We check configuration to decide whether to escape certain characters each time 
write/read a record for custom scripts.
In our benchmark this becomes a hot spot method. And fixing it improves the 
execution of the custom script by 7% (3TB TPCx-BB dataset).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 47242: HIVE-13726

2016-05-12 Thread Rui Li


> On May 12, 2016, 5:48 a.m., Ashutosh Chauhan wrote:
> > ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java, line 3074
> > <https://reviews.apache.org/r/47242/diff/1/?file=1379752#file1379752line3074>
> >
> > It could have been. Feels cleaner to me as non-static. Just a matter of 
> > style. Is there any advantage of keeping it static?

Just thought the method is protected and may be accessed in sub-classes. If 
that's not a concern, I'm OK with non-static.


- Rui


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/47242/#review132822
---


On May 11, 2016, 4:36 p.m., Ashutosh Chauhan wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/47242/
> ---
> 
> (Updated May 11, 2016, 4:36 p.m.)
> 
> 
> Review request for hive and Rui Li.
> 
> 
> Bugs: HIVE-13726
> https://issues.apache.org/jira/browse/HIVE-13726
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> For insert overwrite significant amount might be spent in deleting existing 
> files. This patch parallelizes this task.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/FileUtils.java 71c9188 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java dd14124 
> 
> Diff: https://reviews.apache.org/r/47242/diff/
> 
> 
> Testing
> ---
> 
> existing regression tests. No change in functionality.
> 
> 
> Thanks,
> 
> Ashutosh Chauhan
> 
>



Re: Review Request 47242: HIVE-13726

2016-05-11 Thread Rui Li

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/47242/#review132821
---




ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (line 3074)
<https://reviews.apache.org/r/47242/#comment197074>

Any reason why remove the static? I think the newly added method can be 
static too.



ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (line 3168)
<https://reviews.apache.org/r/47242/#comment197073>

I think FileNotFoundException is not needed here.


- Rui Li


On May 11, 2016, 4:36 p.m., Ashutosh Chauhan wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/47242/
> ---
> 
> (Updated May 11, 2016, 4:36 p.m.)
> 
> 
> Review request for hive and Rui Li.
> 
> 
> Bugs: HIVE-13726
> https://issues.apache.org/jira/browse/HIVE-13726
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> For insert overwrite significant amount might be spent in deleting existing 
> files. This patch parallelizes this task.
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/common/FileUtils.java 71c9188 
>   ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java dd14124 
> 
> Diff: https://reviews.apache.org/r/47242/diff/
> 
> 
> Testing
> ---
> 
> existing regression tests. No change in functionality.
> 
> 
> Thanks,
> 
> Ashutosh Chauhan
> 
>



  1   2   3   4   5   6   7   8   >