[jira] [Commented] (HIVE-7240) Add classifier for avro-mapred jar

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14033559#comment-14033559
 ] 

Hive QA commented on HIVE-7240:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650628/HIVE-7240.1.patch

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 5536 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parquet_columnar
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join_filter
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_ctas
org.apache.hadoop.hive.ql.exec.tez.TestTezTask.testSubmit
org.apache.hive.hcatalog.pig.TestHCatLoader.testReadDataPrimitiveTypes
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/488/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/488/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-488/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12650628

 Add classifier for avro-mapred jar
 --

 Key: HIVE-7240
 URL: https://issues.apache.org/jira/browse/HIVE-7240
 Project: Hive
  Issue Type: Task
  Components: File Formats
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-7240.1.patch, HIVE-7240.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7229) String is compared using equal in HiveMetaStore#HMSHandler#init()

2014-06-17 Thread Kang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kang updated HIVE-7229:
---

Attachment: HIVE-7229

Change string compare operator

 String is compared using equal in HiveMetaStore#HMSHandler#init()
 -

 Key: HIVE-7229
 URL: https://issues.apache.org/jira/browse/HIVE-7229
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Priority: Minor
 Attachments: HIVE-7229


 Around line 423:
 {code}
   if (partitionValidationRegex != null  partitionValidationRegex != ) 
 {
 partitionValidationPattern = 
 Pattern.compile(partitionValidationRegex);
 {code}
 partitionValidationRegex.isEmpty() can be used instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7229) String is compared using equal in HiveMetaStore#HMSHandler#init()

2014-06-17 Thread Kang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kang updated HIVE-7229:
---

Status: Patch Available  (was: Open)

 String is compared using equal in HiveMetaStore#HMSHandler#init()
 -

 Key: HIVE-7229
 URL: https://issues.apache.org/jira/browse/HIVE-7229
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Priority: Minor
 Attachments: HIVE-7229


 Around line 423:
 {code}
   if (partitionValidationRegex != null  partitionValidationRegex != ) 
 {
 partitionValidationPattern = 
 Pattern.compile(partitionValidationRegex);
 {code}
 partitionValidationRegex.isEmpty() can be used instead.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6584) Add HiveHBaseTableSnapshotInputFormat

2014-06-17 Thread zjkyly (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6584?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14033591#comment-14033591
 ] 

zjkyly commented on HIVE-6584:
--

Teng YuTong and I are colleagues. we have a patch for HIVE-6584 and a patch for 
HBASE-11163 ,   and we modify
org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat(line 93)
from: static class TableSnapshotRegionSplit extends InputSplit implements 
Writable
to: public static class TableSnapshotRegionSplit extends InputSplit implements 
Writable 

we can run mapred on snapshot. mapred (count(1)) result :

2014-06-17 16:29:34,540 Stage-1 map = 100%,  reduce = 32%, Cumulative CPU 
2467.57 sec
2014-06-17 16:29:35,578 Stage-1 map = 100%,  reduce = 100%, Cumulative CPU 
2468.35 sec
MapReduce Total cumulative CPU time: 41 minutes 8 seconds 350 msec
Ended Job = job_1402970116480_0015
MapReduce Jobs Launched: 
Job 0: Map: 64  Reduce: 1   Cumulative CPU: 2468.35 sec   HDFS Read: 18334 HDFS 
Write: 9 SUCCESS
Total MapReduce CPU Time Spent: 41 minutes 8 seconds 350 msec
OK
65497163
Time taken: 429.647 seconds, Fetched: 1 row(s)

hbase count result:
Current count: 6540, row: user987684650651905350


65497163 row(s) in 1446.2310 seconds
= 65497163

but hfile has different versions of the record. We can not solve this problem. 
So, we set  the version of hbase table is 1, and run major compact before 
snapshot table.

 Add HiveHBaseTableSnapshotInputFormat
 -

 Key: HIVE-6584
 URL: https://issues.apache.org/jira/browse/HIVE-6584
 Project: Hive
  Issue Type: Improvement
  Components: HBase Handler
Reporter: Nick Dimiduk
Assignee: Nick Dimiduk
 Fix For: 0.14.0

 Attachments: HIVE-6584.0.patch, HIVE-6584.1.patch, HIVE-6584.2.patch, 
 HIVE-6584.3.patch, HIVE-6584.4.patch


 HBASE-8369 provided mapreduce support for reading from HBase table snapsopts. 
 This allows a MR job to consume a stable, read-only view of an HBase table 
 directly off of HDFS. Bypassing the online region server API provides a nice 
 performance boost for the full scan. HBASE-10642 is backporting that feature 
 to 0.94/0.96 and also adding a {{mapred}} implementation. Once that's 
 available, we should add an input format. A follow-on patch could work out 
 how to integrate this functionality into the StorageHandler, similar to how 
 HIVE-6473 integrates the HFileOutputFormat into existing table definitions.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6584) Add HiveHBaseTableSnapshotInputFormat

2014-06-17 Thread zjkyly (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6584?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14033626#comment-14033626
 ] 

zjkyly commented on HIVE-6584:
--

hi Nick Dimiduk  and Teng Yutong. what we can do at repsent is to scan all kv 
records of snapshot.
we weren't able to solve the issue that hfile had different version for record, 
so we set default version of hbase table as 1, and run major compact before 
doing snapshot table.
we don't know the open soure community, is it necessary to solve the problem of 
same kv version? shall we find all kv version ? or just find the latest one?
We will try to solve the problem of multiple versions.



 Add HiveHBaseTableSnapshotInputFormat
 -

 Key: HIVE-6584
 URL: https://issues.apache.org/jira/browse/HIVE-6584
 Project: Hive
  Issue Type: Improvement
  Components: HBase Handler
Reporter: Nick Dimiduk
Assignee: Nick Dimiduk
 Fix For: 0.14.0

 Attachments: HIVE-6584.0.patch, HIVE-6584.1.patch, HIVE-6584.2.patch, 
 HIVE-6584.3.patch, HIVE-6584.4.patch


 HBASE-8369 provided mapreduce support for reading from HBase table snapsopts. 
 This allows a MR job to consume a stable, read-only view of an HBase table 
 directly off of HDFS. Bypassing the online region server API provides a nice 
 performance boost for the full scan. HBASE-10642 is backporting that feature 
 to 0.94/0.96 and also adding a {{mapred}} implementation. Once that's 
 available, we should add an input format. A follow-on patch could work out 
 how to integrate this functionality into the StorageHandler, similar to how 
 HIVE-6473 integrates the HFileOutputFormat into existing table definitions.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread steve, Oh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

steve, Oh updated HIVE-7225:


Attachment: HIVE-7225.1.patch

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread steve, Oh (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

steve, Oh updated HIVE-7225:


Status: Patch Available  (was: Open)

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Raw MapReduce code read RC file using HCatalog

2014-06-17 Thread Fastupload
ps: I use Hive 0.12 release and Hadoop 1.2.1 release.

On Jun 17, 2014, at 5:07 PM, Fastupload fastupl...@outlook.com wrote:

 Hi,
 
 When MapReduce code read RC file using HCatalog, and hive meta store in a 
 remote Oracle database. I write a demo code following by the wiki page, 
 https://cwiki.apache.org/confluence/display/Hive/HCatalog+InputOutput, and 
 package all dependence jar into one jar.
 
 The job got the error while running. some line of failure stack is:
 Caused by: org.datanucleus.exceptions.NucleusUserException: Persistence 
 process has been specified to use a ClassLoaderResolver of name datanucleus 
 yet this has not been found by the DataNucleus plugin mechanism. Please check 
 your CLASSPATH and plugin specification.
 
 It seems that the HCatInputFormat class can not create a 
 JDOPersisitencemanagerFactory object for HiveMetaStoreClient object in 
 InitializeInput.java from line 101 to 106,
 if (conf != null) {
 hiveConf = HCatUtil.getHiveConf(conf);
   } else {
 hiveConf = new HiveConf(HCatInputFormat.class);
   }
   client = HCatUtil.getHiveClient(hiveConf);
 
 The lines of code create a HiveMetaStoreClient object with job configuration 
 or HCatInputFormat properties. So I add code to add hive-site.xml file both 
 job configuration and HCatInputFormat properties. as,
 
   // load hive meta store configuration file, both properties and 
 job config
   Properties prop = new Properties();
   FileInputStream confStream = new FileInputStream(args[4]);
   prop.load(confStream);
   conf.addResource(confStream);
   HCatInputFormat.setInput(job, dbName, 
 tblName).setFilter(filter).setProperties(prop);
 
 But the job still get the same error.  any idea?  
 More error log and code please look at the two attachments.
 error log.txt
 code fragment.txt
 
 Best Regards,
 Link Qian



[jira] [Updated] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-7225:
-

Status: Open  (was: Patch Available)

Steve, thanks for the patch.  A few comments:
I don't think there's a need to explicitly close the ResultSet.  According to 
the JavaDocs on Statement.close(), it closes any associated ResultSets.  So 
making this call seems like a waste of time.
If we're going to do this, it should be done in CompactionTxnHanlder as well, 
since that subclasses this class and uses all of the same methods.
But that if leads to my last comment.  Is this necessary at all?  All of the 
public methods close the db connection as soon as they are done.  Is there any 
value to closing the individual statements in that collection first?

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Question about org.apache.hive.service.cli.Column class

2014-06-17 Thread Mahdi Soleimani
Hi

I want to implement some compression algorithms on the data that is exchanged 
between clients and HS2. Just wanted to make sure that the Get function is 
called on the client side, Right?

Cheers


[jira] [Commented] (HIVE-7219) Improve performance of serialization utils in ORC

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034100#comment-14034100
 ] 

Hive QA commented on HIVE-7219:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650656/HIVE-7219.4.patch

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 5653 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parquet_columnar
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_ctas
org.apache.hadoop.hive.ql.exec.tez.TestTezTask.testSubmit
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/491/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/491/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-491/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12650656

 Improve performance of serialization utils in ORC
 -

 Key: HIVE-7219
 URL: https://issues.apache.org/jira/browse/HIVE-7219
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
 Attachments: HIVE-7219.1.patch, HIVE-7219.2.patch, HIVE-7219.3.patch, 
 HIVE-7219.4.patch, orc-read-perf-jmh-benchmark.png


 ORC uses serialization utils heavily for reading and writing data. The 
 bitpacking and unpacking code in writeInts() and readInts() can be unrolled 
 for better performance. Also double reader/writer performance can be improved 
 by bulk reading/writing from/to byte array.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7219) Improve performance of serialization utils in ORC

2014-06-17 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-7219:
-

   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks [~hagleitn] and [~gopalv] for the code review.

 Improve performance of serialization utils in ORC
 -

 Key: HIVE-7219
 URL: https://issues.apache.org/jira/browse/HIVE-7219
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
 Fix For: 0.14.0

 Attachments: HIVE-7219.1.patch, HIVE-7219.2.patch, HIVE-7219.3.patch, 
 HIVE-7219.4.patch, orc-read-perf-jmh-benchmark.png


 ORC uses serialization utils heavily for reading and writing data. The 
 bitpacking and unpacking code in writeInts() and readInts() can be unrolled 
 for better performance. Also double reader/writer performance can be improved 
 by bulk reading/writing from/to byte array.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034148#comment-14034148
 ] 

Ted Yu commented on HIVE-7225:
--

Does this answer your question ?
http://stackoverflow.com/questions/2708689/impact-of-java-sql-connection-close-on-java-sql-statement-objects-and-the-like

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034192#comment-14034192
 ] 

Alan Gates commented on HIVE-7225:
--

No.  We aren't leaving these connections open for a long time.  We're closing 
them quite quickly at the end of each Thrift RPC.  If we were leaving them open 
for a long time then I would agree that it was important to close the 
statements.

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-7246) Hive transaction manager hardwires bonecp as the JDBC pooling implementation

2014-06-17 Thread Alan Gates (JIRA)
Alan Gates created HIVE-7246:


 Summary: Hive transaction manager hardwires bonecp as the JDBC 
pooling implementation
 Key: HIVE-7246
 URL: https://issues.apache.org/jira/browse/HIVE-7246
 Project: Hive
  Issue Type: Bug
  Components: Transactions
Affects Versions: 0.13.0
Reporter: Alan Gates
Assignee: Alan Gates


Currently TxnManager hardwires BoneCP as the JDBC connection pooling 
implementation.  Instead it should use the same connection pooling that the 
metastore does.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-7094:
-

Status: Open  (was: Patch Available)

Looks like some tests failed. [~davidzchen], can you please take a look? Thanks.

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Status: Open  (was: Patch Available)

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Status: Patch Available  (was: Open)

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Attachment: HIVE-7076.2.patch

Different route. Patch .2 simply excludes the ATSHook from compilation on 
hadoop-1 profile

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7236) Tez progress monitor should indicate running/failed tasks

2014-06-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-7236:
--

Release Note: Log Tez running/failed tasks in TezJobMonitor
  Status: Patch Available  (was: In Progress)

 Tez progress monitor should indicate running/failed tasks
 -

 Key: HIVE-7236
 URL: https://issues.apache.org/jira/browse/HIVE-7236
 Project: Hive
  Issue Type: Bug
  Components: Tez
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Gopal V
Priority: Minor
 Attachments: HIVE-7236.1.patch


 Currently, the only logging in TezJobMonitor is for completed tasks. 
 This makes it hard to locate task stalls and task failures. Failure scenarios 
 are harder to debug, in particular when analyzing query runs on a cluster 
 with bad nodes.
 Change the job monitor to log running  failed tasks as follows.
 {code}
 Map 1: 0(+157,-1)/1755 Reducer 2: 0/1  
 Map 1: 0(+168,-1)/1755 Reducer 2: 0/1  
 Map 1: 0(+189,-1)/1755 Reducer 2: 0/1  
 Map 1: 0(+189,-1)/1755 Reducer 2: 0/1 
 {code}
 That is 189 tasks running, 1 failure and 0 complete.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7119) Extended ACL's should be inherited if warehouse perm inheritance enabled

2014-06-17 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7119?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034293#comment-14034293
 ] 

Gunther Hagleitner commented on HIVE-7119:
--

After this patch itests cannot be compiled against hadoop-1 profile anymore. 
Could you take a look?

$cd itests
$mvn install -DskipTests -Phadoop-1
...
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
(default-testCompile) on project hive-it-unit: Compilation failure: Compilation 
failure:
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
 cannot find symbol
[ERROR] symbol  : class AclStatus
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
 cannot find symbol
[ERROR] symbol  : class AclEntryScope
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
 static import only from classes and interfaces
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
 cannot find symbol
[ERROR] symbol  : class AclEntryType
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
 static import only from classes and interfaces
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
 cannot find symbol
[ERROR] symbol  : class AclEntryType
...

 Extended ACL's should be inherited if warehouse perm inheritance enabled
 

 Key: HIVE-7119
 URL: https://issues.apache.org/jira/browse/HIVE-7119
 Project: Hive
  Issue Type: Bug
Reporter: Szehon Ho
Assignee: Szehon Ho
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-7119.2.patch, HIVE-7119.3.patch, HIVE-7119.4.patch, 
 HIVE-7119.patch


 HDFS recently came out with support for extended ACL's, ie permission for 
 specific group/user in addition to the general owner/group/other permission.
 Hive permission inheritance should also inherit those as well, if user has 
 set them at any point in the warehouse directory.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7105) Enable ReduceRecordProcessor to generate VectorizedRowBatches

2014-06-17 Thread Jitendra Nath Pandey (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7105?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034320#comment-14034320
 ] 

Jitendra Nath Pandey commented on HIVE-7105:


[~rusanu]  Here is the RB link: https://reviews.apache.org/r/22540/

 Enable ReduceRecordProcessor to generate VectorizedRowBatches
 -

 Key: HIVE-7105
 URL: https://issues.apache.org/jira/browse/HIVE-7105
 Project: Hive
  Issue Type: Bug
  Components: Tez, Vectorization
Reporter: Rajesh Balamohan
Assignee: Gopal V
 Fix For: 0.14.0

 Attachments: HIVE-7105.1.patch, HIVE-7105.2.patch


 Currently, ReduceRecordProcessor sends one key,value pair at a time to its 
 operator pipeline.  It would be beneficial to send VectorizedRowBatch to 
 downstream operators. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7232) ReduceSink is emitting NULL keys due to failed keyEval

2014-06-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7232?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034322#comment-14034322
 ] 

Gopal V commented on HIVE-7232:
---

[~navis]: I found out that there are indeed o_orderkey entries which show up as 
214800 in text, which lies outside the range of the TPC-H Identifier column 
spec.

I will reload the data using bigint for o_orderkey soon.

But I still want to locate and confirm the different results between MR and Tez 
here.

 ReduceSink is emitting NULL keys due to failed keyEval
 --

 Key: HIVE-7232
 URL: https://issues.apache.org/jira/browse/HIVE-7232
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Navis
 Attachments: HIVE-7232-extra-logging.patch, q5.explain.txt, q5.sql


 After HIVE-4867 has been merged in, some queries have exhibited a very weird 
 skew towards NULL keys emitted from the ReduceSinkOperator.
 Added extra logging to print expr.column() in ExprNodeColumnEvaluator  in 
 reduce sink.
 {code}
 2014-06-14 00:37:19,186 INFO [TezChild] 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator:
 numDistributionKeys = 1 {null -- ExprNodeColumnEvaluator(_col10)}
 key_row={reducesinkkey0:442}
 {code}
 {code}
   HiveKey firstKey = toHiveKey(cachedKeys[0], tag, null);
   int distKeyLength = firstKey.getDistKeyLength();
   if(distKeyLength = 1) {
 StringBuffer x1 = new StringBuffer();
 x1.append(numDistributionKeys = + numDistributionKeys + \n);
 for (int i = 0; i  numDistributionKeys; i++) {
 x1.append(cachedKeys[0][i] +  --  + keyEval[i] + \n);
 }
 x1.append(key_row=+ SerDeUtils.getJSONString(row, 
 keyObjectInspector));
 LOG.info(GOPAL:  + x1.toString());
   }
 {code}
 The query is tpc-h query5, with extra NULL checks just to be sure.
 {code}
 ELECT n_name,
sum(l_extendedprice * (1 - l_discount)) AS revenue
 FROM customer,
  orders,
  lineitem,
  supplier,
  nation,
  region
 WHERE c_custkey = o_custkey
   AND l_orderkey = o_orderkey
   AND l_suppkey = s_suppkey
   AND c_nationkey = s_nationkey
   AND s_nationkey = n_nationkey
   AND n_regionkey = r_regionkey
   AND r_name = 'ASIA'
   AND o_orderdate = '1994-01-01'
   AND o_orderdate  '1995-01-01'
   and l_orderkey is not null
   and c_custkey is not null
   and l_suppkey is not null
   and c_nationkey is not null
   and s_nationkey is not null
   and n_regionkey is not null
 GROUP BY n_name
 ORDER BY revenue DESC;
 {code}
 The reducer which has the issue has the following plan
 {code}
 Reducer 3
 Reduce Operator Tree:
   Join Operator
 condition map:
  Inner Join 0 to 1
 condition expressions:
   0 {KEY.reducesinkkey0} {VALUE._col2}
   1 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col3}
 outputColumnNames: _col0, _col3, _col10, _col11, _col14
 Statistics: Num rows: 18344 Data size: 95229140992 Basic 
 stats: COMPLETE Column stats: NONE
 Reduce Output Operator
   key expressions: _col10 (type: int)
   sort order: +
   Map-reduce partition columns: _col10 (type: int)
   Statistics: Num rows: 18344 Data size: 95229140992 
 Basic stats: COMPLETE Column stats: NONE
   value expressions: _col0 (type: int), _col3 (type: int), 
 _col11 (type: int), _col14 (type: string)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Status: Open  (was: Patch Available)

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Status: Patch Available  (was: Open)

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch, HIVE-7076.3.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034338#comment-14034338
 ] 

Gunther Hagleitner commented on HIVE-7076:
--

.3 addresses review comments.

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch, HIVE-7076.3.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-7076:
-

Attachment: HIVE-7076.3.patch

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch, HIVE-7076.3.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7190) WebHCat launcher task failure can cause two concurent user jobs to run

2014-06-17 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7190?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-7190:


   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Patch committed to trunk. Thanks for the contribution Ivan, thanks for the 
review Eugene!


 WebHCat launcher task failure can cause two concurent user jobs to run
 --

 Key: HIVE-7190
 URL: https://issues.apache.org/jira/browse/HIVE-7190
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Ivan Mitic
Assignee: Ivan Mitic
 Fix For: 0.14.0

 Attachments: HIVE-7190.2.patch, HIVE-7190.3.patch, HIVE-7190.patch


 Templeton uses launcher jobs to launch the actual user jobs. Launcher jobs 
 are 1-map jobs (a single task jobs) which kick off the actual user job and 
 monitor it until it finishes. Given that the launcher is a task, like any 
 other MR task, it has a retry policy in case it fails (due to a task crash, 
 tasktracker/nodemanager crash, machine level outage, etc.). Further, when 
 launcher task is retried, it will again launch the same user job, *however* 
 the previous attempt user job is already running. What this means is that we 
 can have two identical user jobs running in parallel. 
 In case of MRv2, there will be an MRAppMaster and the launcher task, which 
 are subject to failure. In case any of the two fails, another instance of a 
 user job will be launched again in parallel. 
 Above situation is already a bug.
 Now going further to RM HA, what RM does on failover/restart is that it kills 
 all containers, and it restarts all applications. This means that if our 
 customer had 10 jobs on the cluster (this is 10 launcher jobs and 10 user 
 jobs), on RM failover, all 20 jobs will be restarted, and launcher jobs will 
 queue user jobs again. There are two issues with this design:
 1. There are *possible* chances for corruption of job outputs (it would be 
 useful to analyze this scenario more and confirm this statement).
 2. Cluster resources are spent on jobs redundantly
 To address the issue at least on Yarn (Hadoop 2.0) clusters, webhcat should 
 do the same thing Oozie does in this scenario, and that is to tag all its 
 child jobs with an id, and kill those jobs on task restart before they are 
 kicked off again.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7225) Unclosed Statement's in TxnHandler

2014-06-17 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034400#comment-14034400
 ] 

Ted Yu commented on HIVE-7225:
--

Hive depends on BoneCP 0.8.0

I cloned BoneCP locally and looked at their code.
{code}
public class ConnectionHandle implements Connection,Serializable{
{code}
In ConnectionHandle ctor, I see:
{code}
this.detectUnclosedStatements = 
pool.getConfig().isDetectUnclosedStatements();
this.closeOpenStatements = pool.getConfig().isCloseOpenStatements();
{code}
In close() method:
{code}
if (this.closeOpenStatements){
  for (EntryStatement, String statementEntry: 
this.trackedStatement.entrySet()){
statementEntry.getKey().close();
if (this.detectUnclosedStatements){
  logger.warn(String.format(UNCLOSED_LOG_ERROR_MESSAGE, 
statementEntry.getValue()));
}
  }
  this.trackedStatement.clear();
}
{code}
In Hive codebase, I don't see call that turns on closeOpenStatements.

 Unclosed Statement's in TxnHandler
 --

 Key: HIVE-7225
 URL: https://issues.apache.org/jira/browse/HIVE-7225
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: steve, Oh
 Attachments: HIVE-7225.1.patch


 There're several methods in TxnHandler where Statement (local to the method) 
 is not closed upon return.
 Here're a few examples:
 In compact():
 {code}
 stmt.executeUpdate(s);
 LOG.debug(Going to commit);
 dbConn.commit();
 {code}
 In showCompact():
 {code}
   Statement stmt = dbConn.createStatement();
   String s = select cq_database, cq_table, cq_partition, cq_state, 
 cq_type, cq_worker_id,  +
   cq_start, cq_run_as from COMPACTION_QUEUE;
   LOG.debug(Going to execute query  + s + );
   ResultSet rs = stmt.executeQuery(s);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7231) Improve ORC padding

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034409#comment-14034409
 ] 

Hive QA commented on HIVE-7231:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650660/HIVE-7231.2.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 5536 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parquet_columnar
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_ctas
org.apache.hadoop.hive.ql.exec.tez.TestTezTask.testSubmit
org.apache.hive.hcatalog.pig.TestHCatLoader.testReadDataPrimitiveTypes
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/492/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/492/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-492/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12650660

 Improve ORC padding
 ---

 Key: HIVE-7231
 URL: https://issues.apache.org/jira/browse/HIVE-7231
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-7231.1.patch, HIVE-7231.2.patch


 Current ORC padding is not optimal because of fixed stripe sizes within 
 block. The padding overhead will be significant in some cases. Also padding 
 percentage relative to stripe size is not configurable.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034418#comment-14034418
 ] 

David Chen commented on HIVE-7094:
--

The Hive CLI and QL test failures are unrelated to my patch.

The only one that is concerning to me is the TestOrcHCatLoader failure. 
However, I am not hitting that failure when I run the tests on my machine, but 
I have rebased my changes on the current trunk. Nonetheless, all of the 
HCatalog Core tests pass.

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread David Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Chen updated HIVE-7094:
-

Attachment: HIVE-7094.5.patch

Attaching a new patch rebased on trunk.

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch, 
 HIVE-7094.5.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-7159:


Attachment: HIVE-7159.7.patch

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread David Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Chen updated HIVE-7094:
-

Status: Patch Available  (was: Open)

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch, 
 HIVE-7094.5.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7242) alter table drop partition is acquiring the wrong type of lock

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034420#comment-14034420
 ] 

Hive QA commented on HIVE-7242:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650664/HIVE-7242.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/493/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/493/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-493/

Messages:
{noformat}
 This message was trimmed, see log for full details 
Decision can match input such as LPAREN KW_CASE KW_ARRAY using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE TinyintLiteral using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE KW_STRUCT using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE SmallintLiteral using 
multiple alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:115:5: 
Decision can match input such as KW_CLUSTER KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:127:5: 
Decision can match input such as KW_PARTITION KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:138:5: 
Decision can match input such as KW_DISTRIBUTE KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:149:5: 
Decision can match input such as KW_SORT KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:166:7: 
Decision can match input such as STAR using multiple alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_STRUCT using multiple alternatives: 4, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_UNIONTYPE using multiple alternatives: 5, 
6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_ARRAY using multiple alternatives: 2, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_DATE StringLiteral using multiple 
alternatives: 2, 3

As a result, alternative(s) 3 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_FALSE using multiple alternatives: 3, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_TRUE using multiple alternatives: 3, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_NULL using multiple alternatives: 1, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_OVERWRITE using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_DISTRIBUTE 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_MAP LPAREN 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_INTO using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_LATERAL 
KW_VIEW using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_GROUP 

[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034421#comment-14034421
 ] 

Swarnim Kulkarni commented on HIVE-7230:


Go for it!

 The only comment I had was it would probably be nice to pull down the style 
guide xml file locally so that our builds don't fail in case the remote server 
goes down.

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034426#comment-14034426
 ] 

Harish Butani commented on HIVE-7159:
-

The diff in windowing queries is because of an issue with ColumnPruner for PTF 
and Script Operators.
[~ashutoshc] can you please look at changes to ColumnPruner.

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7231) Improve ORC padding

2014-06-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034434#comment-14034434
 ] 

Gopal V commented on HIVE-7231:
---

The approach results in stray writes across the stripe boundaries.

I think this approach needs to be revisited to disconnect the HDFS block size 
from the ORC stripe size.

The stripe size needs to be a factor of the HDFS block size, but the fraction 
should not remain at 0.5x.

 Improve ORC padding
 ---

 Key: HIVE-7231
 URL: https://issues.apache.org/jira/browse/HIVE-7231
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-7231.1.patch, HIVE-7231.2.patch


 Current ORC padding is not optimal because of fixed stripe sizes within 
 block. The padding overhead will be significant in some cases. Also padding 
 percentage relative to stripe size is not configurable.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034436#comment-14034436
 ] 

Sushanth Sowmyan commented on HIVE-7094:


Some of the HCat tests are flaky when run on the ptest framework (they succeed 
when run on a serial run) - that's on my tofix radar. But in the meanwhile, 
I'll try running the HCat tests with David's patch to verify.

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch, 
 HIVE-7094.5.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034454#comment-14034454
 ] 

David Chen commented on HIVE-7230:
--

I am going to try to point the Maven Eclipse plugin to the local 
eclipse-styles.xml that I checked in with this patch. The code snipped above 
seems to point to the Google Java coding style.

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034466#comment-14034466
 ] 

Harish Butani commented on HIVE-7159:
-

The issue with join pushdown is that now genNotNullFilterForJoinSourcePlan is 
always working of the top QBJoinTree; in the original patch the joinTree was 
passed in, somewhere along the line this got refactored, looking into reverting.

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7185) KeyWrapperFactory#TextKeyWrapper#equals() extracts Text incorrectly when isCopy is false

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034510#comment-14034510
 ] 

Swarnim Kulkarni commented on HIVE-7185:


[~suyeon1222] Any test failures after you made this update? If not can you add 
a test to show what was broken before this change and show how this change 
fixes that?

 KeyWrapperFactory#TextKeyWrapper#equals() extracts Text incorrectly when 
 isCopy is false
 

 Key: HIVE-7185
 URL: https://issues.apache.org/jira/browse/HIVE-7185
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Assignee: SUYEON LEE
Priority: Minor
 Attachments: HIVE-7185.patch


 {code}
   } else {
 t1 = soi_new.getPrimitiveWritableObject(key);
 t2 = soi_copy.getPrimitiveWritableObject(obj);
 {code}
 t2 should be assigned soi_new.getPrimitiveWritableObject(obj)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread Sushanth Sowmyan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034531#comment-14034531
 ] 

Sushanth Sowmyan commented on HIVE-7094:


The only test that fails in my environment with this patch is 
TestHCatPartitionPublish, and that's flaky in how it fails - it succeeds when 
run by itself, as opposed to after other tests run, and that has nothing to do 
with this patch. All other tests under hcatalog succeed.

 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch, 
 HIVE-7094.5.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)
Szehon Ho created HIVE-7247:
---

 Summary: Fix itests using hadoop-1 profile 
 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho






--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19830: HIVE-6694: Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Swarnim Kulkarni

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19830/#review46006
---



beeline/src/java/org/apache/hive/beeline/Commands.java
https://reviews.apache.org/r/19830/#comment81167

Should we be returning a false at this point since the command failed?



itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java
https://reviews.apache.org/r/19830/#comment81168

Is it possible to test a failed command as well?


- Swarnim Kulkarni


On June 17, 2014, 1:49 a.m., Xuefu Zhang wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/19830/
 ---
 
 (Updated June 17, 2014, 1:49 a.m.)
 
 
 Review request for hive.
 
 
 Bugs: HIVE-6694
 https://issues.apache.org/jira/browse/HIVE-6694
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Introduce the new requested feature.
 
 
 Diffs
 -
 
   beeline/src/java/org/apache/hive/beeline/BeeLine.java dcdd6f2 
   beeline/src/java/org/apache/hive/beeline/Commands.java 94f74d2 
   beeline/src/main/resources/BeeLine.properties 697c29a 
   cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java 3357414 
   
 itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java
  4d15e3c 
   itests/util/pom.xml b38c55b 
   itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java 2974327 
   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 32c2fb0 
   ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java 2ce4dbd 
   ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java d2e122d 
 
 Diff: https://reviews.apache.org/r/19830/diff/
 
 
 Testing
 ---
 
 New test cases are added. 
 
 
 Thanks,
 
 Xuefu Zhang
 




[jira] [Commented] (HIVE-7119) Extended ACL's should be inherited if warehouse perm inheritance enabled

2014-06-17 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7119?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034537#comment-14034537
 ] 

Szehon Ho commented on HIVE-7119:
-

Sorry about that, I'll take a look, made HIVE-7247 

 Extended ACL's should be inherited if warehouse perm inheritance enabled
 

 Key: HIVE-7119
 URL: https://issues.apache.org/jira/browse/HIVE-7119
 Project: Hive
  Issue Type: Bug
Reporter: Szehon Ho
Assignee: Szehon Ho
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-7119.2.patch, HIVE-7119.3.patch, HIVE-7119.4.patch, 
 HIVE-7119.patch


 HDFS recently came out with support for extended ACL's, ie permission for 
 specific group/user in addition to the general owner/group/other permission.
 Hive permission inheritance should also inherit those as well, if user has 
 set them at any point in the warehouse directory.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7094) Separate out static/dynamic partitioning code in FileRecordWriterContainer

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034544#comment-14034544
 ] 

David Chen commented on HIVE-7094:
--

Thanks for your help, [~sushanth]! I have a separate ticket open to track down 
the flakiness in TestHCatPartitionPublish (HIVE-7110).

Is there any more feedback before this patch can be committed?



 Separate out static/dynamic partitioning code in FileRecordWriterContainer
 --

 Key: HIVE-7094
 URL: https://issues.apache.org/jira/browse/HIVE-7094
 Project: Hive
  Issue Type: Sub-task
  Components: HCatalog
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7094.1.patch, HIVE-7094.3.patch, HIVE-7094.4.patch, 
 HIVE-7094.5.patch


 There are two major places in FileRecordWriterContainer that have the {{if 
 (dynamicPartitioning)}} condition: the constructor and write().
 This is the approach that I am taking:
 # Move the DP and SP code into two subclasses: 
 DynamicFileRecordWriterContainer and StaticFileRecordWriterContainer.
 # Make FileRecordWriterContainer an abstract class that contains the common 
 code for both implementations. For write(), FileRecordWriterContainer will 
 call an abstract method that will provide the local RecordWriter, 
 ObjectInspector, SerDe, and OutputJobInfo.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6694) Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034543#comment-14034543
 ] 

Swarnim Kulkarni commented on HIVE-6694:


Left few minor comments on RB.

 Beeline should provide a way to execute shell command as Hive CLI does
 --

 Key: HIVE-6694
 URL: https://issues.apache.org/jira/browse/HIVE-6694
 Project: Hive
  Issue Type: Improvement
  Components: CLI, Clients
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.14.0

 Attachments: HIVE-6694.1.patch, HIVE-6694.1.patch, HIVE-6694.2.patch, 
 HIVE-6694.3.patch, HIVE-6694.patch


 Hive CLI allows a user to execute a shell command using ! notation. For 
 instance, !cat myfile.txt. Being able to execute shell command may be 
 important for some users. As a replacement, however, Beeline provides no such 
 capability, possibly because ! notation is reserved for SQLLine commands. 
 It's possible to provide this using a slightly syntactic variation such as 
 !sh cat myfilie.txt.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 22612: HIVE-7234: Handle nulls from decimal columns elegantly

2014-06-17 Thread Swarnim Kulkarni

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/22612/#review46008
---



common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java
https://reviews.apache.org/r/22612/#comment81170

This doesn't seem specific to the HiveDecimal class. Why not simply make it 
private and move to the class where it is actually being used?



itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDecimal.java
https://reviews.apache.org/r/22612/#comment81171

I think this class is nice. But is it possible to throw in a .q file as 
well with queries testing querying of decimals greater than 99


- Swarnim Kulkarni


On June 16, 2014, 2:06 a.m., Ashish Singh wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/22612/
 ---
 
 (Updated June 16, 2014, 2:06 a.m.)
 
 
 Review request for hive, Szehon Ho and Xuefu Zhang.
 
 
 Bugs: HIVE-7234
 https://issues.apache.org/jira/browse/HIVE-7234
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 HIVE-7234: Handle nulls from decimal columns elegantly
 
 
 Diffs
 -
 
   common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java 
 ad0901548217fbb828a01f8f5edda64581ac2c1e 
   data/files/decimal_10_0.txt PRE-CREATION 
   data/files/decimal_9_0.txt PRE-CREATION 
   itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDecimal.java 
 PRE-CREATION 
   serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java 
 78cc3819c61f5a1bcb0cdd3425a0105416c26861 
   serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 
 5a4623729ec955bbe8fcf662503b42ff8735eaad 
 
 Diff: https://reviews.apache.org/r/22612/diff/
 
 
 Testing
 ---
 
 Added unit tests to test the scenario.
 
 
 Thanks,
 
 Ashish Singh
 




[jira] [Created] (HIVE-7248) UNION ALL in hive returns incorrect results on Hbase backed table

2014-06-17 Thread Mala Chikka Kempanna (JIRA)
Mala Chikka Kempanna created HIVE-7248:
--

 Summary: UNION ALL in hive returns incorrect results on Hbase 
backed table
 Key: HIVE-7248
 URL: https://issues.apache.org/jira/browse/HIVE-7248
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.13.1, 0.13.0, 0.12.0
Reporter: Mala Chikka Kempanna


The issue can be recreated with following steps

1) In hbase 
create 'TABLE_EMP','default' 

2) On hive 
sudo -u hive hive 

CREATE EXTERNAL TABLE TABLE_EMP(FIRST_NAME string,LAST_NAME 
string,CDS_UPDATED_DATE string,CDS_PK string) STORED BY 
'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH 
SERDEPROPERTIES(hbase.columns.mapping = 
default:FIRST_NAME,default:LAST_NAME,default:CDS_UPDATED_DATE,:key, 
hbase.scan.cache = 500, hbase.scan.cacheblocks = false ) 
TBLPROPERTIES(hbase.table.name = TABLE_EMP,'serialization.null.format'=''); 


3) On hbase insert the following data 

put 'TABLE_EMP', '1', 'default:FIRST_NAME', 'Srini' 
put 'TABLE_EMP', '1', 'default:LAST_NAME', 'P' 
put 'TABLE_EMP', '1', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 


put 'TABLE_EMP', '2', 'default:FIRST_NAME', 'Aravind' 
put 'TABLE_EMP', '2', 'default:LAST_NAME', 'K' 
put 'TABLE_EMP', '2', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 


4) On hive execute the following query 
hive 
SELECT * 
FROM ( 
SELECT CDS_PK 
FROM TABLE_EMP 
WHERE 
CDS_PK = '0' 
AND CDS_PK = '9' 
AND CDS_UPDATED_DATE IS NOT NULL 
UNION ALL SELECT CDS_PK 
FROM TABLE_EMP 
WHERE 
CDS_PK = 'a' 
AND CDS_PK = 'z' 
AND CDS_UPDATED_DATE IS NOT NULL 
)t ; 


5) Output of the query 

1 
1 
2 
2 

6) Output of just 

SELECT CDS_PK 
FROM TABLE_EMP 
WHERE 
CDS_PK = '0' 
AND CDS_PK = '9' 
AND CDS_UPDATED_DATE IS NOT NULL 

is 

1 
2 

7) Output of just 

SELECT CDS_PK 
FROM TABLE_EMP 
WHERE 
CDS_PK = 'a' 
AND CDS_PK = 'z' 
AND CDS_UPDATED_DATE IS NOT NULL 

Empty 

8) UNION is used to combine the result from multiple SELECT statements into a 
single result set. Hive currently only supports UNION ALL (bag union), in which 
duplicates are not eliminated 

Accordingly above query should return output 
1 
2 

instead it is giving wrong output 
1 
1 
2 
2



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7248) UNION ALL in hive returns incorrect results on Hbase backed table

2014-06-17 Thread Mala Chikka Kempanna (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034555#comment-14034555
 ] 

Mala Chikka Kempanna commented on HIVE-7248:


Current work-around for the problem is to use DISTINCT in one of the sub 
queries.

 UNION ALL in hive returns incorrect results on Hbase backed table
 -

 Key: HIVE-7248
 URL: https://issues.apache.org/jira/browse/HIVE-7248
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.12.0, 0.13.0, 0.13.1
Reporter: Mala Chikka Kempanna

 The issue can be recreated with following steps
 1) In hbase 
 create 'TABLE_EMP','default' 
 2) On hive 
 sudo -u hive hive 
 CREATE EXTERNAL TABLE TABLE_EMP(FIRST_NAME string,LAST_NAME 
 string,CDS_UPDATED_DATE string,CDS_PK string) STORED BY 
 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH 
 SERDEPROPERTIES(hbase.columns.mapping = 
 default:FIRST_NAME,default:LAST_NAME,default:CDS_UPDATED_DATE,:key, 
 hbase.scan.cache = 500, hbase.scan.cacheblocks = false ) 
 TBLPROPERTIES(hbase.table.name = 
 TABLE_EMP,'serialization.null.format'=''); 
 3) On hbase insert the following data 
 put 'TABLE_EMP', '1', 'default:FIRST_NAME', 'Srini' 
 put 'TABLE_EMP', '1', 'default:LAST_NAME', 'P' 
 put 'TABLE_EMP', '1', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 
 put 'TABLE_EMP', '2', 'default:FIRST_NAME', 'Aravind' 
 put 'TABLE_EMP', '2', 'default:LAST_NAME', 'K' 
 put 'TABLE_EMP', '2', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 
 4) On hive execute the following query 
 hive 
 SELECT * 
 FROM ( 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = '0' 
 AND CDS_PK = '9' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 UNION ALL SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = 'a' 
 AND CDS_PK = 'z' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 )t ; 
 5) Output of the query 
 1 
 1 
 2 
 2 
 6) Output of just 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = '0' 
 AND CDS_PK = '9' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 is 
 1 
 2 
 7) Output of just 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = 'a' 
 AND CDS_PK = 'z' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 Empty 
 8) UNION is used to combine the result from multiple SELECT statements into a 
 single result set. Hive currently only supports UNION ALL (bag union), in 
 which duplicates are not eliminated 
 Accordingly above query should return output 
 1 
 2 
 instead it is giving wrong output 
 1 
 1 
 2 
 2



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 22612: HIVE-7234: Handle nulls from decimal columns elegantly

2014-06-17 Thread Swarnim Kulkarni

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/22612/#review46009
---



serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java
https://reviews.apache.org/r/22612/#comment81172

Since it's a public method, some javadoc on this method would be nice.


- Swarnim Kulkarni


On June 16, 2014, 2:06 a.m., Ashish Singh wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/22612/
 ---
 
 (Updated June 16, 2014, 2:06 a.m.)
 
 
 Review request for hive, Szehon Ho and Xuefu Zhang.
 
 
 Bugs: HIVE-7234
 https://issues.apache.org/jira/browse/HIVE-7234
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 HIVE-7234: Handle nulls from decimal columns elegantly
 
 
 Diffs
 -
 
   common/src/java/org/apache/hadoop/hive/common/type/HiveDecimal.java 
 ad0901548217fbb828a01f8f5edda64581ac2c1e 
   data/files/decimal_10_0.txt PRE-CREATION 
   data/files/decimal_9_0.txt PRE-CREATION 
   itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/TestDecimal.java 
 PRE-CREATION 
   serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyHiveDecimal.java 
 78cc3819c61f5a1bcb0cdd3425a0105416c26861 
   serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java 
 5a4623729ec955bbe8fcf662503b42ff8735eaad 
 
 Diff: https://reviews.apache.org/r/22612/diff/
 
 
 Testing
 ---
 
 Added unit tests to test the scenario.
 
 
 Thanks,
 
 Ashish Singh
 




[jira] [Commented] (HIVE-7234) Select on decimal column throws NPE

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7234?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034563#comment-14034563
 ] 

Swarnim Kulkarni commented on HIVE-7234:


Left some minor comments on RB.

 Select on decimal column throws NPE
 ---

 Key: HIVE-7234
 URL: https://issues.apache.org/jira/browse/HIVE-7234
 Project: Hive
  Issue Type: Bug
Reporter: Ashish Kumar Singh
Assignee: Ashish Kumar Singh
 Attachments: HIVE-7234.patch


 Select on decimal column throws NPE for values greater than maximum 
 permissible value (99)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7248) UNION ALL in hive returns incorrect results on Hbase backed table

2014-06-17 Thread Swarnim Kulkarni (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Swarnim Kulkarni updated HIVE-7248:
---

Component/s: HBase Handler

 UNION ALL in hive returns incorrect results on Hbase backed table
 -

 Key: HIVE-7248
 URL: https://issues.apache.org/jira/browse/HIVE-7248
 Project: Hive
  Issue Type: Bug
  Components: HBase Handler
Affects Versions: 0.12.0, 0.13.0, 0.13.1
Reporter: Mala Chikka Kempanna

 The issue can be recreated with following steps
 1) In hbase 
 create 'TABLE_EMP','default' 
 2) On hive 
 sudo -u hive hive 
 CREATE EXTERNAL TABLE TABLE_EMP(FIRST_NAME string,LAST_NAME 
 string,CDS_UPDATED_DATE string,CDS_PK string) STORED BY 
 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH 
 SERDEPROPERTIES(hbase.columns.mapping = 
 default:FIRST_NAME,default:LAST_NAME,default:CDS_UPDATED_DATE,:key, 
 hbase.scan.cache = 500, hbase.scan.cacheblocks = false ) 
 TBLPROPERTIES(hbase.table.name = 
 TABLE_EMP,'serialization.null.format'=''); 
 3) On hbase insert the following data 
 put 'TABLE_EMP', '1', 'default:FIRST_NAME', 'Srini' 
 put 'TABLE_EMP', '1', 'default:LAST_NAME', 'P' 
 put 'TABLE_EMP', '1', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 
 put 'TABLE_EMP', '2', 'default:FIRST_NAME', 'Aravind' 
 put 'TABLE_EMP', '2', 'default:LAST_NAME', 'K' 
 put 'TABLE_EMP', '2', 'default:CDS_UPDATED_DATE', '2014-06-16 00:00:00' 
 4) On hive execute the following query 
 hive 
 SELECT * 
 FROM ( 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = '0' 
 AND CDS_PK = '9' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 UNION ALL SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = 'a' 
 AND CDS_PK = 'z' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 )t ; 
 5) Output of the query 
 1 
 1 
 2 
 2 
 6) Output of just 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = '0' 
 AND CDS_PK = '9' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 is 
 1 
 2 
 7) Output of just 
 SELECT CDS_PK 
 FROM TABLE_EMP 
 WHERE 
 CDS_PK = 'a' 
 AND CDS_PK = 'z' 
 AND CDS_UPDATED_DATE IS NOT NULL 
 Empty 
 8) UNION is used to combine the result from multiple SELECT statements into a 
 single result set. Hive currently only supports UNION ALL (bag union), in 
 which duplicates are not eliminated 
 Accordingly above query should return output 
 1 
 2 
 instead it is giving wrong output 
 1 
 1 
 2 
 2



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034566#comment-14034566
 ] 

David Chen commented on HIVE-7230:
--

[~swarnim] I was not able to generate the formatter settings in the 
.settings/org.eclipse.jdt.core.prefs files even with using code snippet you 
provided. Were you able to use that plugin to generate the correct preference 
settings?

I am not that familiar with Maven. I think if getting Maven to generate the 
correct project files becomes more involved than expected, can we go ahead and 
merge this patch? This patch was for just adding the formatter file to the 
repository.

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7240) Add classifier for avro-mapred jar

2014-06-17 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034569#comment-14034569
 ] 

Gunther Hagleitner commented on HIVE-7240:
--

+1

 Add classifier for avro-mapred jar
 --

 Key: HIVE-7240
 URL: https://issues.apache.org/jira/browse/HIVE-7240
 Project: Hive
  Issue Type: Task
  Components: File Formats
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Attachments: HIVE-7240.1.patch, HIVE-7240.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034576#comment-14034576
 ] 

Swarnim Kulkarni commented on HIVE-7230:


My motivation behind using maven was that when someone builds a project and 
imports it into eclipse, the formatter should get imported as well. If someone 
has to manually import it, it kind of again leaves an open gap for a badly 
formatted file to get merged in.

I'll give this a shot and get back to you.

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034581#comment-14034581
 ] 

David Chen commented on HIVE-7230:
--

Got it. Thanks for your help!

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Harish Butani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harish Butani updated HIVE-7159:


Attachment: HIVE-7159.8.patch

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch, 
 HIVE-7159.8.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Harish Butani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034587#comment-14034587
 ] 

Harish Butani commented on HIVE-7159:
-

[~hagleitn] issues a) and b) were because the check was being made against the 
wrong QBJoinTree.
patch.8 fixes this; I removed the null and empty checks on the nullSafes error.

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch, 
 HIVE-7159.8.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7159) For inner joins push a 'is not null predicate' to the join sources for every non nullSafe join condition

2014-06-17 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034620#comment-14034620
 ] 

Eugene Koifman commented on HIVE-7159:
--

FWIW, you can do the same with outer joins on inner side
R left outer join S on R.r=S.s is the same as R LOJ (select * from S where s is 
not null) as S on R.r=S.s
and symmetrically for ROJ.

 For inner joins push a 'is not null predicate' to the join sources for every 
 non nullSafe join condition
 

 Key: HIVE-7159
 URL: https://issues.apache.org/jira/browse/HIVE-7159
 Project: Hive
  Issue Type: Bug
Reporter: Harish Butani
Assignee: Harish Butani
 Attachments: HIVE-7159.1.patch, HIVE-7159.2.patch, HIVE-7159.3.patch, 
 HIVE-7159.4.patch, HIVE-7159.5.patch, HIVE-7159.6.patch, HIVE-7159.7.patch, 
 HIVE-7159.8.patch


 A join B on A.x = B.y
 can be transformed to
 (A where x is not null) join (B where y is not null) on A.x = B.y
 Apart from avoiding shuffling null keyed rows it also avoids issues with 
 reduce-side skew when there are a lot of null values in the data.
 Thanks to [~gopalv] for the analysis and coming up with the solution.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034628#comment-14034628
 ] 

Swarnim Kulkarni commented on HIVE-7230:


[~davidzchen] What command were you trying to run to generate the files? I used 
the same plugin and used the following command:

{noformat}
mvn clean eclipse:eclipse -pl hbase-handler -P hadoop-1
{noformat}

and my .settings/org.eclipse.jdt.core.prefs was generated[1]

[1] http://pastebin.com/zupmuMrf

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-7249) HiveTxnManager.closeTxnManger() throws if called after commitTxn()

2014-06-17 Thread Eugene Koifman (JIRA)
Eugene Koifman created HIVE-7249:


 Summary: HiveTxnManager.closeTxnManger() throws if called after 
commitTxn()
 Key: HIVE-7249
 URL: https://issues.apache.org/jira/browse/HIVE-7249
 Project: Hive
  Issue Type: Bug
  Components: Locking
Affects Versions: 0.13.1
Reporter: Eugene Koifman
Assignee: Alan Gates


 I openTxn() and acquireLocks() for a query that looks like INSERT INTO T 
PARTITION(p) SELECT * FROM T.
Then I call commitTxn().  Then I call closeTxnManger() I get an exception 
saying lock not found (the only lock in this txn).  So it seems TxnMgr doesn't 
know that commit released the locks.

Here is the stack trace and some log output which maybe useful:
2014-06-17 15:54:40,771 DEBUG mapreduce.TransactionContext 
(TransactionContext.java:onCommitJob(128)) - 
onCommitJob(job_local557130041_0001). this=46719652
2014-06-17 15:54:40,771 DEBUG lockmgr.DbTxnManager 
(DbTxnManager.java:commitTxn(205)) - Committing txn 1
2014-06-17 15:54:40,771 DEBUG txn.TxnHandler (TxnHandler.java:getDbTime(872)) - 
Going to execute query values current_timestamp
2014-06-17 15:54:40,772 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1423)) - Going to execute query select txn_state 
from TXNS where txn_id = 1 for\
 update
2014-06-17 15:54:40,773 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1438)) - Going to execute update update TXNS set 
txn_last_heartbeat = 140304568\
0772 where txn_id = 1
2014-06-17 15:54:40,778 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1440)) - Going to commit
2014-06-17 15:54:40,779 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(344)) - 
Going to execute insert insert into COMPLETED_TXN_COMPONENTS select tc_txn\
id, tc_database, tc_table, tc_partition from TXN_COMPONENTS where tc_txnid = 1
2014-06-17 15:54:40,784 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(352)) - 
Going to execute update delete from TXN_COMPONENTS where tc_txnid = 1
2014-06-17 15:54:40,788 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(356)) - 
Going to execute update delete from HIVE_LOCKS where hl_txnid = 1
2014-06-17 15:54:40,791 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(359)) - 
Going to execute update delete from TXNS where txn_id = 1
2014-06-17 15:54:40,794 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(361)) - 
Going to commit
2014-06-17 15:54:40,795 WARN  mapreduce.TransactionContext 
(TransactionContext.java:cleanup(317)) - 
cleanupJob(JobID=job_local557130041_0001)this=46719652
2014-06-17 15:54:40,795 DEBUG lockmgr.DbLockManager 
(DbLockManager.java:unlock(109)) - Unlocking id:1
2014-06-17 15:54:40,796 DEBUG txn.TxnHandler (TxnHandler.java:getDbTime(872)) - 
Going to execute query values current_timestamp
2014-06-17 15:54:40,796 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatLock(1402)) - Going to execute update update 
HIVE_LOCKS set hl_last_heartbeat = 140\
3045680796 where hl_lock_ext_id = 1
2014-06-17 15:54:40,800 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatLock(1405)) - Going to rollback
2014-06-17 15:54:40,804 ERROR metastore.RetryingHMSHandler 
(RetryingHMSHandler.java:invoke(143)) - NoSuchLockException(message:No such 
lock: 1)
at 
org.apache.hadoop.hive.metastore.txn.TxnHandler.heartbeatLock(TxnHandler.java:1407)
at 
org.apache.hadoop.hive.metastore.txn.TxnHandler.unlock(TxnHandler.java:477)
at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.unlock(HiveMetaStore.java:4817)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:105)
at com.sun.proxy.$Proxy14.unlock(Unknown Source)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.unlock(HiveMetaStoreClient.java:1598)
at 
org.apache.hadoop.hive.ql.lockmgr.DbLockManager.unlock(DbLockManager.java:110)
at 
org.apache.hadoop.hive.ql.lockmgr.DbLockManager.close(DbLockManager.java:162)
at 
org.apache.hadoop.hive.ql.lockmgr.DbTxnManager.destruct(DbTxnManager.java:300)
at 
org.apache.hadoop.hive.ql.lockmgr.HiveTxnManagerImpl.closeTxnManager(HiveTxnManagerImpl.java:39)
at 
org.apache.hadoop.hive.ql.lockmgr.DbTxnManager.closeTxnManager(DbTxnManager.java:43)
at 
org.apache.hive.hcatalog.mapreduce.TransactionContext.cleanup(TransactionContext.java:327)
at 
org.apache.hive.hcatalog.mapreduce.TransactionContext.onCommitJob(TransactionContext.java:142)
at 
org.apache.hive.hcatalog.mapreduce.OutputCommitterContainer.commitJob(OutputCommitterContainer.java:61)
at 

[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034633#comment-14034633
 ] 

David Chen commented on HIVE-7230:
--

I see. I ran the following command from the Hive Developer FAQ 
(https://cwiki.apache.org/confluence/display/Hive/HiveDeveloperFAQ#HiveDeveloperFAQ-HowdoIimportintoeclipse?):

{code}
mvn eclipse:eclipse -DdownloadSources -DdownloadJavadocs -Phadoop-2
{code}

Do I need to specify {{-pl module}} in order for the settings to be generated?

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7249) HiveTxnManager.closeTxnManger() throws if called after commitTxn()

2014-06-17 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-7249:
-

Description: 
 I openTxn() and acquireLocks() for a query that looks like INSERT INTO T 
PARTITION(p) SELECT * FROM T.
Then I call commitTxn().  Then I call closeTxnManger() I get an exception 
saying lock not found (the only lock in this txn).  So it seems TxnMgr doesn't 
know that commit released the locks.

Here is the stack trace and some log output which maybe useful:
{noformat}
2014-06-17 15:54:40,771 DEBUG mapreduce.TransactionContext 
(TransactionContext.java:onCommitJob(128)) - 
onCommitJob(job_local557130041_0001). this=46719652
2014-06-17 15:54:40,771 DEBUG lockmgr.DbTxnManager 
(DbTxnManager.java:commitTxn(205)) - Committing txn 1
2014-06-17 15:54:40,771 DEBUG txn.TxnHandler (TxnHandler.java:getDbTime(872)) - 
Going to execute query values current_timestamp
2014-06-17 15:54:40,772 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1423)) - Going to execute query select txn_state 
from TXNS where txn_id = 1 for\
 update
2014-06-17 15:54:40,773 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1438)) - Going to execute update update TXNS set 
txn_last_heartbeat = 140304568\
0772 where txn_id = 1
2014-06-17 15:54:40,778 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatTxn(1440)) - Going to commit
2014-06-17 15:54:40,779 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(344)) - 
Going to execute insert insert into COMPLETED_TXN_COMPONENTS select tc_txn\
id, tc_database, tc_table, tc_partition from TXN_COMPONENTS where tc_txnid = 1
2014-06-17 15:54:40,784 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(352)) - 
Going to execute update delete from TXN_COMPONENTS where tc_txnid = 1
2014-06-17 15:54:40,788 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(356)) - 
Going to execute update delete from HIVE_LOCKS where hl_txnid = 1
2014-06-17 15:54:40,791 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(359)) - 
Going to execute update delete from TXNS where txn_id = 1
2014-06-17 15:54:40,794 DEBUG txn.TxnHandler (TxnHandler.java:commitTxn(361)) - 
Going to commit
2014-06-17 15:54:40,795 WARN  mapreduce.TransactionContext 
(TransactionContext.java:cleanup(317)) - 
cleanupJob(JobID=job_local557130041_0001)this=46719652
2014-06-17 15:54:40,795 DEBUG lockmgr.DbLockManager 
(DbLockManager.java:unlock(109)) - Unlocking id:1
2014-06-17 15:54:40,796 DEBUG txn.TxnHandler (TxnHandler.java:getDbTime(872)) - 
Going to execute query values current_timestamp
2014-06-17 15:54:40,796 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatLock(1402)) - Going to execute update update 
HIVE_LOCKS set hl_last_heartbeat = 140\
3045680796 where hl_lock_ext_id = 1
2014-06-17 15:54:40,800 DEBUG txn.TxnHandler 
(TxnHandler.java:heartbeatLock(1405)) - Going to rollback
2014-06-17 15:54:40,804 ERROR metastore.RetryingHMSHandler 
(RetryingHMSHandler.java:invoke(143)) - NoSuchLockException(message:No such 
lock: 1)
at 
org.apache.hadoop.hive.metastore.txn.TxnHandler.heartbeatLock(TxnHandler.java:1407)
at 
org.apache.hadoop.hive.metastore.txn.TxnHandler.unlock(TxnHandler.java:477)
at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.unlock(HiveMetaStore.java:4817)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:105)
at com.sun.proxy.$Proxy14.unlock(Unknown Source)
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.unlock(HiveMetaStoreClient.java:1598)
at 
org.apache.hadoop.hive.ql.lockmgr.DbLockManager.unlock(DbLockManager.java:110)
at 
org.apache.hadoop.hive.ql.lockmgr.DbLockManager.close(DbLockManager.java:162)
at 
org.apache.hadoop.hive.ql.lockmgr.DbTxnManager.destruct(DbTxnManager.java:300)
at 
org.apache.hadoop.hive.ql.lockmgr.HiveTxnManagerImpl.closeTxnManager(HiveTxnManagerImpl.java:39)
at 
org.apache.hadoop.hive.ql.lockmgr.DbTxnManager.closeTxnManager(DbTxnManager.java:43)
at 
org.apache.hive.hcatalog.mapreduce.TransactionContext.cleanup(TransactionContext.java:327)
at 
org.apache.hive.hcatalog.mapreduce.TransactionContext.onCommitJob(TransactionContext.java:142)
at 
org.apache.hive.hcatalog.mapreduce.OutputCommitterContainer.commitJob(OutputCommitterContainer.java:61)
at 
org.apache.hive.hcatalog.mapreduce.FileOutputCommitterContainer.commitJob(FileOutputCommitterContainer.java:251)
at 
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:537)

2014-06-17 15:54:40,804 ERROR lockmgr.DbLockManager 
(DbLockManager.java:unlock(114)) - 

[jira] [Commented] (HIVE-7076) Plugin (exec hook) to log to application timeline data to Yarn

2014-06-17 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034646#comment-14034646
 ] 

Thejas M Nair commented on HIVE-7076:
-

+1

 Plugin (exec hook) to log to application timeline data to Yarn
 --

 Key: HIVE-7076
 URL: https://issues.apache.org/jira/browse/HIVE-7076
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Attachments: HIVE-7076.1.patch, HIVE-7076.2.patch, HIVE-7076.3.patch


 See: https://issues.apache.org/jira/browse/YARN-1530
 This is a simple pre/post exec hook to log query + plan information to yarn. 
 This information can be used to build tools and UIs to monitor, track, debug 
 and tune Hive queries.
 Off by default, but can be enabled via:
 hive.exec.pre.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.post.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook
 hive.exec.failure.hooks=ql.src.java.org.apache.hadoop.hive.ql.hooks.ATSHook



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034645#comment-14034645
 ] 

Szehon Ho commented on HIVE-7247:
-

The offending test is heavily using the HDFS ACL API, which is hadoop-2 only, 
so it won't be easy to fix.

All I can think of is to create a new itests sub-proj that gets compiled and 
run only in hadoop-2 profile.  It might be useful, as there could be more 
hadoop-2 features we want to test in the future.  [~hagleitn], [~brocknoland] 
[~ashutoshc], any thoughts on that?

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7247:


Description: 
Currently building itests using -Phadoop-1 profile results in following failure:

{noformat}
$cd itests
$mvn install -DskipTests -Phadoop-1
...
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
(default-testCompile) on project hive-it-unit: Compilation failure: Compilation 
failure:
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
 cannot find symbol
[ERROR] symbol : class AclStatus
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
 cannot find symbol
[ERROR] symbol : class AclEntryScope
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
 static import only from classes and interfaces
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
 cannot find symbol
[ERROR] symbol : class AclEntryType
[ERROR] location: package org.apache.hadoop.fs.permission
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
 static import only from classes and interfaces
[ERROR] 
/Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
 cannot find symbol
[ERROR] symbol : class AclEntryType
{noformat}

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho

 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7241) Wrong lock acquired for alter table rename partition

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034648#comment-14034648
 ] 

Hive QA commented on HIVE-7241:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650679/HIVE-7241.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/495/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/495/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-495/

Messages:
{noformat}
 This message was trimmed, see log for full details 
Decision can match input such as LPAREN KW_CASE KW_ARRAY using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE TinyintLiteral using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE KW_STRUCT using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_CASE SmallintLiteral using 
multiple alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:115:5: 
Decision can match input such as KW_CLUSTER KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:127:5: 
Decision can match input such as KW_PARTITION KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:138:5: 
Decision can match input such as KW_DISTRIBUTE KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:149:5: 
Decision can match input such as KW_SORT KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:166:7: 
Decision can match input such as STAR using multiple alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_STRUCT using multiple alternatives: 4, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_UNIONTYPE using multiple alternatives: 5, 
6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_ARRAY using multiple alternatives: 2, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_DATE StringLiteral using multiple 
alternatives: 2, 3

As a result, alternative(s) 3 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_FALSE using multiple alternatives: 3, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_TRUE using multiple alternatives: 3, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_NULL using multiple alternatives: 1, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_OVERWRITE using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_DISTRIBUTE 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_MAP LPAREN 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_INTO using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_LATERAL 
KW_VIEW using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_GROUP 

[jira] [Commented] (HIVE-7232) ReduceSink is emitting NULL keys due to failed keyEval

2014-06-17 Thread Navis (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7232?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034675#comment-14034675
 ] 

Navis commented on HIVE-7232:
-

Looks like something is wrong in broadcast join. I'll look into this.

 ReduceSink is emitting NULL keys due to failed keyEval
 --

 Key: HIVE-7232
 URL: https://issues.apache.org/jira/browse/HIVE-7232
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Navis
 Attachments: HIVE-7232-extra-logging.patch, q5.explain.txt, q5.sql


 After HIVE-4867 has been merged in, some queries have exhibited a very weird 
 skew towards NULL keys emitted from the ReduceSinkOperator.
 Added extra logging to print expr.column() in ExprNodeColumnEvaluator  in 
 reduce sink.
 {code}
 2014-06-14 00:37:19,186 INFO [TezChild] 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator:
 numDistributionKeys = 1 {null -- ExprNodeColumnEvaluator(_col10)}
 key_row={reducesinkkey0:442}
 {code}
 {code}
   HiveKey firstKey = toHiveKey(cachedKeys[0], tag, null);
   int distKeyLength = firstKey.getDistKeyLength();
   if(distKeyLength = 1) {
 StringBuffer x1 = new StringBuffer();
 x1.append(numDistributionKeys = + numDistributionKeys + \n);
 for (int i = 0; i  numDistributionKeys; i++) {
 x1.append(cachedKeys[0][i] +  --  + keyEval[i] + \n);
 }
 x1.append(key_row=+ SerDeUtils.getJSONString(row, 
 keyObjectInspector));
 LOG.info(GOPAL:  + x1.toString());
   }
 {code}
 The query is tpc-h query5, with extra NULL checks just to be sure.
 {code}
 ELECT n_name,
sum(l_extendedprice * (1 - l_discount)) AS revenue
 FROM customer,
  orders,
  lineitem,
  supplier,
  nation,
  region
 WHERE c_custkey = o_custkey
   AND l_orderkey = o_orderkey
   AND l_suppkey = s_suppkey
   AND c_nationkey = s_nationkey
   AND s_nationkey = n_nationkey
   AND n_regionkey = r_regionkey
   AND r_name = 'ASIA'
   AND o_orderdate = '1994-01-01'
   AND o_orderdate  '1995-01-01'
   and l_orderkey is not null
   and c_custkey is not null
   and l_suppkey is not null
   and c_nationkey is not null
   and s_nationkey is not null
   and n_regionkey is not null
 GROUP BY n_name
 ORDER BY revenue DESC;
 {code}
 The reducer which has the issue has the following plan
 {code}
 Reducer 3
 Reduce Operator Tree:
   Join Operator
 condition map:
  Inner Join 0 to 1
 condition expressions:
   0 {KEY.reducesinkkey0} {VALUE._col2}
   1 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col3}
 outputColumnNames: _col0, _col3, _col10, _col11, _col14
 Statistics: Num rows: 18344 Data size: 95229140992 Basic 
 stats: COMPLETE Column stats: NONE
 Reduce Output Operator
   key expressions: _col10 (type: int)
   sort order: +
   Map-reduce partition columns: _col10 (type: int)
   Statistics: Num rows: 18344 Data size: 95229140992 
 Basic stats: COMPLETE Column stats: NONE
   value expressions: _col0 (type: int), _col3 (type: int), 
 _col11 (type: int), _col14 (type: string)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7240) Add classifier for avro-mapred jar

2014-06-17 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-7240:
---

   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk.

 Add classifier for avro-mapred jar
 --

 Key: HIVE-7240
 URL: https://issues.apache.org/jira/browse/HIVE-7240
 Project: Hive
  Issue Type: Task
  Components: File Formats
Reporter: Ashutosh Chauhan
Assignee: Ashutosh Chauhan
 Fix For: 0.14.0

 Attachments: HIVE-7240.1.patch, HIVE-7240.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7210) NPE with No plan file found when running Driver instances on multiple threads

2014-06-17 Thread Jason Dere (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7210?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034687#comment-14034687
 ] 

Jason Dere commented on HIVE-7210:
--

I've seen several of these tests failing on previous tests. Ran MiniTez 
dynpart_sort_optimization.q, TestHCatLoader#testReadDataPrimitiveTypes, 
TestHiveServer2 locally and these pass for me. Will commit this patch shortly.

 NPE with No plan file found when running Driver instances on multiple 
 threads
 ---

 Key: HIVE-7210
 URL: https://issues.apache.org/jira/browse/HIVE-7210
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Gunther Hagleitner
 Attachments: HIVE-7210.1.patch, HIVE-7210.2.patch


 Informatica has a multithreaded application running multiple instances of 
 CLIDriver.  When running concurrent queries they sometimes hit the following 
 error:
 {noformat}
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.hive.ql.exec.Utilities: No plan file found: 
 hdfs://ICRHHW21NODE1:8020/tmp/hive-qamercury/hive_2014-05-30_10-24-57_346_890014621821056491-2/-mr-10002/6169987c-3263-4737-b5cb-38daab882afb/map.xml
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.mapreduce.JobSubmitter: Cleaning up the staging area 
 /tmp/hadoop-yarn/staging/qamercury/.staging/job_1401360353644_0078
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :ERROR 
 org.apache.hadoop.hive.ql.exec.Task: Job Submission failed with exception 
 'java.lang.NullPointerException(null)'
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
 at 
 org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:271)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:520)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:512)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
 at 
 org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:420)
 at 
 org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:136)
 at 
 org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at 
 org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1504)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1271)
 at 
 org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1089)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:912)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
 at 
 com.informatica.platform.dtm.executor.hive.impl.AbstractHiveDriverBaseImpl.run(AbstractHiveDriverBaseImpl.java:86)
 at 
 com.informatica.platform.dtm.executor.hive.MHiveDriver.executeQuery(MHiveDriver.java:126)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeQuery(HiveTaskHandlerImpl.java:358)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeScript(HiveTaskHandlerImpl.java:247)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeMainScript(HiveTaskHandlerImpl.java:194)
 at 
 

[jira] [Commented] (HIVE-7190) WebHCat launcher task failure can cause two concurent user jobs to run

2014-06-17 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7190?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034698#comment-14034698
 ] 

Ivan Mitic commented on HIVE-7190:
--

Thanks Thejas and Eugene for the review and commit!

 WebHCat launcher task failure can cause two concurent user jobs to run
 --

 Key: HIVE-7190
 URL: https://issues.apache.org/jira/browse/HIVE-7190
 Project: Hive
  Issue Type: Bug
  Components: WebHCat
Affects Versions: 0.13.0
Reporter: Ivan Mitic
Assignee: Ivan Mitic
 Fix For: 0.14.0

 Attachments: HIVE-7190.2.patch, HIVE-7190.3.patch, HIVE-7190.patch


 Templeton uses launcher jobs to launch the actual user jobs. Launcher jobs 
 are 1-map jobs (a single task jobs) which kick off the actual user job and 
 monitor it until it finishes. Given that the launcher is a task, like any 
 other MR task, it has a retry policy in case it fails (due to a task crash, 
 tasktracker/nodemanager crash, machine level outage, etc.). Further, when 
 launcher task is retried, it will again launch the same user job, *however* 
 the previous attempt user job is already running. What this means is that we 
 can have two identical user jobs running in parallel. 
 In case of MRv2, there will be an MRAppMaster and the launcher task, which 
 are subject to failure. In case any of the two fails, another instance of a 
 user job will be launched again in parallel. 
 Above situation is already a bug.
 Now going further to RM HA, what RM does on failover/restart is that it kills 
 all containers, and it restarts all applications. This means that if our 
 customer had 10 jobs on the cluster (this is 10 launcher jobs and 10 user 
 jobs), on RM failover, all 20 jobs will be restarted, and launcher jobs will 
 queue user jobs again. There are two issues with this design:
 1. There are *possible* chances for corruption of job outputs (it would be 
 useful to analyze this scenario more and confirm this statement).
 2. Cluster resources are spent on jobs redundantly
 To address the issue at least on Yarn (Hadoop 2.0) clusters, webhcat should 
 do the same thing Oozie does in this scenario, and that is to tag all its 
 child jobs with an id, and kill those jobs on task restart before they are 
 kicked off again.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HIVE-7250) Adaptive compression buffer size for wide tables in ORC

2014-06-17 Thread Prasanth J (JIRA)
Prasanth J created HIVE-7250:


 Summary: Adaptive compression buffer size for wide tables in ORC
 Key: HIVE-7250
 URL: https://issues.apache.org/jira/browse/HIVE-7250
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J


If the input table is wide (in the order of 1000s), ORC compression buffer size 
overhead becomes significant causing OOM issues. To overcome this issue, buffer 
size should be adaptively chosen based on the available memory and the number 
of columns.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HIVE-7210) NPE with No plan file found when running Driver instances on multiple threads

2014-06-17 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere reassigned HIVE-7210:


Assignee: Jason Dere  (was: Gunther Hagleitner)

 NPE with No plan file found when running Driver instances on multiple 
 threads
 ---

 Key: HIVE-7210
 URL: https://issues.apache.org/jira/browse/HIVE-7210
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Jason Dere
 Attachments: HIVE-7210.1.patch, HIVE-7210.2.patch


 Informatica has a multithreaded application running multiple instances of 
 CLIDriver.  When running concurrent queries they sometimes hit the following 
 error:
 {noformat}
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.hive.ql.exec.Utilities: No plan file found: 
 hdfs://ICRHHW21NODE1:8020/tmp/hive-qamercury/hive_2014-05-30_10-24-57_346_890014621821056491-2/-mr-10002/6169987c-3263-4737-b5cb-38daab882afb/map.xml
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.mapreduce.JobSubmitter: Cleaning up the staging area 
 /tmp/hadoop-yarn/staging/qamercury/.staging/job_1401360353644_0078
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :ERROR 
 org.apache.hadoop.hive.ql.exec.Task: Job Submission failed with exception 
 'java.lang.NullPointerException(null)'
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
 at 
 org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:271)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:520)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:512)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
 at 
 org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:420)
 at 
 org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:136)
 at 
 org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at 
 org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1504)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1271)
 at 
 org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1089)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:912)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
 at 
 com.informatica.platform.dtm.executor.hive.impl.AbstractHiveDriverBaseImpl.run(AbstractHiveDriverBaseImpl.java:86)
 at 
 com.informatica.platform.dtm.executor.hive.MHiveDriver.executeQuery(MHiveDriver.java:126)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeQuery(HiveTaskHandlerImpl.java:358)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeScript(HiveTaskHandlerImpl.java:247)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeMainScript(HiveTaskHandlerImpl.java:194)
 at 
 com.informatica.platform.ldtm.executor.common.workflow.taskhandler.impl.BaseTaskHandlerImpl.run(BaseTaskHandlerImpl.java:126)
 at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
 at java.util.concurrent.FutureTask.run(FutureTask.java:262)
 at 
 

Review Request 22711: HIVE-7247 - Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/22711/
---

Review request for hive.


Bugs: HIVE-7247
https://issues.apache.org/jira/browse/HIVE-7247


Repository: hive-git


Description
---

Create a new project 'hive-unit-hadoop2' for unit-tests that need to rely on 
hadoop-2 API's.


Diffs
-

  itests/hive-unit-hadoop2/pom.xml PRE-CREATION 
  
itests/hive-unit-hadoop2/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java
 PRE-CREATION 
  itests/hive-unit/pom.xml 0eb784b 
  
itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java
 154cec5 
  
itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java
 44a33d8 
  itests/pom.xml 11c6282 

Diff: https://reviews.apache.org/r/22711/diff/


Testing
---


Thanks,

Szehon Ho



[jira] [Updated] (HIVE-7210) NPE with No plan file found when running Driver instances on multiple threads

2014-06-17 Thread Jason Dere (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-7210:
-

   Resolution: Fixed
Fix Version/s: 0.14.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks Gunther/Vikram for the review.

 NPE with No plan file found when running Driver instances on multiple 
 threads
 ---

 Key: HIVE-7210
 URL: https://issues.apache.org/jira/browse/HIVE-7210
 Project: Hive
  Issue Type: Bug
Reporter: Jason Dere
Assignee: Jason Dere
 Fix For: 0.14.0

 Attachments: HIVE-7210.1.patch, HIVE-7210.2.patch


 Informatica has a multithreaded application running multiple instances of 
 CLIDriver.  When running concurrent queries they sometimes hit the following 
 error:
 {noformat}
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.hive.ql.exec.Utilities: No plan file found: 
 hdfs://ICRHHW21NODE1:8020/tmp/hive-qamercury/hive_2014-05-30_10-24-57_346_890014621821056491-2/-mr-10002/6169987c-3263-4737-b5cb-38daab882afb/map.xml
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :INFO 
 org.apache.hadoop.mapreduce.JobSubmitter: Cleaning up the staging area 
 /tmp/hadoop-yarn/staging/qamercury/.staging/job_1401360353644_0078
 2014-05-30 10:24:59 pool-10-thread-1 INFO: Hadoop_Native_Log :ERROR 
 org.apache.hadoop.hive.ql.exec.Task: Job Submission failed with exception 
 'java.lang.NullPointerException(null)'
 java.lang.NullPointerException
 at 
 org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
 at 
 org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:271)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:520)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:512)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
 at 
 org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
 at 
 org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
 at 
 org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:420)
 at 
 org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:136)
 at 
 org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
 at 
 org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1504)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1271)
 at 
 org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1089)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:912)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
 at 
 com.informatica.platform.dtm.executor.hive.impl.AbstractHiveDriverBaseImpl.run(AbstractHiveDriverBaseImpl.java:86)
 at 
 com.informatica.platform.dtm.executor.hive.MHiveDriver.executeQuery(MHiveDriver.java:126)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeQuery(HiveTaskHandlerImpl.java:358)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeScript(HiveTaskHandlerImpl.java:247)
 at 
 com.informatica.platform.dtm.executor.hive.task.impl.HiveTaskHandlerImpl.executeMainScript(HiveTaskHandlerImpl.java:194)
 at 
 com.informatica.platform.ldtm.executor.common.workflow.taskhandler.impl.BaseTaskHandlerImpl.run(BaseTaskHandlerImpl.java:126)
 at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
  

[jira] [Updated] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7247:


Attachment: HIVE-7247.patch

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-7247.patch


 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7247:


Status: Patch Available  (was: Open)

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-7247.patch, HIVE-7247.patch


 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7247:


Attachment: HIVE-7247.patch

Submitting a fix, based on that.

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-7247.patch, HIVE-7247.patch


 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Szehon Ho (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-7247:


Attachment: (was: HIVE-7247.patch)

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-7247.patch


 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7250) Adaptive compression buffer size for wide tables in ORC

2014-06-17 Thread Prasanth J (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7250?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth J updated HIVE-7250:
-

Attachment: HIVE-7250.1.patch

 Adaptive compression buffer size for wide tables in ORC
 ---

 Key: HIVE-7250
 URL: https://issues.apache.org/jira/browse/HIVE-7250
 Project: Hive
  Issue Type: Improvement
  Components: File Formats
Affects Versions: 0.14.0
Reporter: Prasanth J
Assignee: Prasanth J
  Labels: orcfile
 Attachments: HIVE-7250.1.patch


 If the input table is wide (in the order of 1000s), ORC compression buffer 
 size overhead becomes significant causing OOM issues. To overcome this issue, 
 buffer size should be adaptively chosen based on the available memory and the 
 number of columns.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7232) ReduceSink is emitting NULL keys due to failed keyEval

2014-06-17 Thread Navis (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7232?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034734#comment-14034734
 ] 

Navis commented on HIVE-7232:
-

I've reproduced the problem. It occurs on mapjoin + vetorization combination. 

 ReduceSink is emitting NULL keys due to failed keyEval
 --

 Key: HIVE-7232
 URL: https://issues.apache.org/jira/browse/HIVE-7232
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.14.0
Reporter: Gopal V
Assignee: Navis
 Attachments: HIVE-7232-extra-logging.patch, q5.explain.txt, q5.sql


 After HIVE-4867 has been merged in, some queries have exhibited a very weird 
 skew towards NULL keys emitted from the ReduceSinkOperator.
 Added extra logging to print expr.column() in ExprNodeColumnEvaluator  in 
 reduce sink.
 {code}
 2014-06-14 00:37:19,186 INFO [TezChild] 
 org.apache.hadoop.hive.ql.exec.ReduceSinkOperator:
 numDistributionKeys = 1 {null -- ExprNodeColumnEvaluator(_col10)}
 key_row={reducesinkkey0:442}
 {code}
 {code}
   HiveKey firstKey = toHiveKey(cachedKeys[0], tag, null);
   int distKeyLength = firstKey.getDistKeyLength();
   if(distKeyLength = 1) {
 StringBuffer x1 = new StringBuffer();
 x1.append(numDistributionKeys = + numDistributionKeys + \n);
 for (int i = 0; i  numDistributionKeys; i++) {
 x1.append(cachedKeys[0][i] +  --  + keyEval[i] + \n);
 }
 x1.append(key_row=+ SerDeUtils.getJSONString(row, 
 keyObjectInspector));
 LOG.info(GOPAL:  + x1.toString());
   }
 {code}
 The query is tpc-h query5, with extra NULL checks just to be sure.
 {code}
 ELECT n_name,
sum(l_extendedprice * (1 - l_discount)) AS revenue
 FROM customer,
  orders,
  lineitem,
  supplier,
  nation,
  region
 WHERE c_custkey = o_custkey
   AND l_orderkey = o_orderkey
   AND l_suppkey = s_suppkey
   AND c_nationkey = s_nationkey
   AND s_nationkey = n_nationkey
   AND n_regionkey = r_regionkey
   AND r_name = 'ASIA'
   AND o_orderdate = '1994-01-01'
   AND o_orderdate  '1995-01-01'
   and l_orderkey is not null
   and c_custkey is not null
   and l_suppkey is not null
   and c_nationkey is not null
   and s_nationkey is not null
   and n_regionkey is not null
 GROUP BY n_name
 ORDER BY revenue DESC;
 {code}
 The reducer which has the issue has the following plan
 {code}
 Reducer 3
 Reduce Operator Tree:
   Join Operator
 condition map:
  Inner Join 0 to 1
 condition expressions:
   0 {KEY.reducesinkkey0} {VALUE._col2}
   1 {VALUE._col0} {KEY.reducesinkkey0} {VALUE._col3}
 outputColumnNames: _col0, _col3, _col10, _col11, _col14
 Statistics: Num rows: 18344 Data size: 95229140992 Basic 
 stats: COMPLETE Column stats: NONE
 Reduce Output Operator
   key expressions: _col10 (type: int)
   sort order: +
   Map-reduce partition columns: _col10 (type: int)
   Statistics: Num rows: 18344 Data size: 95229140992 
 Basic stats: COMPLETE Column stats: NONE
   value expressions: _col0 (type: int), _col3 (type: int), 
 _col11 (type: int), _col14 (type: string)
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Chen updated HIVE-7230:
-

Attachment: HIVE-7230.2.patch

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch, HIVE-7230.2.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-7230) Add Eclipse formatter file for Hive coding conventions

2014-06-17 Thread David Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034747#comment-14034747
 ] 

David Chen commented on HIVE-7230:
--

[~swarnim] I tried following your steps but was still not able to generate my 
.settings/org.eclipse.jdt.core.prefs that included the formatter settings. It 
only included the following: http://pastebin.com/NpkFpwit

In any case, I have posted a new patch with my changes to the root pom.xml.

 Add Eclipse formatter file for Hive coding conventions
 --

 Key: HIVE-7230
 URL: https://issues.apache.org/jira/browse/HIVE-7230
 Project: Hive
  Issue Type: Improvement
Reporter: David Chen
Assignee: David Chen
 Attachments: HIVE-7230.1.patch, HIVE-7230.2.patch


 Eclipse's formatter is a convenient way to clean up formatting for Java code. 
 Currently, there is no Eclipse formatter file checked into Hive's codebase.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 22590: HIVE-7230: Add Eclipse formatter file and apply Maven Eclipse plugin in root pom.xml.

2014-06-17 Thread David Chen

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/22590/
---

(Updated June 18, 2014, 3:01 a.m.)


Review request for hive.


Changes
---

Apply Maven Eclipse plugin in root pom.xml.


Summary (updated)
-

HIVE-7230: Add Eclipse formatter file and apply Maven Eclipse plugin in root 
pom.xml.


Bugs: HIVE-7230
https://issues.apache.org/jira/browse/HIVE-7230


Repository: hive-git


Description (updated)
---

HIVE-7230: Add Eclipse formatter file and apply Maven Eclipse plugin in root 
pom.xml.


Diffs (updated)
-

  eclipse-styles.xml PRE-CREATION 
  pom.xml 2b91846c2b9511022367ceb41a216c603c8ac281 
  testutils/ptest2/pom.xml 18cd68074a226cbc38e3981886e871d5b99de756 

Diff: https://reviews.apache.org/r/22590/diff/


Testing
---

Manual


Thanks,

David Chen



[jira] [Commented] (HIVE-6561) Beeline should accept -i option to Initializing a SQL file

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6561?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034750#comment-14034750
 ] 

Hive QA commented on HIVE-6561:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650709/HIVE-6561.4.patch.txt

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 5654 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_parquet_columnar
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_root_dir_external_table
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_authorization_ctas
org.apache.hadoop.hive.ql.exec.tez.TestTezTask.testSubmit
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/496/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/496/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-496/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12650709

 Beeline should accept -i option to Initializing a SQL file
 --

 Key: HIVE-6561
 URL: https://issues.apache.org/jira/browse/HIVE-6561
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.10.0, 0.11.0, 0.12.0
Reporter: Xuefu Zhang
Assignee: Navis
 Attachments: HIVE-6561.1.patch.txt, HIVE-6561.2.patch.txt, 
 HIVE-6561.3.patch.txt, HIVE-6561.4.patch.txt


 Hive CLI has -i option. From Hive CLI help:
 {code}
 ...
  -i filenameInitialization SQL file
 ...
 {code}
 However, Beeline has no such option:
 {code}
 xzhang@xzlt:~/apa/hive3$ 
 ./packaging/target/apache-hive-0.14.0-SNAPSHOT-bin/apache-hive-0.14.0-SNAPSHOT-bin/bin/beeline
  -u jdbc:hive2:// -i hive.rc
 ...
 Connected to: Apache Hive (version 0.14.0-SNAPSHOT)
 Driver: Hive JDBC (version 0.14.0-SNAPSHOT)
 Transaction isolation: TRANSACTION_REPEATABLE_READ
 -i (No such file or directory)
 Property url is required
 Beeline version 0.14.0-SNAPSHOT by Apache Hive
 ...
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HIVE-6694) Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034753#comment-14034753
 ] 

Hive QA commented on HIVE-6694:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12650713/HIVE-6694.3.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/497/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-Build/497/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-Build-497/

Messages:
{noformat}
 This message was trimmed, see log for full details 
As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_TRUE using multiple alternatives: 3, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:261:5: 
Decision can match input such as KW_NULL using multiple alternatives: 1, 8

As a result, alternative(s) 8 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_OVERWRITE using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_DISTRIBUTE 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_MAP LPAREN 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_INTO using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_LATERAL 
KW_VIEW using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_GROUP 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as KW_BETWEEN KW_MAP LPAREN using multiple 
alternatives: 8, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_ORDER 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_SORT KW_BY 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_UNION 
KW_ALL using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:393:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_CLUSTER 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:518:5: 
Decision can match input such as {AMPERSAND..BITWISEXOR, DIV..DIVIDE, 
EQUAL..EQUAL_NS, GREATERTHAN..GREATERTHANOREQUALTO, KW_AND, KW_ARRAY, 
KW_BETWEEN..KW_BOOLEAN, KW_CASE, KW_DOUBLE, KW_FLOAT, KW_IF, KW_IN, KW_INT, 
KW_LIKE, KW_MAP, KW_NOT, KW_OR, KW_REGEXP, KW_RLIKE, KW_SMALLINT, 
KW_STRING..KW_STRUCT, KW_TINYINT, KW_UNIONTYPE, KW_WHEN, 
LESSTHAN..LESSTHANOREQUALTO, MINUS..NOTEQUAL, PLUS, STAR, TILDE} using 
multiple alternatives: 1, 3

As a result, alternative(s) 3 were disabled for that input
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.5:process (default) @ hive-exec ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hive-exec 
---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 1 resource
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-antrun-plugin:1.7:run (define-classpath) @ hive-exec ---
[INFO] Executing tasks

main:
[INFO] Executed tasks
[INFO] 
[INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ hive-exec ---
[INFO] Compiling 1698 source files to 
/data/hive-ptest/working/apache-svn-trunk-source/ql/target/classes
[INFO] -
[WARNING] COMPILATION WARNING : 
[INFO] -
[WARNING] 

[jira] [Commented] (HIVE-7247) Fix itests using hadoop-1 profile

2014-06-17 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14034778#comment-14034778
 ] 

Brock Noland commented on HIVE-7247:


+1 

 Fix itests using hadoop-1 profile 
 --

 Key: HIVE-7247
 URL: https://issues.apache.org/jira/browse/HIVE-7247
 Project: Hive
  Issue Type: Bug
  Components: Testing Infrastructure
Reporter: Szehon Ho
Assignee: Szehon Ho
 Attachments: HIVE-7247.patch


 Currently building itests using -Phadoop-1 profile results in following 
 failure:
 {noformat}
 $cd itests
 $mvn install -DskipTests -Phadoop-1
 ...
 [ERROR] Failed to execute goal 
 org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile 
 (default-testCompile) on project hive-it-unit: Compilation failure: 
 Compilation failure:
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/FolderPermissionBase.java:[31,39]
  cannot find symbol
 [ERROR] symbol : class AclStatus
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryScope
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[20,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 [ERROR] location: package org.apache.hadoop.fs.permission
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[21,1]
  static import only from classes and interfaces
 [ERROR] 
 /Users/ghagleitner/Projects/hive-test-trunk/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/TestExtendedAcls.java:[22,46]
  cannot find symbol
 [ERROR] symbol : class AclEntryType
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6694) Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6694:
--

Attachment: HIVE-6694.4.patch

Patch #3 missing new files. Patch #4 fixes that.

 Beeline should provide a way to execute shell command as Hive CLI does
 --

 Key: HIVE-6694
 URL: https://issues.apache.org/jira/browse/HIVE-6694
 Project: Hive
  Issue Type: Improvement
  Components: CLI, Clients
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.14.0

 Attachments: HIVE-6694.1.patch, HIVE-6694.1.patch, HIVE-6694.2.patch, 
 HIVE-6694.3.patch, HIVE-6694.4.patch, HIVE-6694.patch


 Hive CLI allows a user to execute a shell command using ! notation. For 
 instance, !cat myfile.txt. Being able to execute shell command may be 
 important for some users. As a replacement, however, Beeline provides no such 
 capability, possibly because ! notation is reserved for SQLLine commands. 
 It's possible to provide this using a slightly syntactic variation such as 
 !sh cat myfilie.txt.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19830: HIVE-6694: Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19830/#review46038
---



itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java
https://reviews.apache.org/r/19830/#comment81204

Okay. Will create a followup JIRA for this.


- Xuefu Zhang


On June 17, 2014, 1:49 a.m., Xuefu Zhang wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/19830/
 ---
 
 (Updated June 17, 2014, 1:49 a.m.)
 
 
 Review request for hive.
 
 
 Bugs: HIVE-6694
 https://issues.apache.org/jira/browse/HIVE-6694
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 Introduce the new requested feature.
 
 
 Diffs
 -
 
   beeline/src/java/org/apache/hive/beeline/BeeLine.java dcdd6f2 
   beeline/src/java/org/apache/hive/beeline/Commands.java 94f74d2 
   beeline/src/main/resources/BeeLine.properties 697c29a 
   cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java 3357414 
   
 itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java
  4d15e3c 
   itests/util/pom.xml b38c55b 
   itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java 2974327 
   ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 32c2fb0 
   ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java 2ce4dbd 
   ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java d2e122d 
 
 Diff: https://reviews.apache.org/r/19830/diff/
 
 
 Testing
 ---
 
 New test cases are added. 
 
 
 Thanks,
 
 Xuefu Zhang
 




[jira] [Updated] (HIVE-6694) Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6694:
--

Attachment: (was: HIVE-6694.4.patch)

 Beeline should provide a way to execute shell command as Hive CLI does
 --

 Key: HIVE-6694
 URL: https://issues.apache.org/jira/browse/HIVE-6694
 Project: Hive
  Issue Type: Improvement
  Components: CLI, Clients
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.14.0

 Attachments: HIVE-6694.1.patch, HIVE-6694.1.patch, HIVE-6694.2.patch, 
 HIVE-6694.3.patch, HIVE-6694.4.patch, HIVE-6694.patch


 Hive CLI allows a user to execute a shell command using ! notation. For 
 instance, !cat myfile.txt. Being able to execute shell command may be 
 important for some users. As a replacement, however, Beeline provides no such 
 capability, possibly because ! notation is reserved for SQLLine commands. 
 It's possible to provide this using a slightly syntactic variation such as 
 !sh cat myfilie.txt.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HIVE-6694) Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-6694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-6694:
--

Attachment: HIVE-6694.4.patch

 Beeline should provide a way to execute shell command as Hive CLI does
 --

 Key: HIVE-6694
 URL: https://issues.apache.org/jira/browse/HIVE-6694
 Project: Hive
  Issue Type: Improvement
  Components: CLI, Clients
Affects Versions: 0.11.0, 0.12.0, 0.13.0
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Fix For: 0.14.0

 Attachments: HIVE-6694.1.patch, HIVE-6694.1.patch, HIVE-6694.2.patch, 
 HIVE-6694.3.patch, HIVE-6694.4.patch, HIVE-6694.patch


 Hive CLI allows a user to execute a shell command using ! notation. For 
 instance, !cat myfile.txt. Being able to execute shell command may be 
 important for some users. As a replacement, however, Beeline provides no such 
 capability, possibly because ! notation is reserved for SQLLine commands. 
 It's possible to provide this using a slightly syntactic variation such as 
 !sh cat myfilie.txt.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: Review Request 19830: HIVE-6694: Beeline should provide a way to execute shell command as Hive CLI does

2014-06-17 Thread Xuefu Zhang

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/19830/
---

(Updated June 18, 2014, 5:34 a.m.)


Review request for hive.


Bugs: HIVE-6694
https://issues.apache.org/jira/browse/HIVE-6694


Repository: hive-git


Description
---

Introduce the new requested feature.


Diffs (updated)
-

  beeline/src/java/org/apache/hive/beeline/BeeLine.java dcdd6f2 
  beeline/src/java/org/apache/hive/beeline/Commands.java 94f74d2 
  beeline/src/main/resources/BeeLine.properties 697c29a 
  cli/src/java/org/apache/hadoop/hive/cli/CliDriver.java 3357414 
  common/src/java/org/apache/hadoop/hive/common/cli/ShellCmdExecutor.java 
PRE-CREATION 
  common/src/java/org/apache/hive/common/util/StreamPrinter.java PRE-CREATION 
  
itests/hive-unit/src/test/java/org/apache/hive/beeline/TestBeeLineWithArgs.java 
4d15e3c 
  itests/util/pom.xml b38c55b 
  itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java 2974327 
  ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 5e5cf97 
  ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java 2ce4dbd 
  ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapredLocalTask.java d2e122d 

Diff: https://reviews.apache.org/r/19830/diff/


Testing
---

New test cases are added. 


Thanks,

Xuefu Zhang