[jira] [Resolved] (HIVE-3197) Hive compile errors under Java 7 (JDBC 4.1)

2012-11-28 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-3197.


  Resolution: Fixed
Release Note: Dupe of HIVE-3384

 Hive compile errors under Java 7 (JDBC 4.1)
 ---

 Key: HIVE-3197
 URL: https://issues.apache.org/jira/browse/HIVE-3197
 Project: Hive
  Issue Type: Bug
 Environment: Ubuntu 12.04
Reporter: Brian Bloniarz

 Hi, I've been trying to compile Hive trunk from source and getting failures:
 {code}
 [javac] 
 hive-svn/jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveCallableStatement.java:48:
  error: HiveCallableStatement is not abstract and does not override abstract 
 method TgetObject(String,ClassT) in CallableStatement
 [javac] public class HiveCallableStatement implements 
 java.sql.CallableStatement {
 [javac]^
 [javac]   where T is a type-variable:
 [javac] T extends Object declared in method 
 TgetObject(String,ClassT)
 {code}
 I think this is because JDBC 4.1 is part of Java 7, and is not 
 source-compatible with older JDBC versions. Any chance you guys could add 
 JDBC 4.1 support?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3384) HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC specification

2012-11-28 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505283#comment-13505283
 ] 

Phabricator commented on HIVE-3384:
---

ashutoshc has requested changes to the revision [jira] [HIVE-3384] HIVE JDBC 
module won't compile under JDK1.7 as new methods added in JDBC specification.

  Patch looks good. Couple of minor comments.

INLINE COMMENTS
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveConnection.java:49 Unneeded 
import.
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDataSource.java:29 Duplicate 
import
  metastore/build.xml:106 Why UseSplitVerifier is required?  Unless there is a 
strong reason for this I will recommend to get rid of this, since this flag is 
specific to Oracle JVM, we should avoid vendor specific flag for better 
portability.

REVISION DETAIL
  https://reviews.facebook.net/D6873

BRANCH
  trunk

To: cdrome, weidongbian, JIRA, ashutoshc, mbautin


 HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC 
 specification
 --

 Key: HIVE-3384
 URL: https://issues.apache.org/jira/browse/HIVE-3384
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Weidong Bian
Priority: Minor
 Attachments: D6873-0.9.1.patch, D6873.1.patch, HIVE-3384.2.patch, 
 HIVE-3384.patch, HIVE-JDK7-JDBC.patch


 jdbc module couldn't be compiled with jdk7 as it adds some abstract method in 
 the JDBC specification 
 some error info:
  error: HiveCallableStatement is not abstract and does not override abstract
 method TgetObject(String,ClassT) in CallableStatement
 .
 .
 .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3748) QTestUtil should correctly find data files when running in the build directory

2012-11-28 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3748?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505286#comment-13505286
 ] 

Ashutosh Chauhan commented on HIVE-3748:


From description, I am not able to figure out in what scenarios test cases 
fail. I always run test cases from src root dir, why would you want to run 
them from build dir?

 QTestUtil should correctly find data files when running in the build directory
 --

 Key: HIVE-3748
 URL: https://issues.apache.org/jira/browse/HIVE-3748
 Project: Hive
  Issue Type: Improvement
Reporter: Mikhail Bautin
Priority: Minor
 Attachments: D7005.1.patch


 Some parts of the the TestCliDriver test suite (i.e. some jar lookups) 
 require that the current directory is set to the build directory. This makes 
 QTestUtil correctly find data files when running either in the Hive source 
 root or in the build directory.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3748) QTestUtil should correctly find data files when running in the build directory

2012-11-28 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3748?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-3748:
---

Assignee: Mikhail Bautin

 QTestUtil should correctly find data files when running in the build directory
 --

 Key: HIVE-3748
 URL: https://issues.apache.org/jira/browse/HIVE-3748
 Project: Hive
  Issue Type: Improvement
Reporter: Mikhail Bautin
Assignee: Mikhail Bautin
Priority: Minor
 Attachments: D7005.1.patch


 Some parts of the the TestCliDriver test suite (i.e. some jar lookups) 
 require that the current directory is set to the build directory. This makes 
 QTestUtil correctly find data files when running either in the Hive source 
 root or in the build directory.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3552) HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a high number of grouping set keys

2012-11-28 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3552:
-

Summary: HIVE-3552 performant manner for performing cubes/rollups/grouping 
sets for a high number of grouping set keys  (was: performant manner for 
performing cubes and rollups in case of less aggregation)

 HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a 
 high number of grouping set keys
 -

 Key: HIVE-3552
 URL: https://issues.apache.org/jira/browse/HIVE-3552
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain

 This is a follow up for HIVE-3433.
 Had a offline discussion with Sambavi - she pointed out a scenario where the
 implementation in HIVE-3433 will not scale. Assume that the user is performing
 a cube on many columns, say '8' columns. So, each row would generate 256 rows
 for the hash table, which may kill the current group by implementation.
 A better implementation would be to add an additional stage - in the first 
 stage perform the group by assuming there was no cube. Ad another stage, where
 you would perform the cube. The assumption is that the group by would have 
 decreased the output data significantly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3709) Stop storing default ConfVars in temp file

2012-11-28 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505292#comment-13505292
 ] 

Carl Steinbach commented on HIVE-3709:
--

Another option for fixing this problem is to give administrators the ability to 
specify the location where the temporary file should be written (for setuid 
processes running on POSIX systems the conventional location is somewhere under 
/var). For example we could add a configuration property named 
hive.process.local.temporary.dir and have it default to 
${user.home}/.hive/${process_id}.



 Stop storing default ConfVars in temp file
 --

 Key: HIVE-3709
 URL: https://issues.apache.org/jira/browse/HIVE-3709
 Project: Hive
  Issue Type: Improvement
  Components: Configuration
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Attachments: HIVE-3709.1.patch.txt, HIVE-3709.2.patch.txt, 
 HIVE-3709.3.patch.txt


 To work around issues with Hadoop's Configuration object, specifically it's 
 addResource(InputStream), default configurations are written to a temp file 
 (I think HIVE-2362 introduced this).
 This, however, introduces the problem that once that file is deleted from 
 /tmp the client crashes.  This is particularly problematic for long running 
 services like the metastore server.
 Writing a custom InputStream to deal with the problems in the Configuration 
 object should provide a work around, which does not introduce a time bomb 
 into Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3552) HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a high number of grouping set keys

2012-11-28 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3552:
-

Attachment: hive.3552.1.patch

 HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a 
 high number of grouping set keys
 -

 Key: HIVE-3552
 URL: https://issues.apache.org/jira/browse/HIVE-3552
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3552.1.patch


 This is a follow up for HIVE-3433.
 Had a offline discussion with Sambavi - she pointed out a scenario where the
 implementation in HIVE-3433 will not scale. Assume that the user is performing
 a cube on many columns, say '8' columns. So, each row would generate 256 rows
 for the hash table, which may kill the current group by implementation.
 A better implementation would be to add an additional stage - in the first 
 stage perform the group by assuming there was no cube. Ad another stage, where
 you would perform the cube. The assumption is that the group by would have 
 decreased the output data significantly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3552) HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a high number of grouping set keys

2012-11-28 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3552:
-

Status: Patch Available  (was: Open)

 HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a 
 high number of grouping set keys
 -

 Key: HIVE-3552
 URL: https://issues.apache.org/jira/browse/HIVE-3552
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3552.1.patch


 This is a follow up for HIVE-3433.
 Had a offline discussion with Sambavi - she pointed out a scenario where the
 implementation in HIVE-3433 will not scale. Assume that the user is performing
 a cube on many columns, say '8' columns. So, each row would generate 256 rows
 for the hash table, which may kill the current group by implementation.
 A better implementation would be to add an additional stage - in the first 
 stage perform the group by assuming there was no cube. Ad another stage, where
 you would perform the cube. The assumption is that the group by would have 
 decreased the output data significantly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3552) HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a high number of grouping set keys

2012-11-28 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3552?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3552:
-

Description: 
This is a follow up for HIVE-3433.

Had a offline discussion with Sambavi - she pointed out a scenario where the
implementation in HIVE-3433 will not scale. Assume that the user is performing
a cube on many columns, say '8' columns. So, each row would generate 256 rows
for the hash table, which may kill the current group by implementation.

A better implementation would be to add an additional mr job - in the first 
mr job perform the group by assuming there was no cube. Add another mr job, 
where
you would perform the cube. The assumption is that the group by would have 
decreased the output data significantly, and the rows would appear in the order 
of
grouping keys which has a higher probability of hitting the hash table.

  was:
This is a follow up for HIVE-3433.

Had a offline discussion with Sambavi - she pointed out a scenario where the
implementation in HIVE-3433 will not scale. Assume that the user is performing
a cube on many columns, say '8' columns. So, each row would generate 256 rows
for the hash table, which may kill the current group by implementation.

A better implementation would be to add an additional stage - in the first 
stage perform the group by assuming there was no cube. Ad another stage, where
you would perform the cube. The assumption is that the group by would have 
decreased the output data significantly.


 HIVE-3552 performant manner for performing cubes/rollups/grouping sets for a 
 high number of grouping set keys
 -

 Key: HIVE-3552
 URL: https://issues.apache.org/jira/browse/HIVE-3552
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Reporter: Namit Jain
Assignee: Namit Jain
 Attachments: hive.3552.1.patch


 This is a follow up for HIVE-3433.
 Had a offline discussion with Sambavi - she pointed out a scenario where the
 implementation in HIVE-3433 will not scale. Assume that the user is performing
 a cube on many columns, say '8' columns. So, each row would generate 256 rows
 for the hash table, which may kill the current group by implementation.
 A better implementation would be to add an additional mr job - in the first 
 mr job perform the group by assuming there was no cube. Add another mr job, 
 where
 you would perform the cube. The assumption is that the group by would have 
 decreased the output data significantly, and the rows would appear in the 
 order of
 grouping keys which has a higher probability of hitting the hash table.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-3531) Simple lock manager for dedicated hive server

2012-11-28 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach resolved HIVE-3531.
--

   Resolution: Fixed
Fix Version/s: 0.11
 Hadoop Flags: Reviewed

Committed to trunk. Thanks Navis!

 Simple lock manager for dedicated hive server
 -

 Key: HIVE-3531
 URL: https://issues.apache.org/jira/browse/HIVE-3531
 Project: Hive
  Issue Type: Improvement
  Components: Locking, Server Infrastructure
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.11

 Attachments: HIVE-3531.D5871.1.patch, HIVE-3531.D5871.2.patch, 
 HIVE-3531.D5871.3.patch


 In many cases, we uses hive server as a sole proxy for executing all the 
 queries. For that, current default lock manager based on zookeeper seemed a 
 little heavy. Simple in-memory lock manager could be enough.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3729) Error in groupSetExpression rule in Hive grammar

2012-11-28 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505312#comment-13505312
 ] 

Namit Jain commented on HIVE-3729:
--

+1



 Error in groupSetExpression rule in Hive grammar
 

 Key: HIVE-3729
 URL: https://issues.apache.org/jira/browse/HIVE-3729
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: All
Reporter: Harish Butani
Assignee: Harish Butani
Priority: Minor
 Fix For: 0.10.0

 Attachments: HIVE-3729.1.patch.txt

   Original Estimate: 5m
  Remaining Estimate: 5m

 Here is the error:
 Hive.g:1902:38: reference to rewrite element groupByExpression without 
 reference on left of -

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3729) Error in groupSetExpression rule in Hive grammar

2012-11-28 Thread Namit Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505313#comment-13505313
 ] 

Namit Jain commented on HIVE-3729:
--

https://reviews.facebook.net/D7047

 Error in groupSetExpression rule in Hive grammar
 

 Key: HIVE-3729
 URL: https://issues.apache.org/jira/browse/HIVE-3729
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: All
Reporter: Harish Butani
Assignee: Harish Butani
Priority: Minor
 Fix For: 0.10.0

 Attachments: HIVE-3729.1.patch.txt

   Original Estimate: 5m
  Remaining Estimate: 5m

 Here is the error:
 Hive.g:1902:38: reference to rewrite element groupByExpression without 
 reference on left of -

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3531) Simple lock manager for dedicated hive server

2012-11-28 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505315#comment-13505315
 ] 

Phabricator commented on HIVE-3531:
---

navis has closed the revision HIVE-3531 [jira] Simple lock manager for 
dedicated hive server.

  Closed by cws.

CHANGED PRIOR TO COMMIT
  https://reviews.facebook.net/D5871?vs=22401id=22593#differential-review-toc

REVISION DETAIL
  https://reviews.facebook.net/D5871

COMMIT
  https://reviews.facebook.net/rHIVE1414590

To: JIRA, cwsteinbach, navis


 Simple lock manager for dedicated hive server
 -

 Key: HIVE-3531
 URL: https://issues.apache.org/jira/browse/HIVE-3531
 Project: Hive
  Issue Type: Improvement
  Components: Locking, Server Infrastructure
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.11

 Attachments: HIVE-3531.D5871.1.patch, HIVE-3531.D5871.2.patch, 
 HIVE-3531.D5871.3.patch


 In many cases, we uses hive server as a sole proxy for executing all the 
 queries. For that, current default lock manager based on zookeeper seemed a 
 little heavy. Simple in-memory lock manager could be enough.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Review Request: Performance regression introduced by HIVE-3483

2012-11-28 Thread Thejas Nair

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/8246/
---

Review request for hive, Ashutosh Chauhan and namit jain.


Description
---

Solves the issue described in HIVE-3483 without adding any performance 
overhead. It does not attempt to use complete path (including scheme) for 
comparison like earlier solution. It has fix to use consistent way of 
converting path to string, before paths are compared.


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/io/CombineHiveInputFormat.java b774a95 

Diff: https://reviews.apache.org/r/8246/diff/


Testing
---


Thanks,

Thejas Nair



[jira] [Updated] (HIVE-3693) Performance regression introduced by HIVE-3483

2012-11-28 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-3693:


Assignee: Thejas M Nair
  Status: Patch Available  (was: Open)

Review board request - https://reviews.apache.org/r/8246/

 Performance regression introduced by HIVE-3483
 --

 Key: HIVE-3693
 URL: https://issues.apache.org/jira/browse/HIVE-3693
 Project: Hive
  Issue Type: Bug
Reporter: Gang Tim Liu
Assignee: Thejas M Nair
Priority: Minor
 Attachments: HIVE-3693.1.patch


 https://issues.apache.org/jira/browse/HIVE-3483 introduced a performance 
 regression in the client side during split computation.
 The client side spends a lot more time in the split computation phase. The 
 problem is checkFilterPathContains method.
 While investigating, can you create a config to disable it by default?
 thanks

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: hive 0.10 release

2012-11-28 Thread Thejas Nair

Hi Ashutosh,
Can you please include patch in 
https://issues.apache.org/jira/browse/HIVE-3693 in 0.10 if possible ?

It fixes test case failures on windows, and it is a small low risk change.

Thanks,
Thejas


On 11/19/12 3:26 PM, kulkarni.swar...@gmail.com wrote:

There are couple of enhancements that I have been working on mainly related
to the hive/hbase integration. It would be awesome if it is possible at all
to include them in this release. None of them should really be high risk. I
have patches submitted for few of them. Will try to get for others
submitted in next couple of days. Any specific deadline that I should be
looking forward to?

[1] https://issues.apache.org/jira/browse/HIVE-2599 (Patch Available)
[2] https://issues.apache.org/jira/browse/HIVE-3553 (Patch Available)
[3] https://issues.apache.org/jira/browse/HIVE-3211
[4] https://issues.apache.org/jira/browse/HIVE-3555
[5] https://issues.apache.org/jira/browse/HIVE-3725


On Mon, Nov 19, 2012 at 4:55 PM, Ashutosh Chauhan hashut...@apache.orgwrote:


Another quick update. I have created a hive-0.10 branch. At this point,
HIVE-3678 is a blocker to do a 0.10 release. There are few others nice to
have which were there in my previous email. I will be happy to merge new
patches between now and RC if folks request for it and are low risk.

Thanks,
Ashutosh
On Thu, Nov 15, 2012 at 2:29 PM, Ashutosh Chauhan hashut...@apache.org

wrote:



Good progress. Looks like folks are on board. I propose to cut the branch
in next couple of days. There are few jiras which are patch ready which I
want to get into the hive-0.10 release, including HIVE-3255 HIVE-2517
HIVE-3400 HIVE-3678
Ed has already made a request for HIVE-3083.  If folks have other patches
they want see in 0.10, please chime in.
Also, request to other committers to help in review patches. There are
quite a few in Patch Available state.

Thanks,
Ashutosh


On Thu, Nov 8, 2012 at 3:22 PM, Owen O'Malley omal...@apache.org

wrote:



+1


On Thu, Nov 8, 2012 at 3:18 PM, Carl Steinbach c...@cloudera.com

wrote:



+1

On Wed, Nov 7, 2012 at 11:23 PM, Alexander Lorenz 

wget.n...@gmail.com

wrote:



+1, good karma

On Nov 8, 2012, at 4:58 AM, Namit Jain nj...@fb.com wrote:


+1 to the idea

On 11/8/12 6:33 AM, Edward Capriolo edlinuxg...@gmail.com

wrote:



That sounds good. I think this issue needs to be solved as well

as

anything else that produces a bugus query result.

https://issues.apache.org/jira/browse/HIVE-3083

Edward

On Wed, Nov 7, 2012 at 7:50 PM, Ashutosh Chauhan 

hashut...@apache.org

wrote:

Hi,

Its been a while since we released 0.10 more than six months

ago.

All

this
while, lot of action has happened with various cool features

landing

in

trunk. Additionally, I am looking forward to HiveServer2 landing

in

trunk.  So, I propose that we cut the branch for 0.10 soon

afterwards

and
than release it. Thoughts?

Thanks,
Ashutosh




--
Alexander Alten-Lorenz
http://mapredit.blogspot.com
German Hadoop LinkedIn Group: http://goo.gl/N8pCF



















[jira] [Updated] (HIVE-3729) Error in groupSetExpression rule in Hive grammar

2012-11-28 Thread Namit Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Namit Jain updated HIVE-3729:
-

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Committed. Thanks Harish

 Error in groupSetExpression rule in Hive grammar
 

 Key: HIVE-3729
 URL: https://issues.apache.org/jira/browse/HIVE-3729
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: All
Reporter: Harish Butani
Assignee: Harish Butani
Priority: Minor
 Fix For: 0.10.0

 Attachments: HIVE-3729.1.patch.txt

   Original Estimate: 5m
  Remaining Estimate: 5m

 Here is the error:
 Hive.g:1902:38: reference to rewrite element groupByExpression without 
 reference on left of -

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3723) Hive Driver leaks ZooKeeper connections

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505338#comment-13505338
 ] 

Hudson commented on HIVE-3723:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3723 : Hive Driver leaks ZooKeeper connections (Gunther Hagleitner via 
Ashutosh Chauhan) (Revision 1414278)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414278
Files : 
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/Driver.java


 Hive Driver leaks ZooKeeper connections
 ---

 Key: HIVE-3723
 URL: https://issues.apache.org/jira/browse/HIVE-3723
 Project: Hive
  Issue Type: Bug
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
 Fix For: 0.10.0

 Attachments: HIVE-3723.1-r1411423.patch


 In certain error cases (i.e.: statement fails to compile, semantic errors) 
 the hive driver leaks zookeeper connections.
 This can be seen in the TestNegativeCliDriver test which accumulates a large 
 number of open file handles and fails if the max allowed number of file 
 handles isn't at least 2048.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3749) New test cases added by HIVE-3676 in insert1.q is not deterministic

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505339#comment-13505339
 ] 

Hudson commented on HIVE-3749:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3749 : New test cases added by HIVE-3676 in insert1.q is not 
deterministic (Navis via Ashutosh Chauhan) (Revision 1414488)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414488
Files : 
* /hive/trunk/ql/src/test/queries/clientpositive/insert1.q
* /hive/trunk/ql/src/test/results/clientpositive/insert1.q.out


 New test cases added by HIVE-3676 in insert1.q is not deterministic
 ---

 Key: HIVE-3749
 URL: https://issues.apache.org/jira/browse/HIVE-3749
 Project: Hive
  Issue Type: Test
  Components: Tests
Reporter: Navis
Assignee: Navis
 Fix For: 0.10.0

 Attachments: HIVE-3749.D7011.1.patch


 The test case inserts two row and selects those all. But the displaying order 
 can be different from env to env.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3234) getting the reporter in the recordwriter

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3234?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505340#comment-13505340
 ] 

Hudson commented on HIVE-3234:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3234 : getting the reporter in the recordwriter (Owen Omalley via 
Ashutosh Chauhan) (Revision 1414266)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414266
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/AbstractMapJoinOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/FileSinkOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/JoinUtil.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/SMBMapJoinOperator.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/exec/persistence/RowContainer.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/HiveFileFormatUtils.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/merge/BlockMergeTask.java
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/rcfile/merge/RCFileMergeMapper.java
* /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/udf
* /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/udf/Rot13InputFormat.java
* 
/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/io/udf/Rot13OutputFormat.java
* /hive/trunk/ql/src/test/queries/clientpositive/custom_input_output_format.q
* 
/hive/trunk/ql/src/test/results/clientpositive/custom_input_output_format.q.out


 getting the reporter in the recordwriter
 

 Key: HIVE-3234
 URL: https://issues.apache.org/jira/browse/HIVE-3234
 Project: Hive
  Issue Type: Improvement
  Components: Serializers/Deserializers
Affects Versions: 0.9.1
 Environment: any
Reporter: Jimmy Hu
Assignee: Owen O'Malley
  Labels: newbie
 Fix For: 0.10.0

 Attachments: HIVE-3234.D6699.1.patch, HIVE-3234.D6699.2.patch, 
 HIVE-3234.D6987.1.patch

   Original Estimate: 48h
  Remaining Estimate: 48h

 We would like to generate some custom statistics and report back to 
 map/reduce later wen implement the 
  FileSinkOperator.RecordWriter interface. However, the current interface 
 design doesn't allow us to get the map reduce reporter object. Please extend 
 the current FileSinkOperator.RecordWriter interface so that it's close() 
 method passes in a map reduce reporter object. 
 For the same reason, please also extend the RecordReader interface too to 
 include a reporter object so that users can passes in custom map reduce  
 counters.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3742) The derby metastore schema script for 0.10.0 doesn't run

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505342#comment-13505342
 ] 

Hudson commented on HIVE-3742:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3742 : The derby metastore schema script for 0.10.0 doesn't run 
(Prasad Mujumdar via Ashutosh Chauhan) (Revision 1414292)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414292
Files : 
* /hive/trunk/metastore/scripts/upgrade/derby/hive-schema-0.10.0.derby.sql


 The derby metastore schema script for 0.10.0 doesn't run
 

 Key: HIVE-3742
 URL: https://issues.apache.org/jira/browse/HIVE-3742
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.10.0
Reporter: Prasad Mujumdar
Assignee: Prasad Mujumdar
 Fix For: 0.10.0

 Attachments: HIVE-3742-2.patch, HIVE-3742.patch


 The hive-schema-0.10.0.derby.sql contains incorrect alter statement for 
 SKEWED_STRING_LIST which causes the script execution to fail

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3648) HiveMetaStoreFsImpl is not compatible with hadoop viewfs

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505341#comment-13505341
 ] 

Hudson commented on HIVE-3648:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3648 : HiveMetaStoreFsImpl is not compatible with hadoop viewfs (Arup 
Malakar via Ashutosh Chauhan) (Revision 1414353)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414353
Files : 
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreFsImpl.java
* 
/hive/trunk/shims/src/0.20/java/org/apache/hadoop/hive/shims/Hadoop20Shims.java
* 
/hive/trunk/shims/src/0.20S/java/org/apache/hadoop/hive/shims/Hadoop20SShims.java
* 
/hive/trunk/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java
* 
/hive/trunk/shims/src/common-secure/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java
* 
/hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java


 HiveMetaStoreFsImpl is not compatible with hadoop viewfs
 

 Key: HIVE-3648
 URL: https://issues.apache.org/jira/browse/HIVE-3648
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.9.0, 0.10.0
Reporter: Kihwal Lee
Assignee: Arup Malakar
 Fix For: 0.11

 Attachments: HIVE_3648_branch_0.patch, HIVE-3648-trunk-0.patch, 
 HIVE_3648_trunk_1.patch, HIVE-3648-trunk-1.patch


 HiveMetaStoreFsImpl#deleteDir() method calls Trash#moveToTrash(). This may 
 not work when viewfs is used. It needs to call Trash#moveToAppropriateTrash() 
 instead.  Please note that this method is not available in hadoop versions 
 earlier than 0.23.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3645) RCFileWriter does not implement the right function to support Federation

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505343#comment-13505343
 ] 

Hudson commented on HIVE-3645:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3645 : RCFileWriter does not implement the right function to support 
Federation (Arup Malakar via Ashutosh Chauhan) (Revision 1414286)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414286
Files : 
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/io/RCFile.java
* 
/hive/trunk/shims/src/0.20/java/org/apache/hadoop/hive/shims/Hadoop20Shims.java
* 
/hive/trunk/shims/src/0.20S/java/org/apache/hadoop/hive/shims/Hadoop20SShims.java
* 
/hive/trunk/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java
* 
/hive/trunk/shims/src/common-secure/java/org/apache/hadoop/hive/shims/HadoopShimsSecure.java
* 
/hive/trunk/shims/src/common/java/org/apache/hadoop/hive/shims/HadoopShims.java


 RCFileWriter does not implement the right function to support Federation
 

 Key: HIVE-3645
 URL: https://issues.apache.org/jira/browse/HIVE-3645
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.9.0, 0.10.0
 Environment: Hadoop 0.23.3 federation, Hive 0.9 and Pig 0.10
Reporter: Viraj Bhat
Assignee: Arup Malakar
 Fix For: 0.11

 Attachments: HIVE_3645_branch_0.patch, HIVE_3645_trunk_0.patch


 Create a table using Hive DDL
 {code}
 CREATE TABLE tmp_hcat_federated_numbers_part_1 (
   id   int,  
   intnum   int,
   floatnum float
 )partitioned by (
   part1string,
   part2string
 )
 STORED AS rcfile
 LOCATION 'viewfs:///database/tmp_hcat_federated_numbers_part_1';
 {code}
 Populate it using Pig:
 {code}
 A = load 'default.numbers_pig' using org.apache.hcatalog.pig.HCatLoader();
 B = filter A by id =  500;
 C = foreach B generate (int)id, (int)intnum, (float)floatnum;
 store C into
 'default.tmp_hcat_federated_numbers_part_1'
 using org.apache.hcatalog.pig.HCatStorer
('part1=pig, part2=hcat_pig_insert',
 'id: int,intnum: int,floatnum: float');
 {code}
 Generates the following error when running on a Federated Cluster:
 {quote}
 2012-10-29 20:40:25,011 [main] ERROR
 org.apache.pig.tools.pigstats.SimplePigStats - ERROR 2997: Unable to recreate
 exception from backed error: AttemptID:attempt_1348522594824_0846_m_00_3
 Info:Error: org.apache.hadoop.fs.viewfs.NotInMountpointException:
 getDefaultReplication on empty path is invalid
 at
 org.apache.hadoop.fs.viewfs.ViewFileSystem.getDefaultReplication(ViewFileSystem.java:479)
 at org.apache.hadoop.hive.ql.io.RCFile$Writer.init(RCFile.java:723)
 at org.apache.hadoop.hive.ql.io.RCFile$Writer.init(RCFile.java:705)
 at
 org.apache.hadoop.hive.ql.io.RCFileOutputFormat.getRecordWriter(RCFileOutputFormat.java:86)
 at
 org.apache.hcatalog.mapreduce.FileOutputFormatContainer.getRecordWriter(FileOutputFormatContainer.java:100)
 at
 org.apache.hcatalog.mapreduce.HCatOutputFormat.getRecordWriter(HCatOutputFormat.java:228)
 at
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:84)
 at
 org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.init(MapTask.java:587)
 at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:706)
 at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
 at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:157)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:396)
 at
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1212)
 at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:152)
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3676) INSERT INTO regression caused by HIVE-3465

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3676?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505344#comment-13505344
 ] 

Hudson commented on HIVE-3676:
--

Integrated in Hive-trunk-h0.21 #1821 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1821/])
HIVE-3749 : New test cases added by HIVE-3676 in insert1.q is not 
deterministic (Navis via Ashutosh Chauhan) (Revision 1414488)

 Result = FAILURE
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414488
Files : 
* /hive/trunk/ql/src/test/queries/clientpositive/insert1.q
* /hive/trunk/ql/src/test/results/clientpositive/insert1.q.out


 INSERT INTO regression caused by HIVE-3465
 --

 Key: HIVE-3676
 URL: https://issues.apache.org/jira/browse/HIVE-3676
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Carl Steinbach
Assignee: Navis
 Fix For: 0.10.0

 Attachments: HIVE-3676.D6741.1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 1821 - Still Failing

2012-11-28 Thread Apache Jenkins Server
Changes for Build #1777
[kevinwilfong] HIVE-3524. Storing certain Exception objects thrown in 
HiveMetaStore.java in MetaStoreEndFunctionContext. (Maheshwaran Srinivasan via 
kevinwilfong)

[cws] HIVE-1977. DESCRIBE TABLE syntax doesn't support specifying a database 
qualified table name (Zhenxiao Luo via cws)

[cws] HIVE-3674. Test case TestParse broken after recent checkin (Sambavi 
Muthukrishnan via cws)


Changes for Build #1778
[cws] HIVE-1362. Column level scalar valued statistics on Tables and Partitions 
(Shreepadma Venugopalan via cws)


Changes for Build #1779

Changes for Build #1780
[kevinwilfong] HIVE-3686. Fix compile errors introduced by the interaction of 
HIVE-1362 and HIVE-3524. (Shreepadma Venugopalan via kevinwilfong)


Changes for Build #1781
[namit] HIVE-3687 smb_mapjoin_13.q is nondeterministic
(Kevin Wilfong via namit)


Changes for Build #1782
[hashutosh] HIVE-2715: Upgrade Thrift dependency to 0.9.0 (Ashutosh Chauhan)


Changes for Build #1783
[kevinwilfong] HIVE-3654. block relative path access in hive. (njain via 
kevinwilfong)

[hashutosh] HIVE-3658 : Unable to generate the Hbase related unit tests using 
velocity templates on Windows (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3661 : Remove the Windows specific = related swizzle path 
changes from Proxy FileSystems (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3480 : Resource leak: Fix the file handle leaks in Symbolic 
 Symlink related input formats. (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1784
[kevinwilfong] HIVE-3675. NaN does not work correctly for round(n). (njain via 
kevinwilfong)

[cws] HIVE-3651. bucketmapjoin?.q tests fail with hadoop 0.23 (Prasad Mujumdar 
via cws)


Changes for Build #1785
[namit] HIVE-3613 Implement grouping_id function
(Ian Gorbachev via namit)

[namit] HIVE-3692 Update parallel test documentation
(Ivan Gorbachev via namit)

[namit] HIVE-3649 Hive List Bucketing - enhance DDL to specify list bucketing 
table
(Gang Tim Liu via namit)


Changes for Build #1786
[namit] HIVE-3696 Revert HIVE-3483 which causes performance regression
(Gang Tim Liu via namit)


Changes for Build #1787
[kevinwilfong] HIVE-3621. Make prompt in Hive CLI configurable. (Jingwei Lu via 
kevinwilfong)

[kevinwilfong] HIVE-3695. TestParse breaks due to HIVE-3675. (njain via 
kevinwilfong)


Changes for Build #1788
[kevinwilfong] HIVE-3557. Access to external URLs in hivetest.py. (Ivan 
Gorbachev via kevinwilfong)


Changes for Build #1789
[hashutosh] HIVE-3662 : TestHiveServer: testScratchDirShouldClearWhileStartup 
is failing on Windows (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3659 : TestHiveHistory::testQueryloglocParentDirNotExist Test 
fails on Windows because of some resource leaks in ZK (Kanna Karanam via 
Ashutosh Chauhan)

[hashutosh] HIVE-3663 Unable to display the MR Job file path on Windows in case 
of MR job failures.  (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1790

Changes for Build #1791

Changes for Build #1792

Changes for Build #1793
[hashutosh] HIVE-3704 : name of some metastore scripts are not per convention 
(Ashutosh Chauhan)


Changes for Build #1794
[hashutosh] HIVE-3243 : ignore white space between entries of hive/hbase table 
mapping (Shengsheng Huang via Ashutosh Chauhan)

[hashutosh] HIVE-3215 : JobDebugger should use RunningJob.getTrackingURL 
(Bhushan Mandhani via Ashutosh Chauhan)


Changes for Build #1795
[cws] HIVE-3437. 0.23 compatibility: fix unit tests when building against 0.23 
(Chris Drome via cws)

[hashutosh] HIVE-3626 : RetryingHMSHandler should wrap JDOException inside 
MetaException (Bhushan Mandhani via Ashutosh Chauhan)

[hashutosh] HIVE-3560 : Hive always prints a warning message when using remote 
metastore (Travis Crawford via Ashutosh Chauhan)


Changes for Build #1796

Changes for Build #1797
[hashutosh] HIVE-3664 : Avoid to create a symlink for hive-contrib.jar file in 
dist\lib folder. (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1798
[namit] HIVE-3706 getBoolVar in FileSinkOperator can be optimized
(Kevin Wilfong via namit)

[namit] HIVE-3707 Round map/reduce progress down when it is in the range [99.5, 
100)
(Kevin Wilfong via namit)

[namit] HIVE-3471 Implement grouping sets in hive
(Ivan Gorbachev via namit)


Changes for Build #1799
[hashutosh] HIVE-3291 : fix fs resolvers (Ashish Singh via Ashutosh Chauhan)

[hashutosh] HIVE-3680 : Include Table information in Hive's AddPartitionEvent. 
(Mithun Radhakrishnan via Ashutosh Chauhan)


Changes for Build #1800
[hashutosh] HIVE-3520 : ivysettings.xml does not let you override 
.m2/repository (Raja Aluri via Ashutosh Chauhan)

[hashutosh] HIVE-3435 : Get pdk pluginTest passed when triggered from both 
builtin tests and pdk tests on hadoop23  (Zhenxiao Luo via Ashutosh Chauhan)


Changes for Build #1801

Changes for Build #1802

Changes for Build #1803

Changes for Build #1804

Changes for Build #1805

Changes for Build #1806
[hashutosh] Preparing 

[jira] [Commented] (HIVE-3729) Error in groupSetExpression rule in Hive grammar

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505432#comment-13505432
 ] 

Hudson commented on HIVE-3729:
--

Integrated in Hive-trunk-h0.21 #1822 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1822/])
HIVE-3729 Error in groupSetExpression rule in Hive grammar
(Harish Butani via namit) (Revision 1414608)

 Result = ABORTED
namit : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414608
Files : 
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/parse/Hive.g


 Error in groupSetExpression rule in Hive grammar
 

 Key: HIVE-3729
 URL: https://issues.apache.org/jira/browse/HIVE-3729
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.10.0
 Environment: All
Reporter: Harish Butani
Assignee: Harish Butani
Priority: Minor
 Fix For: 0.10.0

 Attachments: HIVE-3729.1.patch.txt

   Original Estimate: 5m
  Remaining Estimate: 5m

 Here is the error:
 Hive.g:1902:38: reference to rewrite element groupByExpression without 
 reference on left of -

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3531) Simple lock manager for dedicated hive server

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505431#comment-13505431
 ] 

Hudson commented on HIVE-3531:
--

Integrated in Hive-trunk-h0.21 #1822 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1822/])
HIVE-3531 [jira] Simple lock manager for dedicated hive server
(Navis Ryu via Carl Steinbach)

Summary:
DPAL-1906 Implement simple lock manager for hive server

In many cases, we uses hive server as a sole proxy for executing all the 
queries. For that, current default lock manager based on zookeeper seemed a 
little heavy. Simple in-memory lock manager could be enough.

Test Plan: TestDedicatedLockManager

Reviewers: JIRA, cwsteinbach

Reviewed By: cwsteinbach

Differential Revision: https://reviews.facebook.net/D5871 (Revision 1414590)

 Result = ABORTED
cws : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414590
Files : 
* 
/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/EmbeddedLockManager.java
* /hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/lockmgr/HiveLockObject.java
* /hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/lockmgr
* 
/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestEmbeddedLockManager.java


 Simple lock manager for dedicated hive server
 -

 Key: HIVE-3531
 URL: https://issues.apache.org/jira/browse/HIVE-3531
 Project: Hive
  Issue Type: Improvement
  Components: Locking, Server Infrastructure
Reporter: Navis
Assignee: Navis
Priority: Trivial
 Fix For: 0.11

 Attachments: HIVE-3531.D5871.1.patch, HIVE-3531.D5871.2.patch, 
 HIVE-3531.D5871.3.patch


 In many cases, we uses hive server as a sole proxy for executing all the 
 queries. For that, current default lock manager based on zookeeper seemed a 
 little heavy. Simple in-memory lock manager could be enough.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3665) Allow URIs without port to be specified in metatool

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3665?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505433#comment-13505433
 ] 

Hudson commented on HIVE-3665:
--

Integrated in Hive-trunk-h0.21 #1822 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1822/])
HIVE-3665: Allow URIs without port to be specified in metatool (Shreepadma 
via Ashutosh Chauhan) (Revision 1414512)

 Result = ABORTED
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414512
Files : 
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/tools/HiveMetaTool.java


 Allow URIs without port to be specified in metatool
 ---

 Key: HIVE-3665
 URL: https://issues.apache.org/jira/browse/HIVE-3665
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Shreepadma Venugopalan
Assignee: Shreepadma Venugopalan
 Fix For: 0.11

 Attachments: HIVE-3665.1.patch.txt


 Metatool should accept input URIs where one URI contains a port and the other 
 doesn't. While metatool today accepts input URIs without the port when both 
 the input URIs (oldLoc and newLoc) don't contain the port, we should make the 
 tool a little more flexible to allow for the case where one URI contains a 
 valid port and the other input URI doesn't. This makes more sense when 
 transitioning to HA and a user chooses to specify the port as part of the 
 oldLoc, but the port doesn't mean much for the newLoc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3678) Add metastore upgrade scripts for column stats schema changes

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3678?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505430#comment-13505430
 ] 

Hudson commented on HIVE-3678:
--

Integrated in Hive-trunk-h0.21 #1822 (See 
[https://builds.apache.org/job/Hive-trunk-h0.21/1822/])
HIVE-3678 : Add metastore upgrade scripts for column stats schema changes 
(Shreepadma via Ashutosh Chauhan) (Revision 1414509)

 Result = ABORTED
hashutosh : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1414509
Files : 
* /hive/trunk/metastore/scripts/upgrade/derby/012-HIVE-1362.derby.sql
* /hive/trunk/metastore/scripts/upgrade/derby/hive-schema-0.10.0.derby.sql
* /hive/trunk/metastore/scripts/upgrade/derby/upgrade-0.9.0-to-0.10.0.derby.sql
* /hive/trunk/metastore/scripts/upgrade/mysql/012-HIVE-1362.mysql.sql
* /hive/trunk/metastore/scripts/upgrade/mysql/hive-schema-0.10.0.mysql.sql
* /hive/trunk/metastore/scripts/upgrade/mysql/upgrade-0.9.0-to-0.10.0.mysql.sql
* /hive/trunk/metastore/scripts/upgrade/oracle/012-HIVE-1362.oracle.sql
* /hive/trunk/metastore/scripts/upgrade/oracle/hive-schema-0.10.0.oracle.sql
* /hive/trunk/metastore/scripts/upgrade/postgres/012-HIVE-1362.postgres.sql
* /hive/trunk/metastore/scripts/upgrade/postgres/hive-schema-0.10.0.postgres.sql
* 
/hive/trunk/metastore/scripts/upgrade/postgres/upgrade-0.9.0-to-0.10.0.postgres.sql
* 
/hive/trunk/metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java
* 
/hive/trunk/metastore/src/model/org/apache/hadoop/hive/metastore/model/MPartitionColumnStatistics.java
* 
/hive/trunk/metastore/src/model/org/apache/hadoop/hive/metastore/model/MTableColumnStatistics.java
* /hive/trunk/metastore/src/model/package.jdo


 Add metastore upgrade scripts for column stats schema changes
 -

 Key: HIVE-3678
 URL: https://issues.apache.org/jira/browse/HIVE-3678
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Reporter: Shreepadma Venugopalan
Assignee: Shreepadma Venugopalan
 Fix For: 0.10.0

 Attachments: HIVE-3678.1.patch.txt, HIVE-3678.2.patch.txt, 
 HIVE-3678.3.patch.txt, HIVE-3678.4.patch.txt


 Add upgrade script for column statistics schema changes for 
 Postgres/MySQL/Oracle/Derby

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3709) Stop storing default ConfVars in temp file

2012-11-28 Thread Chris McConnell (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505455#comment-13505455
 ] 

Chris McConnell commented on HIVE-3709:
---

I was also looking into this with 3596, I was able to fix utilizing a location 
similar to the suggestion Carl made above, however I think that pushes the 
problem to another location, rather than addressing the actual issue. I like 
where Kevin is going with this fix, I had thought about the possibility of 
checking the confVarURL in the copy constructor, removing and re-creating if it 
did not exist, but even that would not be perfect depending upon timing. 

 Stop storing default ConfVars in temp file
 --

 Key: HIVE-3709
 URL: https://issues.apache.org/jira/browse/HIVE-3709
 Project: Hive
  Issue Type: Improvement
  Components: Configuration
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Attachments: HIVE-3709.1.patch.txt, HIVE-3709.2.patch.txt, 
 HIVE-3709.3.patch.txt


 To work around issues with Hadoop's Configuration object, specifically it's 
 addResource(InputStream), default configurations are written to a temp file 
 (I think HIVE-2362 introduced this).
 This, however, introduces the problem that once that file is deleted from 
 /tmp the client crashes.  This is particularly problematic for long running 
 services like the metastore server.
 Writing a custom InputStream to deal with the problems in the Configuration 
 object should provide a work around, which does not introduce a time bomb 
 into Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2206) add a new optimizer for query correlation discovery and optimization

2012-11-28 Thread Yin Huai (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2206?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505495#comment-13505495
 ] 

Yin Huai commented on HIVE-2206:


[~cwsteinbach] I am not sure if unit tests in Hive are comprehensive enough. If 
not, it might be better that we turn on this optimizer by default in future 
after we can use more queries to test it.

I just tested all unit tests with an enabled correlation optimizer. Because, if 
map side aggregation is on, correlation optimizer also requires regular reduce 
side aggregation to be generated, if cube or rollup is used in the query, 
error message 10209 
(org.apache.hadoop.hive.ql.ErrorMsg.HIVE_GROUPING_SETS_AGGR_NOMAPAGGR) will be 
thrown. Seems HIVE-3508 can solve this issue. Except this issue, a few query 
plans need to be re-generated because of changing operator ids.

This jira has taken a long time. Can we wrap it up and I will start to work on 
follow-up jiras.

 add a new optimizer for query correlation discovery and optimization
 

 Key: HIVE-2206
 URL: https://issues.apache.org/jira/browse/HIVE-2206
 Project: Hive
  Issue Type: New Feature
  Components: Query Processor
Affects Versions: 0.10.0
Reporter: He Yongqiang
Assignee: Yin Huai
 Attachments: HIVE-2206.10-r1384442.patch.txt, 
 HIVE-2206.11-r1385084.patch.txt, HIVE-2206.12-r1386996.patch.txt, 
 HIVE-2206.13-r1389072.patch.txt, HIVE-2206.14-r1389704.patch.txt, 
 HIVE-2206.15-r1392491.patch.txt, HIVE-2206.16-r1399936.patch.txt, 
 HIVE-2206.17-r1404933.patch.txt, HIVE-2206.18-r1407720.patch.txt, 
 HIVE-2206.19-r1410581.patch.txt, HIVE-2206.1.patch.txt, 
 HIVE-2206.2.patch.txt, HIVE-2206.3.patch.txt, HIVE-2206.4.patch.txt, 
 HIVE-2206.5-1.patch.txt, HIVE-2206.5.patch.txt, HIVE-2206.6.patch.txt, 
 HIVE-2206.7.patch.txt, HIVE-2206.8.r1224646.patch.txt, 
 HIVE-2206.8-r1237253.patch.txt, testQueries.2.q, YSmartPatchForHive.patch


 This issue proposes a new logical optimizer called Correlation Optimizer, 
 which is used to merge correlated MapReduce jobs (MR jobs) into a single MR 
 job. The idea is based on YSmart (http://ysmart.cse.ohio-state.edu/).The 
 paper and slides of YSmart are linked at the bottom.
 Since Hive translates queries in a sentence by sentence fashion, for every 
 operation which may need to shuffle the data (e.g. join and aggregation 
 operations), Hive will generate a MapReduce job for that operation. However, 
 for those operations which may need to shuffle the data, they may involve 
 correlations explained below and thus can be executed in a single MR job.
 # Input Correlation: Multiple MR jobs have input correlation (IC) if their 
 input relation sets are not disjoint;
 # Transit Correlation: Multiple MR jobs have transit correlation (TC) if they 
 have not only input correlation, but also the same partition key;
 # Job Flow Correlation: An MR has job flow correlation (JFC) with one of its 
 child nodes if it has the same partition key as that child node.
 The current implementation of correlation optimizer only detect correlations 
 among MR jobs for reduce-side join operators and reduce-side aggregation 
 operators (not map only aggregation). A query will be optimized if it 
 satisfies following conditions.
 # There exists a MR job for reduce-side join operator or reduce side 
 aggregation operator which have JFC with all of its parents MR jobs (TCs will 
 be also exploited if JFC exists);
 # All input tables of those correlated MR job are original input tables (not 
 intermediate tables generated by sub-queries); and 
 # No self join is involved in those correlated MR jobs.
 Correlation optimizer is implemented as a logical optimizer. The main reasons 
 are that it only needs to manipulate the query plan tree and it can leverage 
 the existing component on generating MR jobs.
 Current implementation can serve as a framework for correlation related 
 optimizations. I think that it is better than adding individual optimizers. 
 There are several work that can be done in future to improve this optimizer. 
 Here are three examples.
 # Support queries only involve TC;
 # Support queries in which input tables of correlated MR jobs involves 
 intermediate tables; and 
 # Optimize queries involving self join. 
 References:
 Paper and presentation of YSmart.
 Paper: 
 http://www.cse.ohio-state.edu/hpcs/WWW/HTML/publications/papers/TR-11-7.pdf
 Slides: http://sdrv.ms/UpwJJc

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: hive 0.10 release

2012-11-28 Thread Bing Li
When will hive 0.10 be released?

2012/11/28 Thejas Nair the...@hortonworks.com

 Hi Ashutosh,
 Can you please include patch in https://issues.apache.org/**
 jira/browse/HIVE-3693 https://issues.apache.org/jira/browse/HIVE-3693in 
 0.10 if possible ?
 It fixes test case failures on windows, and it is a small low risk change.

 Thanks,
 Thejas



 On 11/19/12 3:26 PM, kulkarni.swar...@gmail.com wrote:

 There are couple of enhancements that I have been working on mainly
 related
 to the hive/hbase integration. It would be awesome if it is possible at
 all
 to include them in this release. None of them should really be high risk.
 I
 have patches submitted for few of them. Will try to get for others
 submitted in next couple of days. Any specific deadline that I should be
 looking forward to?

 [1] 
 https://issues.apache.org/**jira/browse/HIVE-2599https://issues.apache.org/jira/browse/HIVE-2599(Patch
  Available)
 [2] 
 https://issues.apache.org/**jira/browse/HIVE-3553https://issues.apache.org/jira/browse/HIVE-3553(Patch
  Available)
 [3] 
 https://issues.apache.org/**jira/browse/HIVE-3211https://issues.apache.org/jira/browse/HIVE-3211
 [4] 
 https://issues.apache.org/**jira/browse/HIVE-3555https://issues.apache.org/jira/browse/HIVE-3555
 [5] 
 https://issues.apache.org/**jira/browse/HIVE-3725https://issues.apache.org/jira/browse/HIVE-3725


 On Mon, Nov 19, 2012 at 4:55 PM, Ashutosh Chauhan hashut...@apache.org
 wrote:

  Another quick update. I have created a hive-0.10 branch. At this point,
 HIVE-3678 is a blocker to do a 0.10 release. There are few others nice to
 have which were there in my previous email. I will be happy to merge new
 patches between now and RC if folks request for it and are low risk.

 Thanks,
 Ashutosh
 On Thu, Nov 15, 2012 at 2:29 PM, Ashutosh Chauhan hashut...@apache.org

 wrote:


  Good progress. Looks like folks are on board. I propose to cut the
 branch
 in next couple of days. There are few jiras which are patch ready which
 I
 want to get into the hive-0.10 release, including HIVE-3255 HIVE-2517
 HIVE-3400 HIVE-3678
 Ed has already made a request for HIVE-3083.  If folks have other
 patches
 they want see in 0.10, please chime in.
 Also, request to other committers to help in review patches. There are
 quite a few in Patch Available state.

 Thanks,
 Ashutosh


 On Thu, Nov 8, 2012 at 3:22 PM, Owen O'Malley omal...@apache.org

 wrote:


  +1


 On Thu, Nov 8, 2012 at 3:18 PM, Carl Steinbach c...@cloudera.com

 wrote:


  +1

 On Wed, Nov 7, 2012 at 11:23 PM, Alexander Lorenz 

 wget.n...@gmail.com

 wrote:


  +1, good karma

 On Nov 8, 2012, at 4:58 AM, Namit Jain nj...@fb.com wrote:

  +1 to the idea

 On 11/8/12 6:33 AM, Edward Capriolo edlinuxg...@gmail.com

 wrote:


  That sounds good. I think this issue needs to be solved as well

 as

 anything else that produces a bugus query result.

 https://issues.apache.org/**jira/browse/HIVE-3083https://issues.apache.org/jira/browse/HIVE-3083

 Edward

 On Wed, Nov 7, 2012 at 7:50 PM, Ashutosh Chauhan 

 hashut...@apache.org

 wrote:

 Hi,

 Its been a while since we released 0.10 more than six months

 ago.

 All

 this
 while, lot of action has happened with various cool features

 landing

 in

 trunk. Additionally, I am looking forward to HiveServer2 landing

 in

 trunk.  So, I propose that we cut the branch for 0.10 soon

 afterwards

 and
 than release it. Thoughts?

 Thanks,
 Ashutosh



 --
 Alexander Alten-Lorenz
 http://mapredit.blogspot.com
 German Hadoop LinkedIn Group: http://goo.gl/N8pCF














[jira] [Created] (HIVE-3751) Handle enclosing characters in delimited fields

2012-11-28 Thread Mark Grover (JIRA)
Mark Grover created HIVE-3751:
-

 Summary: Handle enclosing characters in delimited fields
 Key: HIVE-3751
 URL: https://issues.apache.org/jira/browse/HIVE-3751
 Project: Hive
  Issue Type: New Feature
  Components: Serializers/Deserializers
Affects Versions: 0.9.0
Reporter: Mark Grover


People have requested in the past to have Hive support enclosing characters in 
delimited row format.

An example of such data:
{code}
1, 123456,  , Charlotte, NC 
{code}

The format of the create table statement may look like:
{{DELIMITED [FIELDS TERMINATED BY char] [COLLECTION ITEMS TERMINATED BY char] 
[FIELDS ENCLOSED BY char]}}

The first pass solution to this JIRA may disallow inclusion of enclosing 
characters within the data. A latter improvement could be to allow escaping of 
enclosing characters within the data.

Reference mailing list thread:
http://mail-archives.apache.org/mod_mbox/hive-user/201204.mbox/%3CCAENxBwwrZrqBSJXtJHpqc_FfcZvwRMoaT9W7dR=jgtyjoxp...@mail.gmail.com%3E

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3751) Handle enclosing characters in delimited fields

2012-11-28 Thread Mark Grover (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3751?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Grover updated HIVE-3751:
--

Description: 
People have requested in the past to have Hive support enclosing characters in 
delimited row format.

An example of such data:
{code}
1, 123456,  , Charlotte, NC 
{code}

The format of the create table statement may look like:
{code}
DELIMITED [FIELDS TERMINATED BY char] [COLLECTION ITEMS TERMINATED BY char] 
[FIELDS ENCLOSED BY char]
{code}

The first pass solution to this JIRA may disallow inclusion of enclosing 
characters within the data. A latter improvement could be to allow escaping of 
enclosing characters within the data.

Reference mailing list thread:
http://mail-archives.apache.org/mod_mbox/hive-user/201204.mbox/%3CCAENxBwwrZrqBSJXtJHpqc_FfcZvwRMoaT9W7dR=jgtyjoxp...@mail.gmail.com%3E

  was:
People have requested in the past to have Hive support enclosing characters in 
delimited row format.

An example of such data:
{code}
1, 123456,  , Charlotte, NC 
{code}

The format of the create table statement may look like:
{{DELIMITED [FIELDS TERMINATED BY char] [COLLECTION ITEMS TERMINATED BY char] 
[FIELDS ENCLOSED BY char]}}

The first pass solution to this JIRA may disallow inclusion of enclosing 
characters within the data. A latter improvement could be to allow escaping of 
enclosing characters within the data.

Reference mailing list thread:
http://mail-archives.apache.org/mod_mbox/hive-user/201204.mbox/%3CCAENxBwwrZrqBSJXtJHpqc_FfcZvwRMoaT9W7dR=jgtyjoxp...@mail.gmail.com%3E


 Handle enclosing characters in delimited fields
 ---

 Key: HIVE-3751
 URL: https://issues.apache.org/jira/browse/HIVE-3751
 Project: Hive
  Issue Type: New Feature
  Components: Serializers/Deserializers
Affects Versions: 0.9.0
Reporter: Mark Grover

 People have requested in the past to have Hive support enclosing characters 
 in delimited row format.
 An example of such data:
 {code}
 1, 123456,  , Charlotte, NC 
 {code}
 The format of the create table statement may look like:
 {code}
 DELIMITED [FIELDS TERMINATED BY char] [COLLECTION ITEMS TERMINATED BY char] 
 [FIELDS ENCLOSED BY char]
 {code}
 The first pass solution to this JIRA may disallow inclusion of enclosing 
 characters within the data. A latter improvement could be to allow escaping 
 of enclosing characters within the data.
 Reference mailing list thread:
 http://mail-archives.apache.org/mod_mbox/hive-user/201204.mbox/%3CCAENxBwwrZrqBSJXtJHpqc_FfcZvwRMoaT9W7dR=jgtyjoxp...@mail.gmail.com%3E

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Build failed in Jenkins: Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false #213

2012-11-28 Thread Apache Jenkins Server
See 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/

--
[...truncated 9912 lines...]

compile-test:
 [echo] Project: serde
[javac] Compiling 26 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/serde/test/classes
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.

create-dirs:
 [echo] Project: service
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/service/src/test/resources
 does not exist.

init:
 [echo] Project: service

ivy-init-settings:
 [echo] Project: service

ivy-resolve:
 [echo] Project: service
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-service-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/ivy/report/org.apache.hive-hive-service-default.html

ivy-retrieve:
 [echo] Project: service

compile:
 [echo] Project: service

ivy-resolve-test:
 [echo] Project: service

ivy-retrieve-test:
 [echo] Project: service

compile-test:
 [echo] Project: service
[javac] Compiling 2 source files to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/service/test/classes

test:
 [echo] Project: hive

test-shims:
 [echo] Project: hive

test-conditions:
 [echo] Project: shims

gen-test:
 [echo] Project: shims

create-dirs:
 [echo] Project: shims
 [copy] Warning: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/test/resources
 does not exist.

init:
 [echo] Project: shims

ivy-init-settings:
 [echo] Project: shims

ivy-resolve:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml
[ivy:report] Processing 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/ivy/resolution-cache/org.apache.hive-hive-shims-default.xml
 to 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/ivy/report/org.apache.hive-hive-shims-default.html

ivy-retrieve:
 [echo] Project: shims

compile:
 [echo] Project: shims
 [echo] Building shims 0.20

build_shims:
 [echo] Project: shims
 [echo] Compiling 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/common/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/0.20/java
 against hadoop 0.20.2 
(https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/hadoopcore/hadoop-0.20.2)

ivy-init-settings:
 [echo] Project: shims

ivy-resolve-hadoop-shim:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml

ivy-retrieve-hadoop-shim:
 [echo] Project: shims
 [echo] Building shims 0.20S

build_shims:
 [echo] Project: shims
 [echo] Compiling 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/common/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/common-secure/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/0.20S/java
 against hadoop 1.0.0 
(https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/hadoopcore/hadoop-1.0.0)

ivy-init-settings:
 [echo] Project: shims

ivy-resolve-hadoop-shim:
 [echo] Project: shims
[ivy:resolve] :: loading settings :: file = 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/ivy/ivysettings.xml

ivy-retrieve-hadoop-shim:
 [echo] Project: shims
 [echo] Building shims 0.23

build_shims:
 [echo] Project: shims
 [echo] Compiling 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/ws/hive/shims/src/common/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/common-secure/java;/home/jenkins/jenkins-slave/workspace/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/hive/shims/src/0.23/java
 against hadoop 0.23.3 
(https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21-keepgoing=false/213/artifact/hive/build/hadoopcore/hadoop-0.23.3)


[jira] [Commented] (HIVE-3646) Add 'IGNORE PROTECTION' predicate for dropping partitions

2012-11-28 Thread Andrew Chalfant (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3646?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505697#comment-13505697
 ] 

Andrew Chalfant commented on HIVE-3646:
---

Namit should I document before or after the change has been pushed?

 Add 'IGNORE PROTECTION' predicate for dropping partitions
 -

 Key: HIVE-3646
 URL: https://issues.apache.org/jira/browse/HIVE-3646
 Project: Hive
  Issue Type: New Feature
  Components: CLI
Affects Versions: 0.9.0
Reporter: Andrew Chalfant
Assignee: Andrew Chalfant
Priority: Minor
 Attachments: HIVE-3646.1.patch.txt

   Original Estimate: 1m
  Remaining Estimate: 1m

 There are cases where it is desirable to move partitions between clusters. 
 Having to undo protection and then re-protect tables in order to delete 
 partitions from a source are multi-step and can leave us in a failed open 
 state where partition and table metadata is dirty. By implementing an 'rm 
 -rf'-like functionality, we can perform these operations atomically.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3646) Add 'IGNORE PROTECTION' predicate for dropping partitions

2012-11-28 Thread Andrew Chalfant (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3646?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505713#comment-13505713
 ] 

Andrew Chalfant commented on HIVE-3646:
---

I spoke with Kevin Wilfong and he said for me to just go ahead and make the 
changes to the docs, but I need a confluence account and he doesn't have 
permissions to make the account for me. Can you please make a confluence 
account with edit permissions for me please?

 Add 'IGNORE PROTECTION' predicate for dropping partitions
 -

 Key: HIVE-3646
 URL: https://issues.apache.org/jira/browse/HIVE-3646
 Project: Hive
  Issue Type: New Feature
  Components: CLI
Affects Versions: 0.9.0
Reporter: Andrew Chalfant
Assignee: Andrew Chalfant
Priority: Minor
 Attachments: HIVE-3646.1.patch.txt

   Original Estimate: 1m
  Remaining Estimate: 1m

 There are cases where it is desirable to move partitions between clusters. 
 Having to undo protection and then re-protect tables in order to delete 
 partitions from a source are multi-step and can leave us in a failed open 
 state where partition and table metadata is dirty. By implementing an 'rm 
 -rf'-like functionality, we can perform these operations atomically.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3384) HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC specification

2012-11-28 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505722#comment-13505722
 ] 

Phabricator commented on HIVE-3384:
---

mbautin has commented on the revision [jira] [HIVE-3384] HIVE JDBC module 
won't compile under JDK1.7 as new methods added in JDBC specification.

INLINE COMMENTS
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveConnection.java:49 Removed.
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDataSource.java:29 Removed.
  metastore/build.xml:106 Unfortunately, this seems to be required in order to 
run unit tests under JDK 7. Without this, verification errors such as P87 start 
happening, presumably because some of the dependencies have been compiled using 
an earlier version of javac and do not pass the more strict verification 
guidelines Java 7 uses by default.

REVISION DETAIL
  https://reviews.facebook.net/D6873

BRANCH
  trunk

To: cdrome, weidongbian, JIRA, ashutoshc, mbautin


 HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC 
 specification
 --

 Key: HIVE-3384
 URL: https://issues.apache.org/jira/browse/HIVE-3384
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Weidong Bian
Priority: Minor
 Attachments: D6873-0.9.1.patch, D6873.1.patch, HIVE-3384.2.patch, 
 HIVE-3384.patch, HIVE-JDK7-JDBC.patch


 jdbc module couldn't be compiled with jdk7 as it adds some abstract method in 
 the JDBC specification 
 some error info:
  error: HiveCallableStatement is not abstract and does not override abstract
 method TgetObject(String,ClassT) in CallableStatement
 .
 .
 .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3384) HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC specification

2012-11-28 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-3384:
--

Attachment: D6873.2.patch

mbautin updated the revision [jira] [HIVE-3384] HIVE JDBC module won't compile 
under JDK1.7 as new methods added in JDBC specification.
Reviewers: cdrome, weidongbian, JIRA, ashutoshc

  Addressing Ashutosh's comments.

REVISION DETAIL
  https://reviews.facebook.net/D6873

AFFECTED FILES
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveBaseResultSet.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveCallableStatement.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveConnection.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDataSource.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDatabaseMetaData.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDriver.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HivePreparedStatement.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveQueryResultSet.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveStatement.java
  metastore/build.xml

To: cdrome, weidongbian, JIRA, ashutoshc, mbautin


 HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC 
 specification
 --

 Key: HIVE-3384
 URL: https://issues.apache.org/jira/browse/HIVE-3384
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Weidong Bian
Priority: Minor
 Attachments: D6873-0.9.1.patch, D6873.1.patch, D6873.2.patch, 
 HIVE-3384.2.patch, HIVE-3384.patch, HIVE-JDK7-JDBC.patch


 jdbc module couldn't be compiled with jdk7 as it adds some abstract method in 
 the JDBC specification 
 some error info:
  error: HiveCallableStatement is not abstract and does not override abstract
 method TgetObject(String,ClassT) in CallableStatement
 .
 .
 .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3646) Add 'IGNORE PROTECTION' predicate for dropping partitions

2012-11-28 Thread Andrew Chalfant (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3646?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Chalfant updated HIVE-3646:
--

Attachment: HIVE-3646.2.patch.txt

New version which includes requested comments

 Add 'IGNORE PROTECTION' predicate for dropping partitions
 -

 Key: HIVE-3646
 URL: https://issues.apache.org/jira/browse/HIVE-3646
 Project: Hive
  Issue Type: New Feature
  Components: CLI
Affects Versions: 0.9.0
Reporter: Andrew Chalfant
Assignee: Andrew Chalfant
Priority: Minor
 Attachments: HIVE-3646.1.patch.txt, HIVE-3646.2.patch.txt

   Original Estimate: 1m
  Remaining Estimate: 1m

 There are cases where it is desirable to move partitions between clusters. 
 Having to undo protection and then re-protect tables in order to delete 
 partitions from a source are multi-step and can leave us in a failed open 
 state where partition and table metadata is dirty. By implementing an 'rm 
 -rf'-like functionality, we can perform these operations atomically.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3705) Adding authorization capability to the metastore

2012-11-28 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505754#comment-13505754
 ] 

Phabricator commented on HIVE-3705:
---

khorgath has commented on the revision HIVE-3705 [jira] Adding authorization 
capability to the metastore.

INLINE COMMENTS
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:49
  Good point. Thought about it a bit, and decided that the best place for this 
constant was in MetaStoreUtils along with some other similar constants there. 
Have refactored to push it there.

  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:77
  Agreed. Removing. I was mimicing existing code in HCatalog's HDFS Auth 
Provider, but you're right, we need to be stricter.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:89
 Agreed, same response.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:111
 Agreed, same response.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:95
 getPath is equivalent to a path constructed on 
table.getTTable().getSd().getLocation() only if it is nonEmpty, which it is in 
the else segment. Have made it clearer by avoiding referencing .getPath 
altogether.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:117
 Done.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:122
 There isn't a case with partition being null, removing that bit as with 77,89 
and 111 above. But it is possible for the location to be null as with the case 
of creating a partition - when the PreEventListener is triggered, it's possible 
for the part location to be null, in which case the correct behaviour is to 
check the table's permissions.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:165
 Interesting question - I assumed it was for creating an Index. That said, this 
is currently unused in Hive - there's no reference to this that I find in the 
codebase.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:168
 Ditto as with 165, but with the exception that HiveOperation does define them 
as read privileges for both LOCKTABLE and UNLOCKTABLE. That doesn't sound 
terribly right to me, as I don't think read privileges are enough to be able to 
perform either of these operations. I'm going to leave this as-is, and ditto 
with the INDEX case above unless you think we should change it. At any rate, 
these are not privileges currently in use, even if LOCK is partially defined.

  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:218
  Agreed, changing across all files in this patch.
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:301
 I was trying to keep changes minimal and not change HiveMetaStore too much, 
but yes, okay, refactoring, moving over to Warehouse
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java:308
 This does not reimplement Warehouse::getDatabasePath, it extends it(it calls 
it) by providing a default path if the location was null.

  
ql/src/test/org/apache/hadoop/hive/ql/security/TestAuthorizationPreEventListener.java:44
 Agreed, done across tests.
  
ql/src/test/org/apache/hadoop/hive/ql/security/TestDefaultHiveMetastoreAuthorizationProvider.java:137
 Added as requested.

REVISION DETAIL
  https://reviews.facebook.net/D6681

BRANCH
  HIVE-3705

To: JIRA, ashutoshc, khorgath


 Adding authorization capability to the metastore
 

 Key: HIVE-3705
 URL: https://issues.apache.org/jira/browse/HIVE-3705
 Project: Hive
  Issue Type: New Feature
  Components: Authorization, Metastore
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: HIVE-3705.D6681.1.patch, HIVE-3705.D6681.2.patch, 
 HIVE-3705.D6681.3.patch, hive-backend-auth.2.git.patch, 
 hive-backend-auth.git.patch, hivesec_investigation.pdf


 In an environment where multiple clients access a single metastore, and we 
 want to evolve hive security to a point where it's no longer simply 
 preventing users from shooting their own foot, we need to be able to 
 authorize metastore calls as well, instead of simply performing every 
 metastore api call that's made.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: 

[jira] [Created] (HIVE-3752) Add a non-sql API in hive to access data.

2012-11-28 Thread Nitay Joffe (JIRA)
Nitay Joffe created HIVE-3752:
-

 Summary: Add a non-sql API in hive to access data.
 Key: HIVE-3752
 URL: https://issues.apache.org/jira/browse/HIVE-3752
 Project: Hive
  Issue Type: Improvement
Reporter: Nitay Joffe


We would like to add an input/output format for accessing Hive data in Hadoop 
directly without having to use e.g. a transform. Using a transform
means having to do a whole map-reduce step with its own disk accesses and its 
imposed structure. It also means needing to have Hive be the base 
infrastructure for the entire system being developed which is not the right fit 
as we only need a small part of it (access to the data).

So we propose adding an API level InputFormat and OutputFormat to Hive that 
will make it trivially easy to select a table with partition spec and read from 
/ write to it. We chose this design to make it compatible with Hadoop so that 
existing systems that work with Hadoop's IO API will just work out of the box.

We need this system for the Giraph graph processing (http://giraph.apache.org/) 
as running graph jobs which read/write from Hive is a common use case.

[~namitjain] [~aching] [~kevinwilfong] [~apresta]


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3752) Add a non-sql API in hive to access data.

2012-11-28 Thread Nitay Joffe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nitay Joffe updated HIVE-3752:
--

Description: 
We would like to add an input/output format for accessing Hive data in Hadoop 
directly without having to use e.g. a transform. Using a transform
means having to do a whole map-reduce step with its own disk accesses and its 
imposed structure. It also means needing to have Hive be the base 
infrastructure for the entire system being developed which is not the right fit 
as we only need a small part of it (access to the data).

So we propose adding an API level InputFormat and OutputFormat to Hive that 
will make it trivially easy to select a table with partition spec and read from 
/ write to it. We chose this design to make it compatible with Hadoop so that 
existing systems that work with Hadoop's IO API will just work out of the box.

We need this system for the Giraph graph processing system 
(http://giraph.apache.org/) as running graph jobs which read/write from Hive is 
a common use case.

[~namitjain] [~aching] [~kevinwilfong] [~apresta]


  was:
We would like to add an input/output format for accessing Hive data in Hadoop 
directly without having to use e.g. a transform. Using a transform
means having to do a whole map-reduce step with its own disk accesses and its 
imposed structure. It also means needing to have Hive be the base 
infrastructure for the entire system being developed which is not the right fit 
as we only need a small part of it (access to the data).

So we propose adding an API level InputFormat and OutputFormat to Hive that 
will make it trivially easy to select a table with partition spec and read from 
/ write to it. We chose this design to make it compatible with Hadoop so that 
existing systems that work with Hadoop's IO API will just work out of the box.

We need this system for the Giraph graph processing (http://giraph.apache.org/) 
as running graph jobs which read/write from Hive is a common use case.

[~namitjain] [~aching] [~kevinwilfong] [~apresta]



 Add a non-sql API in hive to access data.
 -

 Key: HIVE-3752
 URL: https://issues.apache.org/jira/browse/HIVE-3752
 Project: Hive
  Issue Type: Improvement
Reporter: Nitay Joffe

 We would like to add an input/output format for accessing Hive data in Hadoop 
 directly without having to use e.g. a transform. Using a transform
 means having to do a whole map-reduce step with its own disk accesses and its 
 imposed structure. It also means needing to have Hive be the base 
 infrastructure for the entire system being developed which is not the right 
 fit as we only need a small part of it (access to the data).
 So we propose adding an API level InputFormat and OutputFormat to Hive that 
 will make it trivially easy to select a table with partition spec and read 
 from / write to it. We chose this design to make it compatible with Hadoop so 
 that existing systems that work with Hadoop's IO API will just work out of 
 the box.
 We need this system for the Giraph graph processing system 
 (http://giraph.apache.org/) as running graph jobs which read/write from Hive 
 is a common use case.
 [~namitjain] [~aching] [~kevinwilfong] [~apresta]

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3705) Adding authorization capability to the metastore

2012-11-28 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-3705:
--

Attachment: HIVE-3705.D6681.4.patch

khorgath updated the revision HIVE-3705 [jira] Adding authorization capability 
to the metastore.
Reviewers: JIRA, ashutoshc

  Updates per review comments


REVISION DETAIL
  https://reviews.facebook.net/D6681

AFFECTED FILES
  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
  conf/hive-default.xml.template
  metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
  metastore/src/java/org/apache/hadoop/hive/metastore/MetaStoreUtils.java
  metastore/src/java/org/apache/hadoop/hive/metastore/Warehouse.java
  ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveUtils.java
  ql/src/java/org/apache/hadoop/hive/ql/security/HadoopDefaultAuthenticator.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/HadoopDefaultMetastoreAuthenticator.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/HiveMetastoreAuthenticationProvider.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/AuthorizationPreEventListener.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/BitSetCheckedAuthorizationProvider.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/DefaultHiveAuthorizationProvider.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/DefaultHiveMetastoreAuthorizationProvider.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/HiveAuthorizationProviderBase.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/HiveMetastoreAuthorizationProvider.java
  
ql/src/java/org/apache/hadoop/hive/ql/security/authorization/StorageBasedAuthorizationProvider.java
  ql/src/java/org/apache/hadoop/hive/ql/session/SessionState.java
  
ql/src/test/org/apache/hadoop/hive/ql/security/DummyHiveMetastoreAuthorizationProvider.java
  
ql/src/test/org/apache/hadoop/hive/ql/security/InjectableDummyAuthenticator.java
  
ql/src/test/org/apache/hadoop/hive/ql/security/TestAuthorizationPreEventListener.java
  
ql/src/test/org/apache/hadoop/hive/ql/security/TestDefaultHiveMetastoreAuthorizationProvider.java

To: JIRA, ashutoshc, khorgath


 Adding authorization capability to the metastore
 

 Key: HIVE-3705
 URL: https://issues.apache.org/jira/browse/HIVE-3705
 Project: Hive
  Issue Type: New Feature
  Components: Authorization, Metastore
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: HIVE-3705.D6681.1.patch, HIVE-3705.D6681.2.patch, 
 HIVE-3705.D6681.3.patch, HIVE-3705.D6681.4.patch, 
 hive-backend-auth.2.git.patch, hive-backend-auth.git.patch, 
 hivesec_investigation.pdf


 In an environment where multiple clients access a single metastore, and we 
 want to evolve hive security to a point where it's no longer simply 
 preventing users from shooting their own foot, we need to be able to 
 authorize metastore calls as well, instead of simply performing every 
 metastore api call that's made.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3384) HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC specification

2012-11-28 Thread Phabricator (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Phabricator updated HIVE-3384:
--

Attachment: D6873.3.patch

mbautin updated the revision [jira] [HIVE-3384] HIVE JDBC module won't compile 
under JDK1.7 as new methods added in JDBC specification.
Reviewers: cdrome, weidongbian, JIRA, ashutoshc

  Adding another -XX:-UseSplitVerifier for unit tests to work on JDK7.

REVISION DETAIL
  https://reviews.facebook.net/D6873

AFFECTED FILES
  build-common.xml
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveBaseResultSet.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveCallableStatement.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveConnection.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDataSource.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDatabaseMetaData.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveDriver.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HivePreparedStatement.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveQueryResultSet.java
  jdbc/src/java/org/apache/hadoop/hive/jdbc/HiveStatement.java
  metastore/build.xml

To: cdrome, weidongbian, JIRA, ashutoshc, mbautin


 HIVE JDBC module won't compile under JDK1.7 as new methods added in JDBC 
 specification
 --

 Key: HIVE-3384
 URL: https://issues.apache.org/jira/browse/HIVE-3384
 Project: Hive
  Issue Type: Bug
  Components: JDBC
Reporter: Weidong Bian
Priority: Minor
 Attachments: D6873-0.9.1.patch, D6873.1.patch, D6873.2.patch, 
 D6873.3.patch, HIVE-3384.2.patch, HIVE-3384.patch, HIVE-JDK7-JDBC.patch


 jdbc module couldn't be compiled with jdk7 as it adds some abstract method in 
 the JDBC specification 
 some error info:
  error: HiveCallableStatement is not abstract and does not override abstract
 method TgetObject(String,ClassT) in CallableStatement
 .
 .
 .

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3709) Stop storing default ConfVars in temp file

2012-11-28 Thread Kevin Wilfong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13505838#comment-13505838
 ] 

Kevin Wilfong commented on HIVE-3709:
-

I had considered the solution Carl mentioned, but I didn't go that route 
because it could result in clutter in the directory specified, in particular we 
can't guarantee the file will be deleted in the presence of catastrophic 
failures.  To solve this the user would need to set up some sort of periodic 
cleanup which puts us back in the same position.  We might be able to work 
around this by regularly touching the file, but I'm not 100% sure.

Carl, are you still seeing threading problems with the most recent patch?  
TestHiveServerSessions has been succeeding for me consistently.

 Stop storing default ConfVars in temp file
 --

 Key: HIVE-3709
 URL: https://issues.apache.org/jira/browse/HIVE-3709
 Project: Hive
  Issue Type: Improvement
  Components: Configuration
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Attachments: HIVE-3709.1.patch.txt, HIVE-3709.2.patch.txt, 
 HIVE-3709.3.patch.txt


 To work around issues with Hadoop's Configuration object, specifically it's 
 addResource(InputStream), default configurations are written to a temp file 
 (I think HIVE-2362 introduced this).
 This, however, introduces the problem that once that file is deleted from 
 /tmp the client crashes.  This is particularly problematic for long running 
 services like the metastore server.
 Writing a custom InputStream to deal with the problems in the Configuration 
 object should provide a work around, which does not introduce a time bomb 
 into Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Hive-trunk-h0.21 - Build # 1823 - Still Failing

2012-11-28 Thread Apache Jenkins Server
Changes for Build #1779

Changes for Build #1780
[kevinwilfong] HIVE-3686. Fix compile errors introduced by the interaction of 
HIVE-1362 and HIVE-3524. (Shreepadma Venugopalan via kevinwilfong)


Changes for Build #1781
[namit] HIVE-3687 smb_mapjoin_13.q is nondeterministic
(Kevin Wilfong via namit)


Changes for Build #1782
[hashutosh] HIVE-2715: Upgrade Thrift dependency to 0.9.0 (Ashutosh Chauhan)


Changes for Build #1783
[kevinwilfong] HIVE-3654. block relative path access in hive. (njain via 
kevinwilfong)

[hashutosh] HIVE-3658 : Unable to generate the Hbase related unit tests using 
velocity templates on Windows (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3661 : Remove the Windows specific = related swizzle path 
changes from Proxy FileSystems (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3480 : Resource leak: Fix the file handle leaks in Symbolic 
 Symlink related input formats. (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1784
[kevinwilfong] HIVE-3675. NaN does not work correctly for round(n). (njain via 
kevinwilfong)

[cws] HIVE-3651. bucketmapjoin?.q tests fail with hadoop 0.23 (Prasad Mujumdar 
via cws)


Changes for Build #1785
[namit] HIVE-3613 Implement grouping_id function
(Ian Gorbachev via namit)

[namit] HIVE-3692 Update parallel test documentation
(Ivan Gorbachev via namit)

[namit] HIVE-3649 Hive List Bucketing - enhance DDL to specify list bucketing 
table
(Gang Tim Liu via namit)


Changes for Build #1786
[namit] HIVE-3696 Revert HIVE-3483 which causes performance regression
(Gang Tim Liu via namit)


Changes for Build #1787
[kevinwilfong] HIVE-3621. Make prompt in Hive CLI configurable. (Jingwei Lu via 
kevinwilfong)

[kevinwilfong] HIVE-3695. TestParse breaks due to HIVE-3675. (njain via 
kevinwilfong)


Changes for Build #1788
[kevinwilfong] HIVE-3557. Access to external URLs in hivetest.py. (Ivan 
Gorbachev via kevinwilfong)


Changes for Build #1789
[hashutosh] HIVE-3662 : TestHiveServer: testScratchDirShouldClearWhileStartup 
is failing on Windows (Kanna Karanam via Ashutosh Chauhan)

[hashutosh] HIVE-3659 : TestHiveHistory::testQueryloglocParentDirNotExist Test 
fails on Windows because of some resource leaks in ZK (Kanna Karanam via 
Ashutosh Chauhan)

[hashutosh] HIVE-3663 Unable to display the MR Job file path on Windows in case 
of MR job failures.  (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1790

Changes for Build #1791

Changes for Build #1792

Changes for Build #1793
[hashutosh] HIVE-3704 : name of some metastore scripts are not per convention 
(Ashutosh Chauhan)


Changes for Build #1794
[hashutosh] HIVE-3243 : ignore white space between entries of hive/hbase table 
mapping (Shengsheng Huang via Ashutosh Chauhan)

[hashutosh] HIVE-3215 : JobDebugger should use RunningJob.getTrackingURL 
(Bhushan Mandhani via Ashutosh Chauhan)


Changes for Build #1795
[cws] HIVE-3437. 0.23 compatibility: fix unit tests when building against 0.23 
(Chris Drome via cws)

[hashutosh] HIVE-3626 : RetryingHMSHandler should wrap JDOException inside 
MetaException (Bhushan Mandhani via Ashutosh Chauhan)

[hashutosh] HIVE-3560 : Hive always prints a warning message when using remote 
metastore (Travis Crawford via Ashutosh Chauhan)


Changes for Build #1796

Changes for Build #1797
[hashutosh] HIVE-3664 : Avoid to create a symlink for hive-contrib.jar file in 
dist\lib folder. (Kanna Karanam via Ashutosh Chauhan)


Changes for Build #1798
[namit] HIVE-3706 getBoolVar in FileSinkOperator can be optimized
(Kevin Wilfong via namit)

[namit] HIVE-3707 Round map/reduce progress down when it is in the range [99.5, 
100)
(Kevin Wilfong via namit)

[namit] HIVE-3471 Implement grouping sets in hive
(Ivan Gorbachev via namit)


Changes for Build #1799
[hashutosh] HIVE-3291 : fix fs resolvers (Ashish Singh via Ashutosh Chauhan)

[hashutosh] HIVE-3680 : Include Table information in Hive's AddPartitionEvent. 
(Mithun Radhakrishnan via Ashutosh Chauhan)


Changes for Build #1800
[hashutosh] HIVE-3520 : ivysettings.xml does not let you override 
.m2/repository (Raja Aluri via Ashutosh Chauhan)

[hashutosh] HIVE-3435 : Get pdk pluginTest passed when triggered from both 
builtin tests and pdk tests on hadoop23  (Zhenxiao Luo via Ashutosh Chauhan)


Changes for Build #1801

Changes for Build #1802

Changes for Build #1803

Changes for Build #1804

Changes for Build #1805

Changes for Build #1806
[hashutosh] Preparing for 0.11 development

[kevinwilfong] HIVE-3647. map-side groupby wrongly due to HIVE-3432. (njain via 
kevinwilfong)


Changes for Build #1807
[kevinwilfong] HIVE-3679. Unescape partition names returned by show partitions. 
(Wonho Kim via kevinwilfong)


Changes for Build #1808
[kevinwilfong] Remove extra files accidentally committed with HIVE-3679. 
(kevinwilfong)


Changes for Build #1809

Changes for Build #1810

Changes for Build #1811
[namit] HIVE-3735 PTest doesn't work due to hive snapshot version upgrade to 11
(Gang Tim Liu via namit)


Build failed in Jenkins: Hive-0.9.1-SNAPSHOT-h0.21 #213

2012-11-28 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/213/

--
[...truncated 36521 lines...]
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: 
file:/tmp/hudson/hive_2012-11-28_14-14-24_382_6161782169221319366/-mr-1
[junit] OK
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] Hive history 
file=https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/213/artifact/hive/build/service/tmp/hive_job_log_hudson_201211281414_657969550.txt
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] OK
[junit] PREHOOK: query: create table testhivedrivertable (num int)
[junit] PREHOOK: type: DROPTABLE
[junit] Copying file: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/ws/hive/data/files/kv1.txt
[junit] POSTHOOK: query: create table testhivedrivertable (num int)
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: load data local inpath 
'https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/ws/hive/data/files/kv1.txt'
 into table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] Copying data from 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/ws/hive/data/files/kv1.txt
[junit] Loading data to table default.testhivedrivertable
[junit] POSTHOOK: query: load data local inpath 
'https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/ws/hive/data/files/kv1.txt'
 into table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: select * from testhivedrivertable limit 10
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: 
file:/tmp/hudson/hive_2012-11-28_14-14-28_795_3773609989986/-mr-1
[junit] POSTHOOK: query: select * from testhivedrivertable limit 10
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: 
file:/tmp/hudson/hive_2012-11-28_14-14-28_795_3773609989986/-mr-1
[junit] OK
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] Hive history 
file=https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/213/artifact/hive/build/service/tmp/hive_job_log_hudson_201211281414_1311558659.txt
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] OK
[junit] PREHOOK: query: create table testhivedrivertable (num int)
[junit] PREHOOK: type: DROPTABLE
[junit] POSTHOOK: query: create table testhivedrivertable (num int)
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] PREHOOK: Input: default@testhivedrivertable
[junit] PREHOOK: Output: default@testhivedrivertable
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: DROPTABLE
[junit] POSTHOOK: Input: default@testhivedrivertable
[junit] POSTHOOK: Output: default@testhivedrivertable
[junit] OK
[junit] Hive history 
file=https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/213/artifact/hive/build/service/tmp/hive_job_log_hudson_201211281414_697160228.txt
[junit] Hive history 
file=https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/213/artifact/hive/build/service/tmp/hive_job_log_hudson_201211281414_1440676227.txt
[junit] Copying file: 
https://builds.apache.org/job/Hive-0.9.1-SNAPSHOT-h0.21/ws/hive/data/files/kv1.txt
[junit] PREHOOK: query: drop table testhivedrivertable
[junit] PREHOOK: type: DROPTABLE
[junit] POSTHOOK: query: drop table testhivedrivertable
[junit] POSTHOOK: type: 

[jira] [Created] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)
Gang Tim Liu created HIVE-3754:
--

 Summary: Trunk hadoop 23 build fails
 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu


check out the latest code from trunk
svn info 
Path: .
URL: http://svn.apache.org/repos/asf/hive/trunk
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1415005
Node Kind: directory
Schedule: normal
Last Changed Author: namit
Last Changed Rev: 1414608
Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)

ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23

..
{quote}
ivy-retrieve-hadoop-shim:
 [echo] Project: shims
[javac] Compiling 2 source files to 
/Users/gang/hive-trunk-11-28/build/shims/classes
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
 getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied to 
(org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultBlockSize(path);
[javac]  ^
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
 getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
to (org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultReplication(path);
[javac]  ^
[javac] 2 errors

BUILD FAILED
/Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred while 
executing this line:
{quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3754:
---

Description: 
check out the latest code from trunk
svn info 
{quote}
Path: .
URL: http://svn.apache.org/repos/asf/hive/trunk
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1415005
Node Kind: directory
Schedule: normal
Last Changed Author: namit
Last Changed Rev: 1414608
Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
{quote}

ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23


{quote}
ivy-retrieve-hadoop-shim:
 [echo] Project: shims
[javac] Compiling 2 source files to 
/Users/gang/hive-trunk-11-28/build/shims/classes
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
 getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied to 
(org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultBlockSize(path);
[javac]  ^
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
 getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
to (org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultReplication(path);
[javac]  ^
[javac] 2 errors

BUILD FAILED
/Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred while 
executing this line:
{quote}

  was:
check out the latest code from trunk
svn info 
Path: .
URL: http://svn.apache.org/repos/asf/hive/trunk
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1415005
Node Kind: directory
Schedule: normal
Last Changed Author: namit
Last Changed Rev: 1414608
Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)

ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23

..
{quote}
ivy-retrieve-hadoop-shim:
 [echo] Project: shims
[javac] Compiling 2 source files to 
/Users/gang/hive-trunk-11-28/build/shims/classes
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
 getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied to 
(org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultBlockSize(path);
[javac]  ^
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
 getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
to (org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultReplication(path);
[javac]  ^
[javac] 2 errors

BUILD FAILED
/Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred while 
executing this line:
{quote}


 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 svn info 
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gang Tim Liu updated HIVE-3754:
---

Description: 
check out the latest code from trunk
{code}
svn info 
{code}
{quote}
Path: .
URL: http://svn.apache.org/repos/asf/hive/trunk
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1415005
Node Kind: directory
Schedule: normal
Last Changed Author: namit
Last Changed Rev: 1414608
Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
{quote}
{code}
ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23
{code}

{quote}
ivy-retrieve-hadoop-shim:
 [echo] Project: shims
[javac] Compiling 2 source files to 
/Users/gang/hive-trunk-11-28/build/shims/classes
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
 getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied to 
(org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultBlockSize(path);
[javac]  ^
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
 getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
to (org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultReplication(path);
[javac]  ^
[javac] 2 errors

BUILD FAILED
/Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred while 
executing this line:
{quote}

  was:
check out the latest code from trunk
{code}
svn info 
{code}
{quote}
Path: .
URL: http://svn.apache.org/repos/asf/hive/trunk
Repository Root: http://svn.apache.org/repos/asf
Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
Revision: 1415005
Node Kind: directory
Schedule: normal
Last Changed Author: namit
Last Changed Rev: 1414608
Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
{quote}

ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23


{quote}
ivy-retrieve-hadoop-shim:
 [echo] Project: shims
[javac] Compiling 2 source files to 
/Users/gang/hive-trunk-11-28/build/shims/classes
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
 getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied to 
(org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultBlockSize(path);
[javac]  ^
[javac] 
/Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
 getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
to (org.apache.hadoop.fs.Path)
[javac] return fs.getDefaultReplication(path);
[javac]  ^
[javac] 2 errors

BUILD FAILED
/Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred while 
executing this line:
{quote}


 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: 

[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506092#comment-13506092
 ] 

Gang Tim Liu commented on HIVE-3754:


[~ashutoshc] [~cwsteinbach], are you aware of any build command change? Is this 
still the right way to build hadoop 23? It has been working well except recent 
trunk. Just ensure it's right way. thanks

{code}
ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
-Dhadoop.mr.rev=23
{code}

 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2800) NPE in create index without comment clause in external metastore

2012-11-28 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2800?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506120#comment-13506120
 ] 

Phabricator commented on HIVE-2800:
---

khorgath has abandoned the revision HIVE-2800 [jira] NPE in create index 
without comment clause in external metastore.

  Same change implemented by HIVE-3722, redundant now.

REVISION DETAIL
  https://reviews.facebook.net/D3519

To: JIRA, ashutoshc, khorgath


 NPE in create index without comment clause in external metastore
 --

 Key: HIVE-2800
 URL: https://issues.apache.org/jira/browse/HIVE-2800
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.7.1, 0.8.1, 0.9.0
Reporter: Daniel Dai
Assignee: Sushanth Sowmyan
Priority: Minor
 Fix For: 0.10.0


 This happens only when using external metastore (with --hiveconf 
 hive.metastore.uris=thrift://localhost:8088 --hiveconf 
 hive.metastore.local=false). Also if I gave a comment in the statement, this 
 exception go away.
 Here is the statement:
 create index test111 on table hcat_test(name) as 'compact' with deferred 
 rebuild;
 Here is the stack:
 2012-02-10 17:07:42,612 ERROR exec.Task (SessionState.java:printError(380)) - 
 FAILED: Error in metadata: java.lang.NullPointerException
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.NullPointerException
 at org.apache.hadoop.hive.ql.metadata.Hive.createIndex(Hive.java:725)
 at 
 org.apache.hadoop.hive.ql.exec.DDLTask.createIndex(DDLTask.java:822)
 at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:231)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:134)
 at 
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1291)
 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1082)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:933)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:255)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:212)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403)
 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:671)
 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:554)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
 Caused by: java.lang.NullPointerException
 at 
 org.apache.thrift.protocol.TBinaryProtocol.writeString(TBinaryProtocol.java:185)
 at org.apache.hadoop.hive.metastore.api.Index.write(Index.java:1032)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$add_index_args.write(ThriftHiveMetastore.java:47518)
 at org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:63)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_add_index(ThriftHiveMetastore.java:1675)
 at 
 org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.add_index(ThriftHiveMetastore.java:1666)
 at 
 org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createIndex(HiveMetaStoreClient.java:853)
 at org.apache.hadoop.hive.ql.metadata.Hive.createIndex(Hive.java:722)
 ... 17 more

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506141#comment-13506141
 ] 

Ashutosh Chauhan commented on HIVE-3754:


CCing: [~amalakar] This seems to be related to HIVE-3645

 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506149#comment-13506149
 ] 

Gang Tim Liu commented on HIVE-3754:


[~ashutoshc] thank you very much Tim

[~amalakar] [~viraj] would you please check if it is introduced by HIVE-3645? 
If yes, would you please fix it in trunk? It's blocking me. thanks a lot Tim

 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3693) Performance regression introduced by HIVE-3483

2012-11-28 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506158#comment-13506158
 ] 

Ashutosh Chauhan commented on HIVE-3693:


+1. Will commit if tests pass.

 Performance regression introduced by HIVE-3483
 --

 Key: HIVE-3693
 URL: https://issues.apache.org/jira/browse/HIVE-3693
 Project: Hive
  Issue Type: Bug
Reporter: Gang Tim Liu
Assignee: Thejas M Nair
Priority: Minor
 Attachments: HIVE-3693.1.patch


 https://issues.apache.org/jira/browse/HIVE-3483 introduced a performance 
 regression in the client side during split computation.
 The client side spends a lot more time in the split computation phase. The 
 problem is checkFilterPathContains method.
 While investigating, can you create a config to disable it by default?
 thanks

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Viraj Bhat (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506161#comment-13506161
 ] 

Viraj Bhat commented on HIVE-3754:
--

Arup Malakar recently checked this in to enable Federation capability.
Viraj

 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3753) 'CTAS' and INSERT OVERWRITE send different column names to the underlying SerDe

2012-11-28 Thread Mark Grover (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506162#comment-13506162
 ] 

Mark Grover commented on HIVE-3753:
---

Thanks for reporting this Matthew. This looks like a bug. Do you think a query 
like 
{code}
CREATE TABLE foo ROW FORMAT SERDE 'JsonSerDe' SELECT host as host from 
table1;
{code}
could serve as a work around for this?

 'CTAS' and INSERT OVERWRITE send different column names to the underlying 
 SerDe
 ---

 Key: HIVE-3753
 URL: https://issues.apache.org/jira/browse/HIVE-3753
 Project: Hive
  Issue Type: Bug
  Components: Serializers/Deserializers
Affects Versions: 0.9.0
Reporter: Matthew Rathbone
   Original Estimate: 24h
  Remaining Estimate: 24h

 A good example is with a JSON serde 
 (https://github.com/rathboma/Hive-JSON-Serde-1)
 Here is a simple example of how the two results differ:
 CREATE TABLE foo ROW FORMAT SERDE 'JsonSerDe' SELECT host from table1;
 generates = {_col0: localhost}
 CREATE TABLE foo(host string) ROW FORMAT SERDE 'JsonSerDe';
 INSERT OVERWRITE TABLE FOO SELECT host FROM table;
 generates = {host: localhost}
 The SerDe gets passed column names in two places:
 1) The property Constants.LIST_COLUMNS
 2) It gets passed a StructObjectInspector on serialize
 In the CTAS example above, both of these contain '_col0' as the column name. 
 This is not true in the second example, as the LIST_COLUMNS property contains 
 the real column names.
 I'd be happy to help out with this change, but I fear that the solution lies 
 somewhere in SemanticAnalyser.java, and I'm having a hard time finding my way 
 around.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Gang Tim Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506166#comment-13506166
 ] 

Gang Tim Liu commented on HIVE-3754:


Viraj can you check if it breaks Hadoop 23 build? You can use the following to 
check

Sent from my iPhone

On Nov 28, 2012, at 6:25 PM, Viraj Bhat (JIRA) j...@apache.org wrote:



 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3754) Trunk hadoop 23 build fails

2012-11-28 Thread Arup Malakar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3754?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506174#comment-13506174
 ] 

Arup Malakar commented on HIVE-3754:


Hi Gang the API getDefaultBlockSize(Path) and getDefaultReplication(Path) in 
FileSystem is not available on 0.23.1 they are in 0.23.3 though or 2.0.0-alpha. 
That is why they are building fine with default configuration , but won't 
compile when you use Hadoop 0.23.1. Is this a concern?

 Trunk hadoop 23 build fails
 ---

 Key: HIVE-3754
 URL: https://issues.apache.org/jira/browse/HIVE-3754
 Project: Hive
  Issue Type: Bug
  Components: Build Infrastructure
Reporter: Gang Tim Liu

 check out the latest code from trunk
 {code}
 svn info 
 {code}
 {quote}
 Path: .
 URL: http://svn.apache.org/repos/asf/hive/trunk
 Repository Root: http://svn.apache.org/repos/asf
 Repository UUID: 13f79535-47bb-0310-9956-ffa450edef68
 Revision: 1415005
 Node Kind: directory
 Schedule: normal
 Last Changed Author: namit
 Last Changed Rev: 1414608
 Last Changed Date: 2012-11-28 01:36:27 -0800 (Wed, 28 Nov 2012)
 {quote}
 {code}
 ant clean package -Dhadoop.version=0.23.1 -Dhadoop-0.23.version=0.23.1 
 -Dhadoop.mr.rev=23
 {code}
 {quote}
 ivy-retrieve-hadoop-shim:
  [echo] Project: shims
 [javac] Compiling 2 source files to 
 /Users/gang/hive-trunk-11-28/build/shims/classes
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:122:
  getDefaultBlockSize() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultBlockSize(path);
 [javac]  ^
 [javac] 
 /Users/gang/hive-trunk-11-28/shims/src/0.23/java/org/apache/hadoop/hive/shims/Hadoop23Shims.java:127:
  getDefaultReplication() in org.apache.hadoop.fs.FileSystem cannot be applied 
 to (org.apache.hadoop.fs.Path)
 [javac] return fs.getDefaultReplication(path);
 [javac]  ^
 [javac] 2 errors
 BUILD FAILED
 /Users/gang/hive-trunk-11-28/build.xml:302: The following error occurred 
 while executing this line:
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3709) Stop storing default ConfVars in temp file

2012-11-28 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach updated HIVE-3709:
-

   Resolution: Fixed
Fix Version/s: 0.11
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks Kevin!


 Stop storing default ConfVars in temp file
 --

 Key: HIVE-3709
 URL: https://issues.apache.org/jira/browse/HIVE-3709
 Project: Hive
  Issue Type: Improvement
  Components: Configuration
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Fix For: 0.11

 Attachments: HIVE-3709.1.patch.txt, HIVE-3709.2.patch.txt, 
 HIVE-3709.3.patch.txt


 To work around issues with Hadoop's Configuration object, specifically it's 
 addResource(InputStream), default configurations are written to a temp file 
 (I think HIVE-2362 introduced this).
 This, however, introduces the problem that once that file is deleted from 
 /tmp the client crashes.  This is particularly problematic for long running 
 services like the metastore server.
 Writing a custom InputStream to deal with the problems in the Configuration 
 object should provide a work around, which does not introduce a time bomb 
 into Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3709) Stop storing default ConfVars in temp file

2012-11-28 Thread Carl Steinbach (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3709?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13506216#comment-13506216
 ] 

Carl Steinbach commented on HIVE-3709:
--

@Kevin: I tried running TestHiveServerSessions again and wasn't able to provoke 
the failure I saw before.

 Stop storing default ConfVars in temp file
 --

 Key: HIVE-3709
 URL: https://issues.apache.org/jira/browse/HIVE-3709
 Project: Hive
  Issue Type: Improvement
  Components: Configuration
Affects Versions: 0.10.0
Reporter: Kevin Wilfong
Assignee: Kevin Wilfong
 Fix For: 0.11

 Attachments: HIVE-3709.1.patch.txt, HIVE-3709.2.patch.txt, 
 HIVE-3709.3.patch.txt


 To work around issues with Hadoop's Configuration object, specifically it's 
 addResource(InputStream), default configurations are written to a temp file 
 (I think HIVE-2362 introduced this).
 This, however, introduces the problem that once that file is deleted from 
 /tmp the client crashes.  This is particularly problematic for long running 
 services like the metastore server.
 Writing a custom InputStream to deal with the problems in the Configuration 
 object should provide a work around, which does not introduce a time bomb 
 into Hive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-3596) Regression - HiveConf static variable causes issues in long running JVM instances with /tmp/ data

2012-11-28 Thread Carl Steinbach (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carl Steinbach resolved HIVE-3596.
--

Resolution: Duplicate

Resolving this is a duplicate of HIVE-3709.

 Regression - HiveConf static variable causes issues in long running JVM 
 instances with /tmp/ data
 -

 Key: HIVE-3596
 URL: https://issues.apache.org/jira/browse/HIVE-3596
 Project: Hive
  Issue Type: Bug
  Components: Configuration
Affects Versions: 0.8.0, 0.8.1, 0.9.0
Reporter: Chris McConnell
Assignee: Chris McConnell
 Fix For: 0.10.0, 0.9.0, 0.8.1

 Attachments: HIVE-3596.patch


 With Hive 0.8.x, HiveConf was changed to utilize the private, static member 
 confVarURL which points to /tmp/hive-user-tmp_number.xml for job 
 configuration settings. 
 During long running JVMs, such as a Beeswax server, which creates multiple 
 HiveConf objects over time this variable does not properly get updated 
 between jobs and can cause job failure if the OS cleans /tmp/ during a cron 
 job. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira