[jira] [Commented] (HIVE-4825) Separate MapredWork into MapWork and ReduceWork

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721565#comment-13721565
 ] 

Hive QA commented on HIVE-4825:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594510/HIVE-4825.5.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 2653 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_dynamic_partition_skip_default
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/205/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/205/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.CleanupPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests failed with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

 Separate MapredWork into MapWork and ReduceWork
 ---

 Key: HIVE-4825
 URL: https://issues.apache.org/jira/browse/HIVE-4825
 Project: Hive
  Issue Type: Improvement
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4825.1.patch, HIVE-4825.2.code.patch, 
 HIVE-4825.2.testfiles.patch, HIVE-4825.3.testfiles.patch, HIVE-4825.4.patch, 
 HIVE-4825.5.patch


 Right now all the information needed to run an MR job is captured in 
 MapredWork. This class has aliases, tagging info, table descriptors etc.
 For Tez and MRR it will be useful to break this into map and reduce specific 
 pieces. The separation is natural and I think has value in itself, it makes 
 the code easier to understand. However, it will also allow us to reuse these 
 abstractions in Tez where you'll have a graph of these instead of just 1M and 
 0-1R.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4825) Separate MapredWork into MapWork and ReduceWork

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-4825:
-

Attachment: HIVE-4825.6.patch

Unfortunately, I missed updating the golden file for the new skip default 
partition testcase. Patch .6 has the update for the golden file.

 Separate MapredWork into MapWork and ReduceWork
 ---

 Key: HIVE-4825
 URL: https://issues.apache.org/jira/browse/HIVE-4825
 Project: Hive
  Issue Type: Improvement
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4825.1.patch, HIVE-4825.2.code.patch, 
 HIVE-4825.2.testfiles.patch, HIVE-4825.3.testfiles.patch, HIVE-4825.4.patch, 
 HIVE-4825.5.patch, HIVE-4825.6.patch


 Right now all the information needed to run an MR job is captured in 
 MapredWork. This class has aliases, tagging info, table descriptors etc.
 For Tez and MRR it will be useful to break this into map and reduce specific 
 pieces. The separation is natural and I think has value in itself, it makes 
 the code easier to understand. However, it will also allow us to reuse these 
 abstractions in Tez where you'll have a graph of these instead of just 1M and 
 0-1R.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3255) Add DBTokenStore to store Delegation Tokens in DB

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-3255:
-

Attachment: HIVE-3255.6.patch

Re-uploading Ashutosh' patch with different name to trigger pre-commit build. 
(.6 not _6)

 Add DBTokenStore to store Delegation Tokens in DB
 -

 Key: HIVE-3255
 URL: https://issues.apache.org/jira/browse/HIVE-3255
 Project: Hive
  Issue Type: New Feature
  Components: Metastore, Security
Affects Versions: 0.9.0
Reporter: Ashutosh Chauhan
Assignee: Deepesh Khandelwal
 Attachments: hive-3255_2.patch, hive-3255_3.patch, hive-3255_4.patch, 
 hive-3255_5.patch, HIVE-3255_6.patch, HIVE-3255.6.patch, hive-3255.patch


 Before HIVE-1696 metastore was completely stateless and all the data is in 
 backend db. HIVE-1696 added delegation tokens for metastore which metastore 
 needs to keep in memory. HIVE-2467 added support for storing delegation 
 tokens in ZooKeeper via an interface {{DelegationTokenStore}}. This jira is 
 about using DB backend for storing tokens which metastore uses for storing 
 data. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4051) Hive's metastore suffers from 1+N queries when querying partitions is slow

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721584#comment-13721584
 ] 

Hive QA commented on HIVE-4051:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594511/HIVE-4051.D11805.2.patch

{color:red}ERROR:{color} -1 due to 59 failed/errored test(s), 2653 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppr_pushdown2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_or
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_rc
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_insert_into2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_compact_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_and
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat11
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_explain_dependency2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_union_view
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_special_char
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap_auto_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_compact
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_inputddl7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_archive_multi
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mi
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_combine2_hadoop20
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_merge_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge_dynamic_partition3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_columnstats_partlvl
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_bitmap
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats4
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_merge3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_decode_name
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ql_rewrite_gbtoidx
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats_noscan_2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part10
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats12
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_6
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_unused
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_alter_partition_coltype
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_stale_partitioned
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_bucket_groupby
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_explain_dependency
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_concatenate_inherit_table_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auth
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_partition_wise_fileformat5
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input24
org.apache.hcatalog.api.TestHCatClient.testGetPartitionsWithPartialSpec
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part8
org.apache.hcatalog.api.TestHCatClient.testPartitionsHCatClientImpl
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_explain_logical
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_rename_partition_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_select_unquote_not
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_load_dyn_part9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_updateAccessTime
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_stats7
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_index_auto_mult_tables_compact
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_database_drop
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_constant_where
org.apache.hcatalog.api.TestHCatClient.testDropPartitionsWithPartialSpec
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/206/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/206/console

Messages:
{noformat}
Executing 

[jira] [Commented] (HIVE-3255) Add DBTokenStore to store Delegation Tokens in DB

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721617#comment-13721617
 ] 

Hive QA commented on HIVE-3255:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594528/HIVE-3255.6.patch

{color:green}SUCCESS:{color} +1 2646 tests passed

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/207/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/207/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.CleanupPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

 Add DBTokenStore to store Delegation Tokens in DB
 -

 Key: HIVE-3255
 URL: https://issues.apache.org/jira/browse/HIVE-3255
 Project: Hive
  Issue Type: New Feature
  Components: Metastore, Security
Affects Versions: 0.9.0
Reporter: Ashutosh Chauhan
Assignee: Deepesh Khandelwal
 Attachments: hive-3255_2.patch, hive-3255_3.patch, hive-3255_4.patch, 
 hive-3255_5.patch, HIVE-3255_6.patch, HIVE-3255.6.patch, hive-3255.patch


 Before HIVE-1696 metastore was completely stateless and all the data is in 
 backend db. HIVE-1696 added delegation tokens for metastore which metastore 
 needs to keep in memory. HIVE-2467 added support for storing delegation 
 tokens in ZooKeeper via an interface {{DelegationTokenStore}}. This jira is 
 about using DB backend for storing tokens which metastore uses for storing 
 data. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [Discuss] project chop up

2013-07-27 Thread Edward Capriolo
Or feel free to suggest different approach. I am used to managing software
as multi-module maven projects.
From a development standpoint if I was working on beeline, it would be nice
to only require some of the sub-projects to be open in my IDE to do that.
Also managing everything globally is not ideal.

Hive's project layout, build, and test infrastructure is just funky. It has
to do a few interesting things (shims, testing), but I do not think what we
are doing justifies the massive ant build system we have. Ant is so ten
years ago.



On Sat, Jul 27, 2013 at 12:04 AM, Alan Gates ga...@hortonworks.com wrote:

 But I assume they'd still be a part of targets like package, tar, and
 binary?  Making them compile and test separately and explicitly load the
 core Hive jars from maven/ivy seems reasonable.

 Alan.

 On Jul 26, 2013, at 8:40 PM, Brock Noland wrote:

  Hi,
 
  I think thats part of it but I'd like to decouple the downstream projects
  even further so that the only connection is the dependency on the hive
 jars.
 
  Brock
  On Jul 26, 2013 10:10 PM, Alan Gates ga...@hortonworks.com wrote:
 
  I'm not sure how this is different from what hcat does today.  It needs
  Hive's jars to compile, so it's one of the last things in the compile
 step.
  Would moving the other modules you note to be in the same category be
  enough?  Did you want to also make it so that the default ant target
  doesn't compile those?
 
  Alan.
 
  On Jul 26, 2013, at 4:09 PM, Edward Capriolo wrote:
 
  My mistake on saying hcat was a fork metastore. I had a brain fart for
 a
  moment.
 
  One way we could do this is create a folder called downstream. In our
  release step we can execute the downstream builds and then copy the
 files
  we need back. So nothing downstream will be on the classpath of the
 main
  project.
 
  This could help us breakup ql as well. Things like exotic file formats
 ,
  and things that are pluggable like zk locking can go here. That might
 be
  overkill.
 
  For now we can focus on building downstream and hivethrift1might be the
  first thing to try to downstream.
 
 
  On Friday, July 26, 2013, Thejas Nair the...@hortonworks.com wrote:
  +1 to the idea of making the build of core hive and other downstream
  components independent.
 
  bq.  I was under the impression that Hcat and hive-metastore was
  supposed to merge up somehow.
 
  The metastore code was never forked. Hcat was just using
  hive-metastore and making the metadata available to rest of hadoop
  (pig, java MR..).
  A lot of the changes that were driven by hcat goals were being made in
  hive-metastore. You can think of hcat as set of libraries that let pig
  and java MR use hive metastore. Since hcat is closely tied to
  hive-metastore, it makes sense to have them in same project.
 
 
  On Fri, Jul 26, 2013 at 6:33 AM, Edward Capriolo 
 edlinuxg...@gmail.com
 
  wrote:
  Also i believe hcatalog web can fall into the same designation.
 
  Question , hcatalog was initily a big hive-metastore fork. I was
 under
  the
  impression that Hcat and hive-metastore was supposed to merge up
  somehow.
  What is the status on that? I remember that was one of the core
 reasons
  we
  brought it in.
 
  On Friday, July 26, 2013, Edward Capriolo edlinuxg...@gmail.com
  wrote:
  I prefer option 3 as well.
 
 
  On Fri, Jul 26, 2013 at 12:52 AM, Brock Noland br...@cloudera.com
  wrote:
 
  On Thu, Jul 25, 2013 at 9:48 PM, Edward Capriolo 
  edlinuxg...@gmail.com
  wrote:
 
  I have been developing my laptop on a duel core 2 GB Ram laptop
 for
  years
  now. With the addition of hcatalog, hive-thrift2, and some other
  growth
  trying to develop hive in a eclipse on this machine craws,
  especially
  if
  'build automatically' is turned on. As we look to add on more
 things
  this
  is only going to get worse.
 
  I am also noticing issues like this:
 
  https://issues.apache.org/jira/browse/HIVE-4849
 
  What I think we should do is strip down/out optional parts of
 hive.
 
  1) Hive Hbase
  This should really be it's own project to do this right we really
  have to
  have multiple branches since hbase is not backwards compatible.
 
  2) Hive Web Interface
  Now really a big project but not really critical can be just as
  easily
  be
  build separately
 
  3) hive thrift 1
  We have hive thrift 2 now, it is time for the sun to set on
  hivethrift1,
 
  4) odbc
  Not entirely convinced about this one but it is really not
 critical
  to
  running hive.
 
  What I think we should do is create sub-projects for the above
  things
  or
  simply move them into directories that do not build with hive.
  Ideally
  they
  would use maven to pull dependencies.
 
  What does everyone think?
 
 
  I agree that projects like the HBase handler and probably others as
  well
  should somehow be downstream projects which simply depend on the
  hive
  jars.  I see a couple alternatives for this:
 
  * Take the module in question to the Apache Incubator
  * Move the 

[jira] [Commented] (HIVE-2905) Desc table can't show non-ascii comments

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2905?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721641#comment-13721641
 ] 

Edward Capriolo commented on HIVE-2905:
---

+1

 Desc table can't show non-ascii comments
 

 Key: HIVE-2905
 URL: https://issues.apache.org/jira/browse/HIVE-2905
 Project: Hive
  Issue Type: Bug
  Components: CLI
Affects Versions: 0.7.0, 0.10.0
 Environment: hive 0.7.0, mysql 5.1.45
 hive 0.10.0, mysql 5.5.30
Reporter: Sheng Zhou
  Labels: patch
 Attachments: HIVE-2905.D11487.1.patch, utf8-desc-comment.patch


 When desc a table with command line or hive jdbc way, the table's comment 
 can't be read.
 1. I have updated javax.jdo.option.ConnectionURL parameter in hive-site.xml 
 file.
jdbc:mysql://*.*.*.*:3306/hive?characterEncoding=UTF-8
 2. In mysql database, the comment field of COLUMNS table can be read normally.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-2905) Desc table can't show non-ascii comments

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-2905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-2905:
--

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

Thank you everyone involved. Great to see hive adoption in languages I do not 
understand!

 Desc table can't show non-ascii comments
 

 Key: HIVE-2905
 URL: https://issues.apache.org/jira/browse/HIVE-2905
 Project: Hive
  Issue Type: Bug
  Components: CLI
Affects Versions: 0.7.0, 0.10.0
 Environment: hive 0.7.0, mysql 5.1.45
 hive 0.10.0, mysql 5.5.30
Reporter: Sheng Zhou
Assignee: Navis
  Labels: patch
 Fix For: 0.12.0

 Attachments: HIVE-2905.D11487.1.patch, utf8-desc-comment.patch


 When desc a table with command line or hive jdbc way, the table's comment 
 can't be read.
 1. I have updated javax.jdo.option.ConnectionURL parameter in hive-site.xml 
 file.
jdbc:mysql://*.*.*.*:3306/hive?characterEncoding=UTF-8
 2. In mysql database, the comment field of COLUMNS table can be read normally.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4586) [HCatalog] WebHCat should return 404 error for undefined resource

2013-07-27 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721650#comment-13721650
 ] 

Alan Gates commented on HIVE-4586:
--

When I run the unit tests on this I see Test 
org.apache.hcatalog.templeton.TestWebHCatE2e failing.  Other than that 
everything looks good.

 [HCatalog] WebHCat should return 404 error for undefined resource
 -

 Key: HIVE-4586
 URL: https://issues.apache.org/jira/browse/HIVE-4586
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.11.0
Reporter: Daniel Dai
Assignee: Daniel Dai
 Fix For: 0.12.0

 Attachments: HIVE-4586-1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4551) HCatLoader smallint/tinyint promotions to Int have issues with ORC integration

2013-07-27 Thread Alan Gates (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4551?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Gates updated HIVE-4551:
-

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Patch checked in.

 HCatLoader smallint/tinyint promotions to Int have issues with ORC integration
 --

 Key: HIVE-4551
 URL: https://issues.apache.org/jira/browse/HIVE-4551
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Reporter: Sushanth Sowmyan
Assignee: Sushanth Sowmyan
 Attachments: 4551.patch


 This was initially reported from an e2e test run, with the following E2E test:
 {code}
 {
 'name' = 'Hadoop_ORC_Write',
 'tests' = [
 {
  'num' = 1
 ,'hcat_prep'=q\
 drop table if exists hadoop_orc;
 create table hadoop_orc (
 t tinyint,
 si smallint,
 i int,
 b bigint,
 f float,
 d double,
 s string)
 stored as orc;\
 ,'hadoop' = q\
 jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteText -libjars 
 :HCAT_JAR: :THRIFTSERVER: all100k hadoop_orc\,
 ,'result_table' = 'hadoop_orc'
 ,'sql' = q\select * from all100k;\
 ,'floatpostprocess' = 1
 ,'delimiter' = '   '
 },
],
 },
 {code}
 This fails with the following error:
 {code}
 2013-04-26 00:26:07,437 WARN org.apache.hadoop.mapred.Child: Error running 
 child
 org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error 
 converting read value to tuple
   at 
 org.apache.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76)
   at org.apache.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:53)
   at 
 org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211)
   at 
 org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532)
   at 
 org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:765)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:369)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1195)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 Caused by: java.lang.ClassCastException: 
 org.apache.hadoop.hive.serde2.io.ByteWritable cannot be cast to 
 org.apache.hadoop.io.IntWritable
   at 
 org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableIntObjectInspector.getPrimitiveJavaObject(WritableIntObjectInspector.java:45)
   at 
 org.apache.hcatalog.data.HCatRecordSerDe.serializePrimitiveField(HCatRecordSerDe.java:290)
   at 
 org.apache.hcatalog.data.HCatRecordSerDe.serializeField(HCatRecordSerDe.java:192)
   at org.apache.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:53)
   at org.apache.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:97)
   at 
 org.apache.hcatalog.mapreduce.HCatRecordReader.nextKeyValue(HCatRecordReader.java:203)
   at 
 org.apache.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:63)
   ... 12 more
 2013-04-26 00:26:07,440 INFO org.apache.hadoop.mapred.Task: Runnning cleanup 
 for the task
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-2906) Support providing some table properties by user via SQL

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-2906?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721656#comment-13721656
 ] 

Edward Capriolo commented on HIVE-2906:
---

+1. I think this is a very clean way to handle this. This opens up a lot of 
different ways to use input formats and serde

 Support providing some table properties by user via SQL
 ---

 Key: HIVE-2906
 URL: https://issues.apache.org/jira/browse/HIVE-2906
 Project: Hive
  Issue Type: Improvement
  Components: Query Processor
Reporter: Navis
Assignee: Navis
 Attachments: ASF.LICENSE.NOT.GRANTED--HIVE-2906.D2499.1.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2906.D2499.2.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2906.D2499.3.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2906.D2499.4.patch, 
 ASF.LICENSE.NOT.GRANTED--HIVE-2906.D2499.5.patch, HIVE-2906.D2499.6.patch, 
 HIVE-2906.D2499.7.patch


 Some properties are needed to be provided to StorageHandler by user in 
 runtime. It might be an address for remote resource or retry count for access 
 or maximum version count(for hbase), etc.
 For example,  
 {code}
 select emp.empno, emp.ename from hbase_emp ('max.version'='3') emp;
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [Discuss] project chop up

2013-07-27 Thread Brock Noland
I am open to any approach to making the build and project more modular.
 Maven is no silver bullet but it does have some big positives. Most
importantly I have been a little distressed with Ivy when changing library
versions as it caused hundreds of tests to fail for no good reason in
HIVE-3632.  I have hit this several times on Ivy projects.  As has been
iterated before offline and eclipse support are built in.  IMO alignment
with the other Hadoop ecosystem projects is another win.


On Sat, Jul 27, 2013 at 9:03 AM, Edward Capriolo edlinuxg...@gmail.comwrote:

 Or feel free to suggest different approach. I am used to managing software
 as multi-module maven projects.
 From a development standpoint if I was working on beeline, it would be nice
 to only require some of the sub-projects to be open in my IDE to do that.
 Also managing everything globally is not ideal.

 Hive's project layout, build, and test infrastructure is just funky. It has
 to do a few interesting things (shims, testing), but I do not think what we
 are doing justifies the massive ant build system we have. Ant is so ten
 years ago.



 On Sat, Jul 27, 2013 at 12:04 AM, Alan Gates ga...@hortonworks.com
 wrote:

  But I assume they'd still be a part of targets like package, tar, and
  binary?  Making them compile and test separately and explicitly load the
  core Hive jars from maven/ivy seems reasonable.
 
  Alan.
 
  On Jul 26, 2013, at 8:40 PM, Brock Noland wrote:
 
   Hi,
  
   I think thats part of it but I'd like to decouple the downstream
 projects
   even further so that the only connection is the dependency on the hive
  jars.
  
   Brock
   On Jul 26, 2013 10:10 PM, Alan Gates ga...@hortonworks.com wrote:
  
   I'm not sure how this is different from what hcat does today.  It
 needs
   Hive's jars to compile, so it's one of the last things in the compile
  step.
   Would moving the other modules you note to be in the same category be
   enough?  Did you want to also make it so that the default ant target
   doesn't compile those?
  
   Alan.
  
   On Jul 26, 2013, at 4:09 PM, Edward Capriolo wrote:
  
   My mistake on saying hcat was a fork metastore. I had a brain fart
 for
  a
   moment.
  
   One way we could do this is create a folder called downstream. In our
   release step we can execute the downstream builds and then copy the
  files
   we need back. So nothing downstream will be on the classpath of the
  main
   project.
  
   This could help us breakup ql as well. Things like exotic file
 formats
  ,
   and things that are pluggable like zk locking can go here. That might
  be
   overkill.
  
   For now we can focus on building downstream and hivethrift1might be
 the
   first thing to try to downstream.
  
  
   On Friday, July 26, 2013, Thejas Nair the...@hortonworks.com
 wrote:
   +1 to the idea of making the build of core hive and other downstream
   components independent.
  
   bq.  I was under the impression that Hcat and hive-metastore was
   supposed to merge up somehow.
  
   The metastore code was never forked. Hcat was just using
   hive-metastore and making the metadata available to rest of hadoop
   (pig, java MR..).
   A lot of the changes that were driven by hcat goals were being made
 in
   hive-metastore. You can think of hcat as set of libraries that let
 pig
   and java MR use hive metastore. Since hcat is closely tied to
   hive-metastore, it makes sense to have them in same project.
  
  
   On Fri, Jul 26, 2013 at 6:33 AM, Edward Capriolo 
  edlinuxg...@gmail.com
  
   wrote:
   Also i believe hcatalog web can fall into the same designation.
  
   Question , hcatalog was initily a big hive-metastore fork. I was
  under
   the
   impression that Hcat and hive-metastore was supposed to merge up
   somehow.
   What is the status on that? I remember that was one of the core
  reasons
   we
   brought it in.
  
   On Friday, July 26, 2013, Edward Capriolo edlinuxg...@gmail.com
   wrote:
   I prefer option 3 as well.
  
  
   On Fri, Jul 26, 2013 at 12:52 AM, Brock Noland 
 br...@cloudera.com
   wrote:
  
   On Thu, Jul 25, 2013 at 9:48 PM, Edward Capriolo 
   edlinuxg...@gmail.com
   wrote:
  
   I have been developing my laptop on a duel core 2 GB Ram laptop
  for
   years
   now. With the addition of hcatalog, hive-thrift2, and some other
   growth
   trying to develop hive in a eclipse on this machine craws,
   especially
   if
   'build automatically' is turned on. As we look to add on more
  things
   this
   is only going to get worse.
  
   I am also noticing issues like this:
  
   https://issues.apache.org/jira/browse/HIVE-4849
  
   What I think we should do is strip down/out optional parts of
  hive.
  
   1) Hive Hbase
   This should really be it's own project to do this right we
 really
   have to
   have multiple branches since hbase is not backwards compatible.
  
   2) Hive Web Interface
   Now really a big project but not really critical can be just as
   easily
   be
   build 

[jira] [Commented] (HIVE-4947) Add myself to committer list

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721685#comment-13721685
 ] 

Edward Capriolo commented on HIVE-4947:
---

+1. Can you please set my company to 'm6d' while you are at it? thanks

 Add myself to committer list
 

 Key: HIVE-4947
 URL: https://issues.apache.org/jira/browse/HIVE-4947
 Project: Hive
  Issue Type: Task
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4947.patch


 Update website to include myself.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4931) MAX/MIN aggregates yield incorrect results

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721686#comment-13721686
 ] 

Edward Capriolo commented on HIVE-4931:
---

Why no tests? 

 MAX/MIN aggregates yield incorrect results
 --

 Key: HIVE-4931
 URL: https://issues.apache.org/jira/browse/HIVE-4931
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: vectorization-branch
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
 Fix For: vectorization-branch

 Attachments: HIVE-4931.1.patch


 With the dataset attached by Tony, running:
  
 select max(ctinyint), max(csmallint), max(cint), max(cbigint), max(cfloat), 
 max(cdouble) from alltypes_orc;
  
 this is the result in rowmode:
 OK
 62  16379   1073737829  2146661184  62.016376.0
  
 This is the result in vector mode:
 OK
 1   16379   1071926788  -106115134  62.016361.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (HIVE-4931) MAX/MIN aggregates yield incorrect results

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721686#comment-13721686
 ] 

Edward Capriolo edited comment on HIVE-4931 at 7/27/13 5:08 PM:


Why no tests? What is to prevent a regression?

  was (Author: appodictic):
Why no tests? 
  
 MAX/MIN aggregates yield incorrect results
 --

 Key: HIVE-4931
 URL: https://issues.apache.org/jira/browse/HIVE-4931
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: vectorization-branch
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
 Fix For: vectorization-branch

 Attachments: HIVE-4931.1.patch


 With the dataset attached by Tony, running:
  
 select max(ctinyint), max(csmallint), max(cint), max(cbigint), max(cfloat), 
 max(cdouble) from alltypes_orc;
  
 this is the result in rowmode:
 OK
 62  16379   1073737829  2146661184  62.016376.0
  
 This is the result in vector mode:
 OK
 1   16379   1071926788  -106115134  62.016361.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4931) MAX/MIN aggregates yield incorrect results

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721688#comment-13721688
 ] 

Edward Capriolo commented on HIVE-4931:
---

{code}
-@Override
+@Override
{code}
And ^ means the last commit on this file went in without check-style.

 MAX/MIN aggregates yield incorrect results
 --

 Key: HIVE-4931
 URL: https://issues.apache.org/jira/browse/HIVE-4931
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: vectorization-branch
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
 Fix For: vectorization-branch

 Attachments: HIVE-4931.1.patch


 With the dataset attached by Tony, running:
  
 select max(ctinyint), max(csmallint), max(cint), max(cbigint), max(cfloat), 
 max(cdouble) from alltypes_orc;
  
 this is the result in rowmode:
 OK
 62  16379   1073737829  2146661184  62.016376.0
  
 This is the result in vector mode:
 OK
 1   16379   1071926788  -106115134  62.016361.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HIVE-4931) MAX/MIN aggregates yield incorrect results

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4931?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo reopened HIVE-4931:
---


 MAX/MIN aggregates yield incorrect results
 --

 Key: HIVE-4931
 URL: https://issues.apache.org/jira/browse/HIVE-4931
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: vectorization-branch
Reporter: Remus Rusanu
Assignee: Remus Rusanu
Priority: Minor
 Fix For: vectorization-branch

 Attachments: HIVE-4931.1.patch


 With the dataset attached by Tony, running:
  
 select max(ctinyint), max(csmallint), max(cint), max(cbigint), max(cfloat), 
 max(cdouble) from alltypes_orc;
  
 this is the result in rowmode:
 OK
 62  16379   1073737829  2146661184  62.016376.0
  
 This is the result in vector mode:
 OK
 1   16379   1071926788  -106115134  62.016361.0

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-4936) PTest increase performance by parallelizing rsync

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland resolved HIVE-4936.


Resolution: Duplicate

 PTest increase performance by parallelizing rsync
 -

 Key: HIVE-4936
 URL: https://issues.apache.org/jira/browse/HIVE-4936
 Project: Hive
  Issue Type: Bug
Reporter: Brock Noland
Assignee: Brock Noland

 Currently the bottom neck in PTest2 when adding hosts is the rsync to the 
 slaves. Today if we have 3 threads which will execute on a slave we rsync 
 the source from the master to the slave 3 times. This causes the process to 
 be CPU bound on the master.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-4920:
---

Description: 
We should handle spot price increases more gracefully and parallelize rsync to 
slaves better

NO PRECOMMIT TESTS

  was:We should handle spot price increases more gracefully and parallelize 
rsync to slaves better


 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better
 NO PRECOMMIT TESTS

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-4920:
---

Description: We should handle spot price increases more gracefully and 
parallelize rsync to slaves better  (was: Today the price for m1.xlarge 
instances has been varying dramatically. We should fall back on c1.xlarge 
(which is more powerful and is cheaper at present) and then on on-demand 
instances.)
Summary: PTest2 handle Spot Price increases gracefully and improve 
rsync paralllelsim  (was: PTest2 spot instances should fall back on c1.xlarge 
and then on-demand instances)

 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721694#comment-13721694
 ] 

Brock Noland commented on HIVE-4920:


Also this change allows us to go much wider (16 slaves hosts) with linear 
increases in test time (2 hours - 1 hour).

 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: HIVE-4920.patch, Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better
 NO PRECOMMIT TESTS

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-4920:
---

Status: Patch Available  (was: Open)

 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: HIVE-4920.patch, Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better
 NO PRECOMMIT TESTS

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland updated HIVE-4920:
---

Attachment: HIVE-4920.patch

Alright last non-critical fix to PTest2 :)

I have tested this code a number of times including terminating all spot 
instances during a test and it worked quite well.

 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: HIVE-4920.patch, Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better
 NO PRECOMMIT TESTS

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4051) Hive's metastore suffers from 1+N queries when querying partitions is slow

2013-07-27 Thread Phabricator (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4051?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721696#comment-13721696
 ] 

Phabricator commented on HIVE-4051:
---

brock has commented on the revision HIVE-4051 [jira] Hive's metastore suffers 
from 1+N queries when querying partitions  is slow.

  Comments about handling of error.

INLINE COMMENTS
  metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java:2074 In 
general I am supportive of this patch but we should not be catching Throwable 
here. We should at most catch Exceptuon.

  Additionally log4j now supports passing exceptions directly, the 
stringifyException is no longer needed, you can pass the exception as the 
second arg of the log method.

REVISION DETAIL
  https://reviews.facebook.net/D11805

To: JIRA, sershe
Cc: brock


 Hive's metastore suffers from 1+N queries when querying partitions  is slow
 

 Key: HIVE-4051
 URL: https://issues.apache.org/jira/browse/HIVE-4051
 Project: Hive
  Issue Type: Bug
  Components: Clients, Metastore
 Environment: RHEL 6.3 / EC2 C1.XL
Reporter: Gopal V
Assignee: Sergey Shelukhin
 Attachments: HIVE-4051.D11805.1.patch, HIVE-4051.D11805.2.patch


 Hive's query client takes a long time to initialize  start planning queries 
 because of delays in creating all the MTable/MPartition objects.
 For a hive db with 1800 partitions, the metastore took 6-7 seconds to 
 initialize - firing approximately 5900 queries to the mysql database.
 Several of those queries fetch exactly one row to create a single object on 
 the client.
 The following 12 queries were repeated for each partition, generating a storm 
 of SQL queries 
 {code}
 4 Query SELECT 
 `A0`.`SD_ID`,`B0`.`INPUT_FORMAT`,`B0`.`IS_COMPRESSED`,`B0`.`IS_STOREDASSUBDIRECTORIES`,`B0`.`LOCATION`,`B0`.`NUM_BUCKETS`,`B0`.`OUTPUT_FORMAT`,`B0`.`SD_ID`
  FROM `PARTITIONS` `A0` LEFT OUTER JOIN `SDS` `B0` ON `A0`.`SD_ID` = 
 `B0`.`SD_ID` WHERE `A0`.`PART_ID` = 3945
 4 Query SELECT `A0`.`CD_ID`,`B0`.`CD_ID` FROM `SDS` `A0` LEFT OUTER JOIN 
 `CDS` `B0` ON `A0`.`CD_ID` = `B0`.`CD_ID` WHERE `A0`.`SD_ID` =4871
 4 Query SELECT COUNT(*) FROM `COLUMNS_V2` THIS WHERE THIS.`CD_ID`=1546 
 AND THIS.`INTEGER_IDX`=0
 4 Query SELECT 
 `A0`.`COMMENT`,`A0`.`COLUMN_NAME`,`A0`.`TYPE_NAME`,`A0`.`INTEGER_IDX` AS 
 NUCORDER0 FROM `COLUMNS_V2` `A0` WHERE `A0`.`CD_ID` = 1546 AND 
 `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT `A0`.`SERDE_ID`,`B0`.`NAME`,`B0`.`SLIB`,`B0`.`SERDE_ID` 
 FROM `SDS` `A0` LEFT OUTER JOIN `SERDES` `B0` ON `A0`.`SERDE_ID` = 
 `B0`.`SERDE_ID` WHERE `A0`.`SD_ID` =4871
 4 Query SELECT COUNT(*) FROM `SORT_COLS` THIS WHERE THIS.`SD_ID`=4871 AND 
 THIS.`INTEGER_IDX`=0
 4 Query SELECT `A0`.`COLUMN_NAME`,`A0`.`ORDER`,`A0`.`INTEGER_IDX` AS 
 NUCORDER0 FROM `SORT_COLS` `A0` WHERE `A0`.`SD_ID` =4871 AND 
 `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT COUNT(*) FROM `SKEWED_VALUES` THIS WHERE 
 THIS.`SD_ID_OID`=4871 AND THIS.`INTEGER_IDX`=0
 4 Query SELECT 'org.apache.hadoop.hive.metastore.model.MStringList' AS 
 NUCLEUS_TYPE,`A1`.`STRING_LIST_ID`,`A0`.`INTEGER_IDX` AS NUCORDER0 FROM 
 `SKEWED_VALUES` `A0` INNER JOIN `SKEWED_STRING_LIST` `A1` ON 
 `A0`.`STRING_LIST_ID_EID` = `A1`.`STRING_LIST_ID` WHERE `A0`.`SD_ID_OID` 
 =4871 AND `A0`.`INTEGER_IDX` = 0 ORDER BY NUCORDER0
 4 Query SELECT COUNT(*) FROM `SKEWED_COL_VALUE_LOC_MAP` WHERE `SD_ID` 
 =4871 AND `STRING_LIST_ID_KID` IS NOT NULL
 4 Query SELECT 'org.apache.hadoop.hive.metastore.model.MStringList' AS 
 NUCLEUS_TYPE,`A0`.`STRING_LIST_ID` FROM `SKEWED_STRING_LIST` `A0` INNER JOIN 
 `SKEWED_COL_VALUE_LOC_MAP` `B0` ON `A0`.`STRING_LIST_ID` = 
 `B0`.`STRING_LIST_ID_KID` WHERE `B0`.`SD_ID` =4871
 4 Query SELECT `A0`.`STRING_LIST_ID_KID`,`A0`.`LOCATION` FROM 
 `SKEWED_COL_VALUE_LOC_MAP` `A0` WHERE `A0`.`SD_ID` =4871 AND NOT 
 (`A0`.`STRING_LIST_ID_KID` IS NULL)
 {code}
 This data is not detached or cached, so this operation is performed during 
 every query plan for the partitions, even in the same hive client.
 The queries are automatically generated by JDO/DataNucleus which makes it 
 nearly impossible to rewrite it into a single denormalized join operation  
 process it locally.
 Attempts to optimize this with JDO fetch-groups did not bear fruit in 
 improving the query count.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-4338) Create Table fails after upgrade from 0.9 to 0.10

2013-07-27 Thread Brock Noland (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4338?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brock Noland resolved HIVE-4338.


Resolution: Not A Problem

Agreed with Harsh J, it looks like the upgrade script was not run. Closing.

 Create Table fails after upgrade from 0.9 to 0.10
 -

 Key: HIVE-4338
 URL: https://issues.apache.org/jira/browse/HIVE-4338
 Project: Hive
  Issue Type: Bug
  Components: Database/Schema
Affects Versions: 0.10.0
 Environment: Ubuntu  3.2.0-23-generic #36-Ubuntu on AMD
Reporter: Geula Vainappel

 I ran apt-get upgrade on a relatively old cdh installation. Many things were 
 upgraded, among them hadoop, hdfs and hive (from 0.9 to 0.10). After the 
 upgrade, CREATE TABLE started failing. I rebooted the machine, and it is 
 still not working. The error I am receiving is:
 hive create table ttt(line string);
 FAILED: Error in metadata: 
 MetaException(message:javax.jdo.JDODataStoreException: Error(s) were found 
 while auto-creating/validating the datastore for classes. The errors are 
 printed in the log, and are attached to this exception.
 NestedThrowables:
 java.sql.SQLSyntaxErrorException: In an ALTER TABLE statement, the column 
 'IS_STOREDASSUBDIRECTORIES' has been specified as NOT NULL and either the 
 DEFAULT clause was not specified or was specified as DEFAULT NULL.)
 FAILED: Execution Error, return code 1 from 
 org.apache.hadoop.hive.ql.exec.DDLTask

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4941) PTest2 Investigate Ignores

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721697#comment-13721697
 ] 

Brock Noland commented on HIVE-4941:


TestJdbcDriver2 ran just fine, I have enabled it.

 PTest2 Investigate Ignores
 --

 Key: HIVE-4941
 URL: https://issues.apache.org/jira/browse/HIVE-4941
 Project: Hive
  Issue Type: Task
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor

 Currently we excluding the following tests:
 unitTests.exclude = TestHiveMetaStore TestSerDe TestBeeLineDriver 
 TestHiveServer2Concurrency TestJdbcDriver2 TestHiveServer2Concurrency 
 TestBeeLineDriver
 some of them we got from the build files but I am not sure about 
 TestJdbcDriver2 for example. We should investigate why these are excluded.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4947) Add myself to committer list

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4947?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-4947:
-

Attachment: HIVE-4947.2.patch

Sure. Added m6d as your organization.

 Add myself to committer list
 

 Key: HIVE-4947
 URL: https://issues.apache.org/jira/browse/HIVE-4947
 Project: Hive
  Issue Type: Task
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4947.2.patch, HIVE-4947.patch


 Update website to include myself.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4947) Add myself to committer list

2013-07-27 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721700#comment-13721700
 ] 

Gunther Hagleitner commented on HIVE-4947:
--

[~brocknoland] I've added you too, since I didn't see you in there. Can you 
confirm I got the entry right?

 Add myself to committer list
 

 Key: HIVE-4947
 URL: https://issues.apache.org/jira/browse/HIVE-4947
 Project: Hive
  Issue Type: Task
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4947.2.patch, HIVE-4947.patch


 Update website to include myself.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


Re: [Discuss] project chop up

2013-07-27 Thread Mohammad Islam


Yes hive build and test cases got convoluted as the project scope gradually 
increased. This is the time to take action!

Based on my other Apache experiences, I prefer the option #3 Breakup the 
projects within our own source tree. Make multiple modules or sub-projects. By 
default, only key modules will be built. 

Maven could be a possible candidate.

Regards,
Mohammad




 From: Edward Capriolo edlinuxg...@gmail.com
To: dev@hive.apache.org dev@hive.apache.org 
Sent: Saturday, July 27, 2013 7:03 AM
Subject: Re: [Discuss] project chop up
 

Or feel free to suggest different approach. I am used to managing software
as multi-module maven projects.
From a development standpoint if I was working on beeline, it would be nice
to only require some of the sub-projects to be open in my IDE to do that.
Also managing everything globally is not ideal.

Hive's project layout, build, and test infrastructure is just funky. It has
to do a few interesting things (shims, testing), but I do not think what we
are doing justifies the massive ant build system we have. Ant is so ten
years ago.



On Sat, Jul 27, 2013 at 12:04 AM, Alan Gates ga...@hortonworks.com wrote:

 But I assume they'd still be a part of targets like package, tar, and
 binary?  Making them compile and test separately and explicitly load the
 core Hive jars from maven/ivy seems reasonable.

 Alan.

 On Jul 26, 2013, at 8:40 PM, Brock Noland wrote:

  Hi,
 
  I think thats part of it but I'd like to decouple the downstream projects
  even further so that the only connection is the dependency on the hive
 jars.
 
  Brock
  On Jul 26, 2013 10:10 PM, Alan Gates ga...@hortonworks.com wrote:
 
  I'm not sure how this is different from what hcat does today.  It needs
  Hive's jars to compile, so it's one of the last things in the compile
 step.
  Would moving the other modules you note to be in the same category be
  enough?  Did you want to also make it so that the default ant target
  doesn't compile those?
 
  Alan.
 
  On Jul 26, 2013, at 4:09 PM, Edward Capriolo wrote:
 
  My mistake on saying hcat was a fork metastore. I had a brain fart for
 a
  moment.
 
  One way we could do this is create a folder called downstream. In our
  release step we can execute the downstream builds and then copy the
 files
  we need back. So nothing downstream will be on the classpath of the
 main
  project.
 
  This could help us breakup ql as well. Things like exotic file formats
 ,
  and things that are pluggable like zk locking can go here. That might
 be
  overkill.
 
  For now we can focus on building downstream and hivethrift1might be the
  first thing to try to downstream.
 
 
  On Friday, July 26, 2013, Thejas Nair the...@hortonworks.com wrote:
  +1 to the idea of making the build of core hive and other downstream
  components independent.
 
  bq.  I was under the impression that Hcat and hive-metastore was
  supposed to merge up somehow.
 
  The metastore code was never forked. Hcat was just using
  hive-metastore and making the metadata available to rest of hadoop
  (pig, java MR..).
  A lot of the changes that were driven by hcat goals were being made in
  hive-metastore. You can think of hcat as set of libraries that let pig
  and java MR use hive metastore. Since hcat is closely tied to
  hive-metastore, it makes sense to have them in same project.
 
 
  On Fri, Jul 26, 2013 at 6:33 AM, Edward Capriolo 
 edlinuxg...@gmail.com
 
  wrote:
  Also i believe hcatalog web can fall into the same designation.
 
  Question , hcatalog was initily a big hive-metastore fork. I was
 under
  the
  impression that Hcat and hive-metastore was supposed to merge up
  somehow.
  What is the status on that? I remember that was one of the core
 reasons
  we
  brought it in.
 
  On Friday, July 26, 2013, Edward Capriolo edlinuxg...@gmail.com
  wrote:
  I prefer option 3 as well.
 
 
  On Fri, Jul 26, 2013 at 12:52 AM, Brock Noland br...@cloudera.com
  wrote:
 
  On Thu, Jul 25, 2013 at 9:48 PM, Edward Capriolo 
  edlinuxg...@gmail.com
  wrote:
 
  I have been developing my laptop on a duel core 2 GB Ram laptop
 for
  years
  now. With the addition of hcatalog, hive-thrift2, and some other
  growth
  trying to develop hive in a eclipse on this machine craws,
  especially
  if
  'build automatically' is turned on. As we look to add on more
 things
  this
  is only going to get worse.
 
  I am also noticing issues like this:
 
  https://issues.apache.org/jira/browse/HIVE-4849
 
  What I think we should do is strip down/out optional parts of
 hive.
 
  1) Hive Hbase
  This should really be it's own project to do this right we really
  have to
  have multiple branches since hbase is not backwards compatible.
 
  2) Hive Web Interface
  Now really a big project but not really critical can be just as
  easily
  be
  build separately
 
  3) hive thrift 1
  We have hive thrift 2 now, it is time for the sun to set on
  hivethrift1,
 
  4) odbc
  Not entirely 

[jira] [Updated] (HIVE-3325) serde-reported partition cols should not be persisted in metastore

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3325?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-3325:
--

Resolution: Fixed
Status: Resolved  (was: Patch Available)

This was committed on


r1505418 | ecapriolo | 2013-07-21 12:50:25 -0400 (Sun, 21 Jul 2013) | 5 lines

Serde-reported partition cols should not be persisted in metastore (Travis 
Crawford via egc)

Sorry that I did not close out this ticket already. Thanks Travis.

 serde-reported partition cols should not be persisted in metastore
 --

 Key: HIVE-3325
 URL: https://issues.apache.org/jira/browse/HIVE-3325
 Project: Hive
  Issue Type: Bug
  Components: Metastore
Affects Versions: 0.10.0
Reporter: Travis Crawford
Assignee: Travis Crawford
 Attachments: HIVE-3325.patch.txt, 
 HIVE-3325_serde_reported_part_cols.1.patch


 In HIVE-3279 an issue was fixed where serde-reported columns were not added 
 to partitions. However, the fix in that issue caused serde-reported columns 
 to be stored in the partition storage descriptor.
 Serde-reported columns should be dynamic and only reported at runtime (not 
 stored in the partition storage descriptor).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4920) PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4920?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721732#comment-13721732
 ] 

Edward Capriolo commented on HIVE-4920:
---

+1. I will review a bit more and commit in 24 hours.

 PTest2 handle Spot Price increases gracefully and improve rsync paralllelsim
 

 Key: HIVE-4920
 URL: https://issues.apache.org/jira/browse/HIVE-4920
 Project: Hive
  Issue Type: Improvement
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Critical
 Attachments: HIVE-4920.patch, Screen Shot 2013-07-23 at 3.35.00 PM.png


 We should handle spot price increases more gracefully and parallelize rsync 
 to slaves better
 NO PRECOMMIT TESTS

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo reassigned HIVE-4299:
-

Assignee: Edward Capriolo  (was: Sho Shimauchi)

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4299:
--

Attachment: HIVE-4299.4.patch.txt

Changed name to . from _. Also date format is not thread safe, moved it inside 
the method just to be safe.

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


hive windowing should not have been committed in it's current state

2013-07-27 Thread Edward Capriolo
I was looking through the function registery today and I noticed this:

UDAFS_IMPLY_ORDER.add(rank);
UDAFS_IMPLY_ORDER.add(dense_rank);
UDAFS_IMPLY_ORDER.add(percent_rank);
UDAFS_IMPLY_ORDER.add(cume_dist);
UDAFS_IMPLY_ORDER.add(LEAD_FUNC_NAME);
UDAFS_IMPLY_ORDER.add(LAG_FUNC_NAME);
UDAFS_IMPLY_ORDER.add(first_value);
UDAFS_IMPLY_ORDER.add(last_value);

There are several things to point out first.

1) I can not follow why some method names are constant and some are not.
2) A hard coded list of UDAFS_IMPLY_ORDER seems to indicate that a window
function of this type can not be added.
3) Every other container in function registry is a synchronized collection,
but this is not
-  /*
-   * UDAFS that only work when the input rows have an order.
-   */
-  public static final HashSetString UDAFS_IMPLY_ORDER = new
HashSetString();

4) Our style rules were not followed '{' sometimes on the line
sometimes tab, sometimes two spaces, sometimes 8 tabs,

  public static void registerTableFunction(String name, Class? extends
TableFunctionResolver tFnCls)
  {
FunctionInfo tInfo = new FunctionInfo(name, tFnCls);
mFunctions.put(name.toLowerCase(), tInfo);
  }

I have opened up https://issues.apache.org/jira/browse/HIVE-4879 to address
these things.

Personally I am not a crazy stickler for check-style and I rarely kick a
patch back at someone for a couple of style violations, but this really
needed the once over. Especially since we just spent a lot of effort
cleaning up un thread safe things so hiveserver2 can be happy.

Keep in mind that patches  with flagrant check style violations, no tests,
and little javadoc should be a no go.

I am adding an annotation for the classes so we do not need to maintain a
static list.
Edward


[jira] [Commented] (HIVE-4586) [HCatalog] WebHCat should return 404 error for undefined resource

2013-07-27 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4586?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721754#comment-13721754
 ] 

Eugene Koifman commented on HIVE-4586:
--

That test suite was added after Daniel made the patch

 [HCatalog] WebHCat should return 404 error for undefined resource
 -

 Key: HIVE-4586
 URL: https://issues.apache.org/jira/browse/HIVE-4586
 Project: Hive
  Issue Type: Bug
  Components: HCatalog
Affects Versions: 0.11.0
Reporter: Daniel Dai
Assignee: Daniel Dai
 Fix For: 0.12.0

 Attachments: HIVE-4586-1.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HIVE-4055) add Date data type

2013-07-27 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan resolved HIVE-4055.


   Resolution: Fixed
Fix Version/s: 0.12.0

Committed to trunk. Thanks, Sun for initial work and Jason for following it up 
to see through this completion.

 add Date data type
 --

 Key: HIVE-4055
 URL: https://issues.apache.org/jira/browse/HIVE-4055
 Project: Hive
  Issue Type: Sub-task
  Components: JDBC, Query Processor, Serializers/Deserializers, UDF
Reporter: Sun Rui
Assignee: Jason Dere
 Fix For: 0.12.0

 Attachments: Date.pdf, HIVE-4055.1.patch.txt, HIVE-4055.2.patch.txt, 
 HIVE-4055.3.patch.txt, HIVE-4055.4.patch, HIVE-4055.4.patch.txt, 
 HIVE-4055.D11547.1.patch


 Add Date data type, a new primitive data type which supports the standard SQL 
 date type.
 Basically, the implementation can take HIVE-2272 and HIVE-2957 as references.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721768#comment-13721768
 ] 

Hive QA commented on HIVE-4299:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594554/HIVE-4299.4.patch.txt

{color:red}ERROR:{color} -1 due to 50 failed/errored test(s), 2723 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_02_all_part_over_overlap
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_09_part_spec_nonoverlap
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_23_import_exist_authfail
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_04_nonpart_noncompat_colnumber
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_00_nonpart_empty
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_17_part_spec_underspec
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_16_part_noncompat_schema
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_18_part_spec_missing
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_16_part_external
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_06_nonpart_noncompat_storage
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_17_part_managed
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_10_nonpart_noncompat_bucketing
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_19_00_part_external_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_23_import_part_authsuccess
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_13_nonnative_import
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_10_external_managed
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_02_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_22_import_exist_authsuccess
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_25_import_nonexist_authfail
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_14_nonpart_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_07_all_part_over_nonoverlap
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_14_managed_location_over_existing
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_12_external_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_02_00_part_empty
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_08_nonpart_noncompat_serde
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_09_nonpart_noncompat_serdeparam
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_24_import_part_authfail
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_20_managed_location_over_existing
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_19_part_external_location
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_15_part_nonpart
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_03_nonpart_noncompat_colschema
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_07_nonpart_noncompat_ifof
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_11_nonpart_noncompat_sorting
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_20_part_managed_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_03_nonpart_over_compat
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_01_nonpart_over_loaded
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_04_evolved_parts
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_21_part_managed_external
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testNegativeCliDriver_exim_05_nonpart_noncompat_coltype
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_18_part_external
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_04_all_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_06_one_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_24_import_nonexist_authsuccess
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_01_nonpart
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_13_managed_location
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_11_managed_external
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_15_external_part
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_08_nonpart_rename
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_exim_05_some_part

[jira] [Commented] (HIVE-4941) PTest2 Investigate Ignores

2013-07-27 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721772#comment-13721772
 ] 

Ashutosh Chauhan commented on HIVE-4941:


In recent runs, I noticed Hive QA runs around ~2653 tests while when I run my 
tests old style (ie via ant test) it reports running ~2692 tests. It will be 
good to verify the same number of tests are run in either cases.

 PTest2 Investigate Ignores
 --

 Key: HIVE-4941
 URL: https://issues.apache.org/jira/browse/HIVE-4941
 Project: Hive
  Issue Type: Task
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor

 Currently we excluding the following tests:
 unitTests.exclude = TestHiveMetaStore TestSerDe TestBeeLineDriver 
 TestHiveServer2Concurrency TestJdbcDriver2 TestHiveServer2Concurrency 
 TestBeeLineDriver
 some of them we got from the build files but I am not sure about 
 TestJdbcDriver2 for example. We should investigate why these are excluded.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4941) PTest2 Investigate Ignores

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721775#comment-13721775
 ] 

Brock Noland commented on HIVE-4941:


Will do

 PTest2 Investigate Ignores
 --

 Key: HIVE-4941
 URL: https://issues.apache.org/jira/browse/HIVE-4941
 Project: Hive
  Issue Type: Task
Reporter: Brock Noland
Assignee: Brock Noland
Priority: Minor

 Currently we excluding the following tests:
 unitTests.exclude = TestHiveMetaStore TestSerDe TestBeeLineDriver 
 TestHiveServer2Concurrency TestJdbcDriver2 TestHiveServer2Concurrency 
 TestBeeLineDriver
 some of them we got from the build files but I am not sure about 
 TestJdbcDriver2 for example. We should investigate why these are excluded.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721774#comment-13721774
 ] 

Brock Noland commented on HIVE-4299:


At least one failure is:
{noformat}
 [junit] junit.framework.AssertionFailedError: Client Execution failed with 
error code = 10027
{noformat}

which is invalid path.

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4929) the type of all numeric constants is changed to double in the plan

2013-07-27 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-4929:
---

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Thanks, Sergey!

 the type of all numeric constants is changed to double in the plan
 --

 Key: HIVE-4929
 URL: https://issues.apache.org/jira/browse/HIVE-4929
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.12.0

 Attachments: HIVE-4929.patch


 There's code which, after the numeric type for a constant in where clause has 
 been chosen as the most restricted one or based on suffix, tries to change 
 the type to match the numeric column  which the constant is being compared 
 with. However, due to a hack from HIVE-3059 every column type shows up as 
 string in that code, causing it to always change the constant type to double. 
 This should not be done (regardless of the hack).
 Spinoff from HIVE-2702, large number of query outputs change so it will be a 
 big patch

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4825) Separate MapredWork into MapWork and ReduceWork

2013-07-27 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721779#comment-13721779
 ] 

Ashutosh Chauhan commented on HIVE-4825:


+1

 Separate MapredWork into MapWork and ReduceWork
 ---

 Key: HIVE-4825
 URL: https://issues.apache.org/jira/browse/HIVE-4825
 Project: Hive
  Issue Type: Improvement
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4825.1.patch, HIVE-4825.2.code.patch, 
 HIVE-4825.2.testfiles.patch, HIVE-4825.3.testfiles.patch, HIVE-4825.4.patch, 
 HIVE-4825.5.patch, HIVE-4825.6.patch


 Right now all the information needed to run an MR job is captured in 
 MapredWork. This class has aliases, tagging info, table descriptors etc.
 For Tez and MRR it will be useful to break this into map and reduce specific 
 pieces. The separation is natural and I think has value in itself, it makes 
 the code easier to understand. However, it will also allow us to reuse these 
 abstractions in Tez where you'll have a graph of these instead of just 1M and 
 0-1R.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4520) java.lang.NegativeArraySizeException when query on hive-0.11.0, hbase-0.94.6.1

2013-07-27 Thread Swarnim Kulkarni (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721783#comment-13721783
 ] 

Swarnim Kulkarni commented on HIVE-4520:


[~mayanhui] This bug seems to be a duplicate of HIVE-4515. Do you agree to 
resolve this one as duplicate?

 java.lang.NegativeArraySizeException when query on hive-0.11.0, hbase-0.94.6.1
 --

 Key: HIVE-4520
 URL: https://issues.apache.org/jira/browse/HIVE-4520
 Project: Hive
  Issue Type: Bug
  Components: HBase Handler
Affects Versions: 0.11.0
 Environment: hive-0.11.0
 hbase-0.94.6.1
 zookeeper-3.4.3
 hadoop-1.0.4
 centos-5.7
Reporter: Yanhui Ma
Priority: Critical

 After integration hive-0.11.0+hbase-0.94.6.1, these commands could be 
 executed sucessfully:
 create table
 insert overwrite table
 select * from table
 However, when execute select count(*) from table, throws exception:
 hive select count(*) from test; 
 Total MapReduce jobs = 1
 Launching Job 1 out of 1
 Number of reduce tasks determined at compile time: 1
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapred.reduce.tasks=number
 Starting Job = job_201305061042_0028, Tracking URL = 
 http://master0:50030/jobdetails.jsp?jobid=job_201305061042_0028
 Kill Command = /opt/modules/hadoop/hadoop-1.0.4/libexec/../bin/hadoop job  
 -kill job_201305061042_0028
 Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 
 1
 2013-05-07 18:41:42,649 Stage-1 map = 0%,  reduce = 0%
 2013-05-07 18:42:14,789 Stage-1 map = 100%,  reduce = 100%
 Ended Job = job_201305061042_0028 with errors
 Error during job, obtaining debugging information...
 Job Tracking URL: 
 http://master0:50030/jobdetails.jsp?jobid=job_201305061042_0028
 Examining task ID: task_201305061042_0028_m_02 (and more) from job 
 job_201305061042_0028
 Task with the most failures(4): 
 -
 Task ID:
   task_201305061042_0028_m_00
 URL:
   
 http://master0:50030/taskdetails.jsp?jobid=job_201305061042_0028tipid=task_201305061042_0028_m_00
 -
 Diagnostic Messages for this Task:
 java.lang.NegativeArraySizeException: -1
   at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:148)
   at 
 org.apache.hadoop.hbase.mapreduce.TableSplit.readFields(TableSplit.java:133)
   at 
 org.apache.hadoop.hive.hbase.HBaseSplit.readFields(HBaseSplit.java:53)
   at 
 org.apache.hadoop.hive.ql.io.HiveInputFormat$HiveInputSplit.readFields(HiveInputFormat.java:150)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
   at org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:396)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:412)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.MapRedTask
 MapReduce Jobs Launched: 
 Job 0: Map: 1  Reduce: 1   HDFS Read: 0 HDFS Write: 0 FAIL
 Total MapReduce CPU Time Spent: 0 msec
 ==
 The log of tasktracker:
 stderr logs
 13/05/07 18:43:20 INFO util.NativeCodeLoader: Loaded the native-hadoop library
 13/05/07 18:43:20 INFO mapred.TaskRunner: Creating symlink: 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/distcache/107328478296390_-1298160740_2123690974/master0/tmp/hive-hadoop/hive_2013-05-07_18-41-30_290_832140779606816147/-mr-10003/fd22448b-e923-498c-bc00-2164ca68447d
  - 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/attempt_201305061042_0028_m_00_0/work/HIVE_PLANfd22448b-e923-498c-bc00-2164ca68447d
 13/05/07 18:43:20 INFO filecache.TrackerDistributedCacheManager: Creating 
 symlink: 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/jars/javolution
  - 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/attempt_201305061042_0028_m_00_0/work/javolution
 13/05/07 18:43:20 INFO filecache.TrackerDistributedCacheManager: 

[jira] [Resolved] (HIVE-4947) Add myself to committer list

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4947?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner resolved HIVE-4947.
--

Resolution: Fixed

 Add myself to committer list
 

 Key: HIVE-4947
 URL: https://issues.apache.org/jira/browse/HIVE-4947
 Project: Hive
  Issue Type: Task
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4947.2.patch, HIVE-4947.patch


 Update website to include myself.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4947) Add myself to committer list

2013-07-27 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721785#comment-13721785
 ] 

Gunther Hagleitner commented on HIVE-4947:
--

Committed to hive-site. Thanks!

 Add myself to committer list
 

 Key: HIVE-4947
 URL: https://issues.apache.org/jira/browse/HIVE-4947
 Project: Hive
  Issue Type: Task
Reporter: Gunther Hagleitner
Assignee: Gunther Hagleitner
Priority: Minor
 Attachments: HIVE-4947.2.patch, HIVE-4947.patch


 Update website to include myself.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721787#comment-13721787
 ] 

Edward Capriolo commented on HIVE-4299:
---

Lets try again with the _ then.

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4299:
--

Attachment: HIVE-4299.5.patch.txt

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.5.patch.txt, HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721808#comment-13721808
 ] 

Hive QA commented on HIVE-4299:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594564/HIVE-4299.5.patch.txt

{color:red}ERROR:{color} -1 due to 56 failed/errored test(s), 2736 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testDriverProperties
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testDatabaseMetaData
org.apache.hive.jdbc.TestJdbcDriver2.testSelectAll
org.apache.hive.jdbc.TestJdbcDriver2.testResultSetMetaData
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetColumnsMetaData
org.apache.hive.jdbc.TestJdbcDriver2.testDescribeTable
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testSelectAllFetchSize
org.apache.hive.jdbc.TestJdbcDriver2.testNullType
org.apache.hive.jdbc.TestJdbcDriver2.testDuplicateColumnNameOrder
org.apache.hive.jdbc.TestJdbcDriver2.testProccedures
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testSelectAllMaxRows
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testErrorMessages
org.apache.hive.jdbc.TestJdbcDriver2.testSelectAllFetchSize
org.apache.hive.jdbc.TestJdbcDriver2.testDataTypes2
org.apache.hive.jdbc.TestJdbcDriver2.testOutOfBoundCols
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testSelectAll
org.apache.hive.jdbc.TestJdbcDriver2.testDriverProperties
org.apache.hive.jdbc.TestJdbcDriver2.testErrorDiag
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testNullType
org.apache.hive.jdbc.TestJdbcDriver2.testProcCols
org.apache.hive.jdbc.TestJdbcDriver2.testBuiltInUDFCol
org.apache.hive.jdbc.TestJdbcDriver2.testPostClose
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetColumnsMetaData
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetTableTypes
org.apache.hive.jdbc.TestJdbcDriver2.testErrorMessages
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testSelectAllPartioned
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetColumns
org.apache.hive.jdbc.TestJdbcDriver2.testPrimaryKeys
org.apache.hive.jdbc.TestJdbcDriver2.testShowTables
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetTables
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testResultSetMetaData
org.apache.hive.jdbc.TestJdbcDriver2.testExplainStmt
org.apache.hive.jdbc.TestJdbcDriver2.testBadURL
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetCatalogs
org.apache.hive.jdbc.TestJdbcDriver2.testImportedKeys
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testShowTables
org.apache.hive.jdbc.TestJdbcDriver2.testPrepareStatement
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetCatalogs
org.apache.hive.jdbc.TestJdbcDriver2.testInvalidURL
org.apache.hive.jdbc.TestJdbcDriver2.testSetCommand
org.apache.hive.jdbc.TestJdbcDriver2.testSelectAllMaxRows
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testExplainStmt
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testDescribeTable
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testDataTypes
org.apache.hive.jdbc.TestJdbcDriver2.testDataTypes
org.apache.hive.jdbc.TestJdbcDriver2.testSelectAllPartioned
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testSetCommand
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetTables
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetColumns
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testPrepareStatement
org.apache.hive.jdbc.TestJdbcDriver2.testDatabaseMetaData
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetSchemas
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testMetaDataGetSchemas
org.apache.hadoop.hive.jdbc.TestJdbcDriver.testConversionsBaseResultSet
org.apache.hive.jdbc.TestJdbcDriver2.testExprCol
org.apache.hive.jdbc.TestJdbcDriver2.testMetaDataGetTableTypes
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/211/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/211/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests failed with: TestsFailedException: 56 tests failed
{noformat}

This message is automatically generated.

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.5.patch.txt, HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table 

[jira] [Updated] (HIVE-4879) Window functions that imply order can only be registered at compile time

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4879:
--

Description: Adding an annotation for impliesOrder  (was: It seems like the 
windowing (PTF) commit never used the apache code style. Curly braces in 
particular are not formatted according to standards. Maybe we can bundle a 
style commit up
 
{code}
  for (OrderExpression oExpr : spec.getExpressions())
  {
  OrderExpressionDef oexpDef = translate(inpShape, oExpr);
  def.addExpression(oexpDef);
}
{code})
   Assignee: Edward Capriolo
 Issue Type: Improvement  (was: Task)
Summary: Window functions that imply order can only be registered at 
compile time  (was: Windowing commit is need of a code-style cleanup)

 Window functions that imply order can only be registered at compile time
 

 Key: HIVE-4879
 URL: https://issues.apache.org/jira/browse/HIVE-4879
 Project: Hive
  Issue Type: Improvement
Reporter: Edward Capriolo
Assignee: Edward Capriolo

 Adding an annotation for impliesOrder

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4879) Window functions that imply order can only be registered at compile time

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4879:
--

Attachment: HIVE-4879.1.patch.txt

 Window functions that imply order can only be registered at compile time
 

 Key: HIVE-4879
 URL: https://issues.apache.org/jira/browse/HIVE-4879
 Project: Hive
  Issue Type: Improvement
Reporter: Edward Capriolo
Assignee: Edward Capriolo
 Attachments: HIVE-4879.1.patch.txt


 Adding an annotation for impliesOrder

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-4879) Window functions that imply order can only be registered at compile time

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4879:
--

Fix Version/s: 0.12.0
Affects Version/s: 0.11.0
   Status: Patch Available  (was: Open)

 Window functions that imply order can only be registered at compile time
 

 Key: HIVE-4879
 URL: https://issues.apache.org/jira/browse/HIVE-4879
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.11.0
Reporter: Edward Capriolo
Assignee: Edward Capriolo
 Fix For: 0.12.0

 Attachments: HIVE-4879.1.patch.txt


 Adding an annotation for impliesOrder

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4299) exported metadata by HIVE-3068 cannot be imported because of wrong file name

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4299?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721816#comment-13721816
 ] 

Brock Noland commented on HIVE-4299:


The tests failed with
{noformat}
DATE and DATETIME types aren't supported yet. Please use TIMESTAMP instead at
{noformat}
which is believe is a problem on trunk. Running again to see.

 exported metadata by HIVE-3068 cannot be imported because of wrong file name
 

 Key: HIVE-4299
 URL: https://issues.apache.org/jira/browse/HIVE-4299
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Affects Versions: 0.11.0
Reporter: Sho Shimauchi
Assignee: Edward Capriolo
 Attachments: HIVE-4299.1.patch.txt, HIVE-4299.4.patch.txt, 
 HIVE-4299.5.patch.txt, HIVE-4299.patch


 h2. Symptom
 When DROP TABLE a table, metadata of the table is generated to be able to 
 import the dropped table again.
 However, the exported metadata name is 'table name.metadata'.
 Since ImportSemanticAnalyzer allows only '_metadata' as metadata filename, 
 user have to rename the metadata file to import the table.
 h2. How to reproduce
 Set the following setting to hive-site.xml:
 {code}
  property
namehive.metastore.pre.event.listeners/name
valueorg.apache.hadoop.hive.ql.parse.MetaDataExportListener/value
  /property
 {code}
 Then run the following queries:
 {code}
  CREATE TABLE test_table (id INT, name STRING);
  DROP TABLE test_table;
  IMPORT TABLE test_table_imported FROM '/path/to/metadata/file';
 FAILED: SemanticException [Error 10027]: Invalid path
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-07-27 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721815#comment-13721815
 ] 

Gunther Hagleitner commented on HIVE-3926:
--

[~navis] sorry patch doesn't apply cleanly anymore. I believe the conflict is 
with HIVE-4929. Can you please rebase?

 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: HIVE-3926.D8121.1.patch, HIVE-3926.D8121.2.patch, 
 HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4055) add Date data type

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721817#comment-13721817
 ] 

Brock Noland commented on HIVE-4055:


We appear to be seeing the following failures on trunk

{noformat}
DATE and DATETIME types aren't supported yet. Please use TIMESTAMP instead at
{noformat}

for TestJdbcDriver\{,2\}

 add Date data type
 --

 Key: HIVE-4055
 URL: https://issues.apache.org/jira/browse/HIVE-4055
 Project: Hive
  Issue Type: Sub-task
  Components: JDBC, Query Processor, Serializers/Deserializers, UDF
Reporter: Sun Rui
Assignee: Jason Dere
 Fix For: 0.12.0

 Attachments: Date.pdf, HIVE-4055.1.patch.txt, HIVE-4055.2.patch.txt, 
 HIVE-4055.3.patch.txt, HIVE-4055.4.patch, HIVE-4055.4.patch.txt, 
 HIVE-4055.D11547.1.patch


 Add Date data type, a new primitive data type which supports the standard SQL 
 date type.
 Basically, the implementation can take HIVE-2272 and HIVE-2957 as references.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3255) Add DBTokenStore to store Delegation Tokens in DB

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-3255:
-

   Resolution: Fixed
Fix Version/s: 0.12.0
   Status: Resolved  (was: Patch Available)

 Add DBTokenStore to store Delegation Tokens in DB
 -

 Key: HIVE-3255
 URL: https://issues.apache.org/jira/browse/HIVE-3255
 Project: Hive
  Issue Type: New Feature
  Components: Metastore, Security
Affects Versions: 0.9.0
Reporter: Ashutosh Chauhan
Assignee: Deepesh Khandelwal
 Fix For: 0.12.0

 Attachments: hive-3255_2.patch, hive-3255_3.patch, hive-3255_4.patch, 
 hive-3255_5.patch, HIVE-3255_6.patch, HIVE-3255.6.patch, hive-3255.patch


 Before HIVE-1696 metastore was completely stateless and all the data is in 
 backend db. HIVE-1696 added delegation tokens for metastore which metastore 
 needs to keep in memory. HIVE-2467 added support for storing delegation 
 tokens in ZooKeeper via an interface {{DelegationTokenStore}}. This jira is 
 about using DB backend for storing tokens which metastore uses for storing 
 data. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3255) Add DBTokenStore to store Delegation Tokens in DB

2013-07-27 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3255?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721819#comment-13721819
 ] 

Gunther Hagleitner commented on HIVE-3255:
--

Committed to trunk. Thanks Ashutosh and Deepesh!

 Add DBTokenStore to store Delegation Tokens in DB
 -

 Key: HIVE-3255
 URL: https://issues.apache.org/jira/browse/HIVE-3255
 Project: Hive
  Issue Type: New Feature
  Components: Metastore, Security
Affects Versions: 0.9.0
Reporter: Ashutosh Chauhan
Assignee: Deepesh Khandelwal
 Attachments: hive-3255_2.patch, hive-3255_3.patch, hive-3255_4.patch, 
 hive-3255_5.patch, HIVE-3255_6.patch, HIVE-3255.6.patch, hive-3255.patch


 Before HIVE-1696 metastore was completely stateless and all the data is in 
 backend db. HIVE-1696 added delegation tokens for metastore which metastore 
 needs to keep in memory. HIVE-2467 added support for storing delegation 
 tokens in ZooKeeper via an interface {{DelegationTokenStore}}. This jira is 
 about using DB backend for storing tokens which metastore uses for storing 
 data. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-07-27 Thread Gunther Hagleitner (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721822#comment-13721822
 ] 

Gunther Hagleitner commented on HIVE-3926:
--

Correction - I meant HIVE-4878.

 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: HIVE-3926.D8121.1.patch, HIVE-3926.D8121.2.patch, 
 HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-07-27 Thread Gunther Hagleitner (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gunther Hagleitner updated HIVE-3926:
-

Attachment: HIVE-3926.6.patch

 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: HIVE-3926.6.patch, HIVE-3926.D8121.1.patch, 
 HIVE-3926.D8121.2.patch, HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, 
 HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4055) add Date data type

2013-07-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721829#comment-13721829
 ] 

Brock Noland commented on HIVE-4055:


Something quite strange is going on with the TestJdbcDriver* tests. I cannot 
TestJdbcDriver2 to fail locally but it does fail on ptest: 
https://builds.apache.org/user/brock/my-views/view/hive/job/Hive-trunk-hadoop1-ptest/103/
 I'll take a look tomorrow unless someone else figures it out first.

 add Date data type
 --

 Key: HIVE-4055
 URL: https://issues.apache.org/jira/browse/HIVE-4055
 Project: Hive
  Issue Type: Sub-task
  Components: JDBC, Query Processor, Serializers/Deserializers, UDF
Reporter: Sun Rui
Assignee: Jason Dere
 Fix For: 0.12.0

 Attachments: Date.pdf, HIVE-4055.1.patch.txt, HIVE-4055.2.patch.txt, 
 HIVE-4055.3.patch.txt, HIVE-4055.4.patch, HIVE-4055.4.patch.txt, 
 HIVE-4055.D11547.1.patch


 Add Date data type, a new primitive data type which supports the standard SQL 
 date type.
 Basically, the implementation can take HIVE-2272 and HIVE-2957 as references.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3926) PPD on virtual column of partitioned table is not working

2013-07-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721861#comment-13721861
 ] 

Hive QA commented on HIVE-3926:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12594569/HIVE-3926.6.patch

{color:green}SUCCESS:{color} +1 2730 tests passed

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/214/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/214/console

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

 PPD on virtual column of partitioned table is not working
 -

 Key: HIVE-3926
 URL: https://issues.apache.org/jira/browse/HIVE-3926
 Project: Hive
  Issue Type: Bug
  Components: Query Processor
Reporter: Navis
Assignee: Navis
Priority: Minor
 Attachments: HIVE-3926.6.patch, HIVE-3926.D8121.1.patch, 
 HIVE-3926.D8121.2.patch, HIVE-3926.D8121.3.patch, HIVE-3926.D8121.4.patch, 
 HIVE-3926.D8121.5.patch


 {code}
 select * from src where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 is working, but
 {code}
 select * from srcpart where BLOCK__OFFSET__INSIDE__FILE100;
 {code}
 throws SemanticException. Disabling PPD makes it work.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-4520) java.lang.NegativeArraySizeException when query on hive-0.11.0, hbase-0.94.6.1

2013-07-27 Thread Bryan Peterson (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-4520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721865#comment-13721865
 ] 

Bryan Peterson commented on HIVE-4520:
--

Seems so. The only difference is version numbers for both hive and hbase.

 java.lang.NegativeArraySizeException when query on hive-0.11.0, hbase-0.94.6.1
 --

 Key: HIVE-4520
 URL: https://issues.apache.org/jira/browse/HIVE-4520
 Project: Hive
  Issue Type: Bug
  Components: HBase Handler
Affects Versions: 0.11.0
 Environment: hive-0.11.0
 hbase-0.94.6.1
 zookeeper-3.4.3
 hadoop-1.0.4
 centos-5.7
Reporter: Yanhui Ma
Priority: Critical

 After integration hive-0.11.0+hbase-0.94.6.1, these commands could be 
 executed sucessfully:
 create table
 insert overwrite table
 select * from table
 However, when execute select count(*) from table, throws exception:
 hive select count(*) from test; 
 Total MapReduce jobs = 1
 Launching Job 1 out of 1
 Number of reduce tasks determined at compile time: 1
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapred.reduce.tasks=number
 Starting Job = job_201305061042_0028, Tracking URL = 
 http://master0:50030/jobdetails.jsp?jobid=job_201305061042_0028
 Kill Command = /opt/modules/hadoop/hadoop-1.0.4/libexec/../bin/hadoop job  
 -kill job_201305061042_0028
 Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 
 1
 2013-05-07 18:41:42,649 Stage-1 map = 0%,  reduce = 0%
 2013-05-07 18:42:14,789 Stage-1 map = 100%,  reduce = 100%
 Ended Job = job_201305061042_0028 with errors
 Error during job, obtaining debugging information...
 Job Tracking URL: 
 http://master0:50030/jobdetails.jsp?jobid=job_201305061042_0028
 Examining task ID: task_201305061042_0028_m_02 (and more) from job 
 job_201305061042_0028
 Task with the most failures(4): 
 -
 Task ID:
   task_201305061042_0028_m_00
 URL:
   
 http://master0:50030/taskdetails.jsp?jobid=job_201305061042_0028tipid=task_201305061042_0028_m_00
 -
 Diagnostic Messages for this Task:
 java.lang.NegativeArraySizeException: -1
   at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:148)
   at 
 org.apache.hadoop.hbase.mapreduce.TableSplit.readFields(TableSplit.java:133)
   at 
 org.apache.hadoop.hive.hbase.HBaseSplit.readFields(HBaseSplit.java:53)
   at 
 org.apache.hadoop.hive.ql.io.HiveInputFormat$HiveInputSplit.readFields(HiveInputFormat.java:150)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
   at org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:396)
   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:412)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.MapRedTask
 MapReduce Jobs Launched: 
 Job 0: Map: 1  Reduce: 1   HDFS Read: 0 HDFS Write: 0 FAIL
 Total MapReduce CPU Time Spent: 0 msec
 ==
 The log of tasktracker:
 stderr logs
 13/05/07 18:43:20 INFO util.NativeCodeLoader: Loaded the native-hadoop library
 13/05/07 18:43:20 INFO mapred.TaskRunner: Creating symlink: 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/distcache/107328478296390_-1298160740_2123690974/master0/tmp/hive-hadoop/hive_2013-05-07_18-41-30_290_832140779606816147/-mr-10003/fd22448b-e923-498c-bc00-2164ca68447d
  - 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/attempt_201305061042_0028_m_00_0/work/HIVE_PLANfd22448b-e923-498c-bc00-2164ca68447d
 13/05/07 18:43:20 INFO filecache.TrackerDistributedCacheManager: Creating 
 symlink: 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/jars/javolution
  - 
 /tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201305061042_0028/attempt_201305061042_0028_m_00_0/work/javolution
 13/05/07 18:43:20 INFO filecache.TrackerDistributedCacheManager: Creating 
 symlink: 
 

[jira] [Updated] (HIVE-4879) Window functions that imply order can only be registered at compile time

2013-07-27 Thread Edward Capriolo (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-4879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Capriolo updated HIVE-4879:
--

Attachment: HIVE-4879.2.patch.txt

One more try

 Window functions that imply order can only be registered at compile time
 

 Key: HIVE-4879
 URL: https://issues.apache.org/jira/browse/HIVE-4879
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.11.0
Reporter: Edward Capriolo
Assignee: Edward Capriolo
 Fix For: 0.12.0

 Attachments: HIVE-4879.1.patch.txt, HIVE-4879.2.patch.txt


 Adding an annotation for impliesOrder

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HIVE-3561) Build a full SQL-compliant parser for Hive

2013-07-27 Thread Edward Capriolo (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-3561?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13721883#comment-13721883
 ] 

Edward Capriolo commented on HIVE-3561:
---

I think supporting two query languages is nuts. What is the obsession with SQL 
compliance? I like hive's query language. It has less constructs and that makes 
it easier to read and write queries. Also, knowing hive means learning and 
understanding map reduce you can make things too easy for people and they get 
into anti-patterns.  Having a new parser or two parsers activated by a switch 
sounds not good to me either. 

I really do not see the big win. For the people who can write the complicated 
sub queries that hive can not currently do there are usually smart enough to 
look in the language manual and figure out a way to re-write the query. 



 Build a full SQL-compliant parser for Hive
 --

 Key: HIVE-3561
 URL: https://issues.apache.org/jira/browse/HIVE-3561
 Project: Hive
  Issue Type: Sub-task
  Components: Query Processor
Affects Versions: 0.10.0
Reporter: Shane Huang

 To build a full SQL compliant engine on Hive, we'll need a full SQL complant 
 parser. The current Hive parser missed a lot of grammar units from standard 
 SQL. To support full SQL there're possibly four approaches:
 1.Extend the existing Hive parser to support full SQL constructs. We need to 
 modify the current Hive.g and add any missing grammar units and resolve 
 conflicts. 
 2.Reuse an existing open source SQL compliant parser and extend it to support 
 Hive extensions. We may need to adapt Semantic Analyzers to the new AST 
 structure.  
 3.Reuse an existing SQL compliant parser and make it co-exist with the 
 existing Hive parser. Both parsers share the same CliDriver interface. Use a 
 query mode configuration to switch the query mode between SQL and HQL (this 
 is the approach we're now using in the 0.9.0 demo project)
 4.Reuse an existing SQL compliant parser and make it co-exist with the 
 existing Hive parser. Use a separate xxxCliDriver interface for standard SQL. 
  
 Let's discuss which is the best approach. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira