Re: HBase as input AND output?

2010-10-13 Thread Tim Robertson
That's right. Hive can use an HBase table as an input format to the hive query regardless of output format, and can also write the output to an HBase table regardless of the input format. You can also supposedly do a join in Hive that uses 1 side of the join from an HBase table, and the other

Exception in hive startup

2010-10-13 Thread hdev ml
Hi all, I installed Hadoop 0.20.2 and installed hive 0.5.0. I followed all the instructions on Hive's getting started page for setting up environment variables like HADOOP_HOME When I run from command prompt in the hive installation folder as bin/hive it gives me following exception Exception

Re: HBase as input AND output?

2010-10-13 Thread John Sichi
If your query only accesses HBase tables, then yes, Hive does not access any source data directly from HDFS (although of course it may put intermediate results in HDFS, e.g. for the result of a join). However, if your query does something like join a HBase table with a native Hive table, then

RE: Exception in hive startup

2010-10-13 Thread Steven Wong
You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive. From: hdev ml [mailto:hde...@gmail.com] Sent: Wednesday, October 13, 2010 2:18 PM To: hive-u...@hadoop.apache.org Subject: Exception in hive startup Hi all, I installed Hadoop 0.20.2 and installed hive 0.5.0. I followed

Re: Exception in hive startup

2010-10-13 Thread Ted Yu
This should be documented in README.txt On Wed, Oct 13, 2010 at 6:14 PM, Steven Wong sw...@netflix.com wrote: You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive. *From:* hdev ml [mailto:hde...@gmail.com] *Sent:* Wednesday, October 13, 2010 2:18 PM *To:*

Re: Exception in hive startup

2010-10-13 Thread Edward Capriolo
On Thu, Oct 14, 2010 at 12:49 AM, Ted Yu yuzhih...@gmail.com wrote: This should be documented in README.txt On Wed, Oct 13, 2010 at 6:14 PM, Steven Wong sw...@netflix.com wrote: You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive. From: hdev ml

Got question after deploy hadoop-0.21.0

2010-10-14 Thread SingoWong
Hi, I got some question after deploy hadoop-0.21.0 need to help. There is a new deploy not update, and i execute start-hdfs.sh, start-mapred.sh, got the message in the log file below: 2010-10-14 16:59:55,354 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 9000: readAndProcess threw

when is hive0.6 released

2010-10-14 Thread lei liu
This month our system will be released, so I want to use newest hive0.6 version, I want to know hive0.6 whether can be released in next week. Thanks, LiuLei

Re: when is hive0.6 released

2010-10-14 Thread Carl Steinbach
Hi, We're working on building the first release candidate for 0.6. I expect that the official release will happen soon, but I can't promise that it will occur it in the next week. Thanks. Carl On Thu, Oct 14, 2010 at 7:30 PM, lei liu liulei...@gmail.com wrote: This month our system will be

Re: when is hive0.6 released

2010-10-14 Thread lei liu
Hi Carl Steinbach , Thank you for your reply quickly. I look the metastore code is updated in 06/Oct/10 (please see HIVE-1364https://issues.apache.org/jira/browse/HIVE-1364 issues), could you tell me whether the metasote code in svn will not be updated again when hive0.6 is released? Thanks,

Help with last 30 day unique user query

2010-10-15 Thread Vijay
Hi, I need help with this scenario. We have a table of events which has columns date, event (not important for this discussion), and user_id. It is obviously easy to find number of unique users for each day. I also need to find number of unique users in the last 30 days for each day. This is also

Re: Help with last 30 day unique user query

2010-10-15 Thread Alex Boisvert
As far as I know, Hive has no built-in support for sliding-window analytics. There is an enhancement request here: https://issues.apache.org/jira/browse/HIVE-896 https://issues.apache.org/jira/browse/HIVE-896Without such support, the brute force way of doing things is, SELECT COUNT(DISTINCT

Re: Multiple insert statement and levels of aggregation

2010-10-15 Thread Ning Zhang
In the multi-insert statement, you cannot put another FROM clause. What you can do is to put both UDTF in the FROM clause: FROM foo lateral view someUDTF(foo.a) as t1_a lateral view anotherUDTF(foo.a) as T2_a INSERT ... SELECT a,b,c,count(1), t1_a .. SELECT a,b,c,count(1), t2_a .. On Oct 15,

UDAF modes

2010-10-15 Thread Alex Boisvert
Hi, I'm writing a UDAF and I'm a little unclear about the PARTIAL1, PARTIAL2, FINAL and COMPLETE modes. I've read the extent of the Javadoc ;) and looked at some of the built-in UDAFs in the Hive source tree and I'm still unclear about the properties of the input data in each aggregation step.

Re: Merging small files with dynamic partitions

2010-10-15 Thread Ning Zhang
The output file shows it only have 2 jobs (the mapreduce job and the move task). This indicates that the plan does not have merge enabled. Merge should consists of a ConditionalTask and 2 sub tasks (a MR task and a move task). Can you send the plan of the query? One thing I noticed is that

RE: I define one UDF function, the UDf retunr List ,but When I use ResultSet to receive result hive throw exception

2010-10-19 Thread Howell, Nathan
Try returning a non-empty string from getDisplayString, I've run across this issue before. I use a helper class that looks similar to this Scala function: override def getDisplayString (children: Array[String]): String = children.mkString(FunctionName(, , , )) From: Steven Wong

Re: Key values not importing for SequenceFiles

2010-10-20 Thread Chris Bates
Oh and we are using CDH2 which I believe is using version 0.4.1 On Wed, Oct 20, 2010 at 5:38 PM, Chris Bates christopher.andrew.ba...@gmail.com wrote: Hi all, I've done a lot of research about this today, but haven't found the solution. I'm importing a bunch of sequencefiles that were

Cannot access hive docs r0.5.0

2010-10-21 Thread Manhee Jo
Hi all, I cannot access http://hive.apache.org/docs/r0.5.0/ due to 404 Not Found error. Can anyone tell me what's happening about it? Thanks, Manhee

RE: [VOTE] Bylaws for Apache Hive Project

2010-10-22 Thread Namit Jain
+1 From: Ashish Thusoo [mailto:athu...@facebook.com] Sent: Friday, October 22, 2010 2:51 PM To: user@hive.apache.org Subject: [VOTE] Bylaws for Apache Hive Project Hi Folks, I propose that we adopt the following bylaws for the Apache Hive Project https://cwiki.apache.org/HIVE/bylaws.html

Re: [VOTE] Bylaws for Apache Hive Project

2010-10-22 Thread Ashish Thusoo
I knew I was going to miss a pig somewhere... :) Ashish Sent from my iPhone On Oct 22, 2010, at 2:55 PM, John Sichi jsi...@facebook.com wrote: Hive users etc are encouraged to vote too :) JVS (gotta love cut-and-paste) On Oct 22, 2010, at 2:51 PM, Ashish Thusoo wrote: Hi Folks, I

RE: [VOTE] Bylaws for Apache Hive Project

2010-10-22 Thread Joydeep Sen Sarma
Under 'Code change', 'Binding votes says Lazy approval (not counting the vote of the contributor) But Lazy Approval says it's an action that that is implicitly allowed except for -1. I guess the Code change rule wants to say one +1 is required - but that's not very clear. Maybe Lazy Approval

Re: [VOTE] Bylaws for Apache Hive Project

2010-10-22 Thread Edward Capriolo
Good point about knowing which list to vote on. As for the voting new committers, it would be very hard for me to imagine a nominee being rejected. I would not call it a formality, but essentially the only criteria is having x patches. (Its not dancing with the stars :) Ed On Friday, October 22,

Hive and Hadoop 0.21.0

2010-10-22 Thread Sonal Goyal
Hi, I need to get Hive working on a 0.21.0 Hadoop cluster. Can someone please let me know how it can be done. I tried HIVE-1612 but it did not work for me. Am I missing something? Thanks and Regards, Sonal Sonal Goyal | Founder and CEO | Nube Technologies LLP http://www.nubetech.co |

Re: Hive connection issues on CDH3b3

2010-10-24 Thread 김영우
Hi ray, https://issues.apache.org/jira/browse/HIVE-1264 You should upgrade Hive to 0.7(trunk). CDH3beta3 includes hadoop security features. - Youngwoo 2010/10/25 Ray Duong ray.du...@gmail.com Hi, I'm getting the following error message after upgrading to CDH3b3. Does anyone know how to

Re: Hive connection issues on CDH3b3

2010-10-24 Thread Philip Zeyliger
You can also run the Hive that's part of CDH3b3 (downloadable at http://archive.cloudera.com/cdh/3/hive-0.5.0+32.tar.gz), which works with the security stuff. -- Philip On Sun, Oct 24, 2010 at 7:39 PM, 김영우 warwit...@gmail.com wrote: Hi ray, https://issues.apache.org/jira/browse/HIVE-1264

Re: Hive connection issues on CDH3b3

2010-10-25 Thread Ray Duong
Thank Youngwoo, I checked out the last Hive build from trunk which contains the Hive1264 patch. I was able to get pass last error message and able to query the Hbase table. However, when I try to copy the data from Hbase into a Hive table, I get the following error message. Does anyone know

Re: Granting Access

2010-10-29 Thread Pavan
IMHO, there is no straight forward way of doing this in Hadoop except that you need to install Hadoop components such as MapReduce and HDFS as different users . This is an ongoing development priority. The available access related configuration options (before Kerberos V5) are : -

Re: meeting minutes for 25-Oct-2010 contributor meeting

2010-10-29 Thread Edward Capriolo
On Fri, Oct 29, 2010 at 3:42 PM, John Sichi jsi...@facebook.com wrote: http://wiki.apache.org/hadoop/Hive/Development/ContributorsMeetings/HiveContributorsMinutes101025 JVS Carl Steinbach proposed making 0.7.0 a time-based release (rather than a feature-based release), and that we should

Single map task per file in an external table

2010-10-29 Thread phil young
I'm about to investigate the following situation, but I'd appreciate any insight that can be given. We have an external table which is comprised of 3 HDFS files. We then run an INSERT OVERWRITE which is just a SELECT * from the external table. The table being overwritten has N buckets. The issue

Re: UDF help

2010-11-02 Thread Tim Robertson
Thanks Edward. I'll poke around there. On Tue, Nov 2, 2010 at 6:40 PM, Edward Capriolo edlinuxg...@gmail.com wrote: On Tue, Nov 2, 2010 at 12:47 PM, Tim Robertson timrobertson...@gmail.com wrote: Hi all, Is the following a valid UDF please? When I run it I get the following so I presume

RE: Does output directory remain in case of map/reduce task failures

2010-11-02 Thread Namit Jain
Hive writes to a temporary directory first, and if the UDF fails, the temp. directory is removed. The expected final directory is not touched. -namit From: Pradeep Kamath [prade...@yahoo-inc.com] Sent: Tuesday, November 02, 2010 3:26 PM To:

When I create one table, hive throw one Exception

2010-11-03 Thread lei liu
When I create one table, hive throw below Exception: 2010-11-03 09:42:02,939 ERROR exec.DDLTask (SessionState.java:printError(277)) - FAILED: Error in metadata: java.lang.RuntimeExcepti on: commitTransaction was called but openTransactionCalls = 0. This probably indicates that there are

longest matching prefix

2010-11-04 Thread Eyal
Hi, In MySQL to match '12684041234' to the longest prefix I would do SELECT num_prefix FROM nums WHERE '12684041234' LIKE CONCAT(num_prefix, '%') AND LENGTH(num_prefix) = ( SELECT MAX(LENGTH(num_prefix)) FROM nums WHERE '12684041234' LIKE CONCAT(num_prefix, '%') ) ** Table nums has a column

Upgrading Hypertable Hive connector

2010-11-04 Thread Sanjit Jhala
Hi, I'm taking another pass over the Hypertable Hive storage handler and adding a bit more functionality to it (including regexp filtering). I was also planning to upgrade the Hive version it uses from something between 0.4 0.5 to 0.6. In this regard, my preference is to upgrade to 0.6 but I'm

Re: How to add jar to Hive globally?

2010-11-11 Thread Marcello de Sales
I like the approach to have the properties in the hive-site.xml, as I have it under version control... Thanks for the suggestions... Marcello de Sales On Thu, Nov 11, 2010 at 8:45 PM, Edward Capriolo edlinuxg...@gmail.comwrote: On Thu, Nov 11, 2010 at 11:42 PM, Shrijeet Paliwal

Re: Merging small files with dynamic partitions

2010-11-12 Thread Dave Brondsema
It seems that I can't use this with Hadoop 0.18 since the Hadoop18Shims.getCombineFileInputFormat returns null, and SemanticAnalyzer.java sets HIVEMERGEMAPREDFILES to false if CombineFileInputFormat is not supported. Is that right? Maybe I can copy the Hadoop19Shims implementation of

Re: Merging small files with dynamic partitions

2010-11-12 Thread Dave Brondsema
I copied Hadoop19Shims' implementation of getCombineFileInputFormat (HIVE-1121) into Hadoop18Shims and it worked, if anyone is interested. And hopefully we can upgrade our Hadoop version soon :) On Fri, Nov 12, 2010 at 12:44 PM, Dave Brondsema dbronds...@geek.netwrote: It seems that I can't

Re: How to generate global unique ID?

2010-11-16 Thread Tim Robertson
Does it need to be a sequential INT? If not, then a UUID works very well. Cheers, Tim On Tue, Nov 16, 2010 at 8:55 AM, afancy grou...@gmail.com wrote: Hi, Zhang, How to integrate this snowflake with Hive?  Thanks! Regards, afancy On Mon, Nov 15, 2010 at 10:35 AM, Jeff Zhang

Re: How to generate global unique ID?

2010-11-16 Thread afancy
Hi, Tim, I think sequential integer would be better. Is it difficult to generate a global integer? BTW: how to generate UUID in Hive? thanks Regards, afancy On Tue, Nov 16, 2010 at 9:11 AM, Tim Robertson timrobertson...@gmail.comwrote: Does it need to be a sequential INT? If not, then a UUID

RE: EXTERNAL:Re: unable to create table

2010-11-16 Thread Gerlach, Hannah L (IS)
Dear Ted, Maybe I am missing something, but 'Exception in hive startup' appears to be a different problem. When I run install-dir/bin/hive, it starts fine. The problem arises when I try to create a table once hive is running. Best, Hannah From: Ted Yu

Re: Hive/HBase integration issue.

2010-11-18 Thread afancy
Hi, Does the INSERT clause have to include the OVERWRITE, which means that the new data will overwrite the previous data? How to implement the indeed INSERT operation, instead of OVERWRITE? BTW: How to implement the DELETE operator? thanks afancy

Re: Hive/HBase integration issue.

2010-11-18 Thread John Sichi
As noted here, when writing to HBase, existing rows are overwritten, but old rows are not deleted. http://wiki.apache.org/hadoop/Hive/HBaseIntegration#Overwrite There is not yet any deletion support. JVS On Nov 18, 2010, at 1:00 AM, afancy wrote: Hi, Does the INSERT clause have to

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-18 Thread Leo Alekseyev
Hi Ning, For the dataset I'm experimenting with, the total size of the output is 2mb, and the files are at most a few kb in size. My hive.input.format was set to default HiveInputFormat; however, when I set it to CombineHiveInputFormat, it only made the first stage of the job use fewer mappers.

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-18 Thread Leo Alekseyev
I thought I was running Hive with those changes merged in, but to make sure, I built the latest trunk version. The behavior changed somewhat (as in, it runs 2 stages instead of 1), but it still generates the same number of files (# of files generated is equal to the number of the original

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-19 Thread Dave Brondsema
What version of Hadoop are you on? On Thu, Nov 18, 2010 at 10:48 PM, Leo Alekseyev dnqu...@gmail.com wrote: I thought I was running Hive with those changes merged in, but to make sure, I built the latest trunk version. The behavior changed somewhat (as in, it runs 2 stages instead of 1), but

problem with jdbc connection

2010-11-19 Thread Peter Limbach
Hi, i have a problem selecting tables over a jdbc connection. when i try to select a partitioned Table i get the following error: Error in semantic analysis: javax.jdo.JDODataStoreException: Error adding class org.apache.hadoop.hive.metastore.model.MPartition to list of persistence-managed

How to develop custom Avro InputFormat and Deserializer?

2010-11-19 Thread Denitsa Tsolova
Hi all, We have a large amount of data in HDFS stored in Avro format. We don't want to convert it to Hive supported format. That is why we have developed a custom InputFormat and Deserializer. Our custom InputFormat does not support file split, because the Avro schema (which describes the

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-19 Thread Leo Alekseyev
I'm using Hadoop 0.20.2. Merge jobs (with static partitions) have worked for me in the past. Again, what's strange here is with the latest Hive build the merge stage appears to run, but it doesn't actually merge -- it's a quick map-only job that, near as I can tell, doesn't do anything. On Fri,

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-19 Thread yongqiang he
These are the parameters that control the behavior. (Try to set them to different values if it does not work in your environment.) set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; set mapred.min.split.size.per.node=10; set

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-19 Thread yongqiang he
I can not think this could be the cause. The problem should be: your files can not be merged. I mean the file size is bigger than the split size On Friday, November 19, 2010, Leo Alekseyev dnqu...@gmail.com wrote: Folks, thanks for your help.  I've narrowed the problem down to compression.  

Re: Using jdbc in embedded mode - Can't find warehouse directory [SOLVED]

2010-11-19 Thread Stuart Smith
Hello, Just wanted to let people know I tracked this one down: It looks like it was not picking up the *hadoop* core-site.xml configuration file. - So the variable fs.default.name was never set - So the warehouse dir became file://[hive.metastore.warehouse.dir] instead of [hdfs

Number of simultaneous hive users

2010-11-20 Thread David Lary
How many simultaneous users of hive can there be? Are different invocations of hive by the same user likely to conflict? I ask as I have millions of files to load into hive and so it makes sense to set off several simultaneous hive jobs to load this data in, e.g. with commands like LOAD DATA

Re: Hive produces very small files despite hive.merge...=true settings

2010-11-23 Thread Ning Zhang
This should be expected. Compressed text files are not splittable so that CombineHiveInputFormat cannot read multiple files per mapper. CombinedHiveInputFormat is used when hive.merge.maponly=true. If you set it to false, we'll use HiveInputFormat and that should be able to merge compressed

Querying HBase

2010-11-23 Thread Xavier Stevens
I'm trying to create an external table to a pre-existing HBase table using Hive trunk, like so: CREATE EXTERNAL TABLE hbase_metrics (key string, value mapstring,int) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES (hbase.columns.mapping = :key,counters:)

RE: Hive/HBase integration issue.

2010-11-24 Thread Vivek Mishra
Added 1 post at: http://mevivs.wordpress.com/2010/11/24/hivehbase-integration/ Sharing it if is useful. Vivek -Original Message- From: Vivek Mishra Sent: Friday, November 19, 2010 10:36 AM To: user@hive.apache.org Subject: RE: Hive/HBase integration issue. Hi, Just found that, It is

Does hive have batch processing mode ?

2010-11-24 Thread Jeff Zhang
Hi all, I have a bunch of files, and want to import them into one table each partition per file. Currently, I have to enter each add partition statement in cli. So I wonder whether hive have a batch processing mode, so that I can put the sql statements in one file, and execute the file using one

Re: Does hive have batch processing mode ?

2010-11-25 Thread Jeff Zhang
Thanks, it works On Thu, Nov 25, 2010 at 3:49 PM, james warren ja...@rockyou.com wrote: Try the following: % hive -f myhive.file cheers, -James On Wed, Nov 24, 2010 at 11:37 PM, Jeff Zhang zjf...@gmail.com wrote: Hi all, I have a bunch of files, and want to import them into one table

HIVE and HADOOP Installations on different server

2010-11-25 Thread vaibhav negi
Hi, I have set up a 3 node hadoop cluster excluding 1 name node . Currently, i have installed hive and hive metastore on name node. But now due to some space issue, i want to move hive out of name node and install it on different server. How to go for it? Thanks and Regards Vaibhav Negi

Re: Error while fetching Hive Metadata

2010-11-25 Thread Adarsh Sharma
Carl Steinbach wrote: Hi Adarsh, Hibernate will not work with Hive because Hibernate depends on the ability to execute row-level insert, update and delete operations. None of these operations are supported by Hive. Carl On Tue, Sep 21, 2010 at 3:18 AM, Bennie Schut bsc...@ebuddy.com

Re: hive not stable,weird exception

2010-11-29 Thread Dave Brondsema
Have you looked in the hiveserver logs? I've found that Hive doesn't return very meaningful error messages to the client, and I have to look in hive's logs to find out what happened. On Sun, Nov 28, 2010 at 11:11 PM, shangan shan...@corp.kaixin001.comwrote: I use hive client through php, it

Hive thrift C++ client

2010-11-29 Thread Saptarshi Guha
Hello, Does anyone have some example code of using building this? My steps, I'm trying to build a shared library using the thrift api 1. Checked out from SVN (trunk) 2. Made a directory called thrift, into which I place fb303.thrift, hive_metastore.thrift,hive_service.thrift,queryplan.thrift

Defining the CORR(X, Y) UDAF in version 0.5

2010-11-29 Thread Eric Halpern
I am interested in using the correlation UDAF available in version 0.7 (see https://issues.apache.org/jira/browse/HIVE-1549) but am using AWS Elastic Map-Reduce which is currently running version 0.5. Is there a way to install this function in a older version of Hive? If not, can anyone suggest

Re: Deserializing results from ThriftHiveClient::fetch*

2010-11-30 Thread Saptarshi Guha
And sorry for the terrible grammar. Should have proofread. Cheers J From: Saptarshi Guha saptar...@revolution-computing.com Date: November 30, 2010 4:51:35 PM PST To: user@hive.apache.org Subject: Deserializing results from ThriftHiveClient::fetch* Hello A few questions. Once i get the

Twitter Search + big Hadoop, Dec. 8th at Seattle Scalability Meetup

2010-11-30 Thread Bradford Stephens
Greetings, The Seattle Scalability Meetup isn't slacking for the holidays. We've got an awesome lineup for Wed, December 8 at 7pm: http://www.meetup.com/Seattle-Hadoop-HBase-NoSQL-Meetup/ -Jake Mannix from Twitter will talk about the Twitter Search infrastructure (with distributed Lucene)

Re: Deserializing results from ThriftHiveClient::fetch*

2010-12-01 Thread Saptarshi Guha
Hi Carl, I had played around with the HBase thrift api and it was a matter of generating the cpp files and using them. I assumed Hive thrift would be something similar - hence my messing around with that. But you correctly guessed my intentions : a c++ client. So i'll concentrate on

Backtrace when running build/odbc/test/HiveClientTestC (from trunk)

2010-12-01 Thread Saptarshi Guha
Hello, Does anyone get a backtrace when running HiveClientTestC? On an EC2 instance with Hive (0.6.0), i checked out the trunk (RELEASE_NOTEST.txt says 0.6.0) of Hive and compiled the client. ant compile-cpp -Dthrift.home=/usr/local Copied the odbc/testdata/ to /tmp and ran

Re: Backtrace when running build/odbc/test/HiveClientTestC (from trunk)

2010-12-01 Thread Saptarshi Guha
Hello, This very small code also fails (with similar errors) #include stdio.h #include stdlib.h #include assert.h #include hiveclient.h //x.cpp int main(char **argv, int argc){ fprintf(stderr, Running %s...\n, __FUNCTION__); char err_buf[MAX_HIVE_ERR_MSG_LEN]; HiveReturn retval;

seattle hadoop announce: meeting December 2nd @ 7:15 pm, Mahout and S4

2010-12-01 Thread sean jensen-grey
Hello Fellow Mappers and Reducers, We are meeting at 7:15 pm on December 2nd at the University Heights Community Center 5031 University Way NE Seattle WA 98105 Room #110 The meetings are informal and highly conversational. If you have questions about Hadoop and map reduce this is a great place

Re: Backtrace when running build/odbc/test/HiveClientTestC (from trunk)

2010-12-02 Thread Carl Steinbach
Hi Satparshi, It looks like glibc's MALLOC_CHECK feature is broken on x86_64 systems, at least for some versions. Some background: * http://www.novell.com/support/viewContent.do?externalId=3113982sliceId=1 * http://www.pubbs.net/200911/kde/23500-deactivate-malloccheck-on-x8664.html Setting

Re: Backtrace when running build/odbc/test/HiveClientTestC (from trunk)

2010-12-02 Thread Saptarshi Guha
Hi Carl, Excellent - setting to 0/1 worked. The links were informative. Thanks again for the help. Regards Saptarshi From: Carl Steinbach c...@cloudera.com Date: December 2, 2010 8:12:47 AM PST To: user@hive.apache.org Subject: Re: Backtrace when running build/odbc/test/HiveClientTestC (from

Running a HiveClient with create external table HBase

2010-12-03 Thread Vivek Mishra
Hi, Currently I am facing random behavior while trying to create a java client for Hive hbase integration. Case: I am trying to create a hive table for existing HBase table. So i have started hiveserver via /hive -service hiveserver. In logs I can see it is printing by my sql with CREATE

Does anyone get past TestEmbeededHiveMetaStore in ant test?

2010-12-03 Thread Edward Capriolo
test: [junit] Running org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore [junit] Tests run: 11, Failures: 0, Errors: 0, Time elapsed: 34.192 sec [junit] BR.recoverFromMismatchedToken [junit] Running org.apache.hadoop.hive.metastore.TestRemoteHiveMetaStore [junit]

Re: Does anyone get past TestEmbeededHiveMetaStore in ant test?

2010-12-03 Thread Ashutosh Chauhan
This happens because you have a slower machine :) In this test HiveMetaStore is brought up in separate thread and then current thread sleeps for 5 secs, hoping MetaStore will be up by then. On heavily loaded machine (or slow machines) 5 seconds may be too less. Change TestRemoteHiveMetaStore.java

Re: Caused by: javax.jdo.JDODataStoreException: Exception thrown while querying indices for table=DBS: HELP need

2010-12-03 Thread Carl Steinbach
Hi Tali, Did you run the metastore schema upgrade script? This is a requirement if you are upgrading from an older version of Hive to version 0.6. Thanks. Carl On Fri, Dec 3, 2010 at 1:31 PM, Tali K ncherr...@hotmail.com wrote: Hi All, We installed a new hive distribution - 0.6. We

Re: Query output formatting

2010-12-06 Thread yongqiang he
You can use collect_set() udaf. (And use lateral view join and explode if you want operate on the set data.) On Mon, Dec 6, 2010 at 1:01 PM, Steven Wong sw...@netflix.com wrote: I have this query to calculate some averages: select hour, cdn, avg(bitrate) from fact_table group by hour, cdn

Re: Failure when using insert overwrite after upgrading to Hive 0.6.0

2010-12-07 Thread Ryan LeCompte
I just put the Hive log4j config file on DEBUG, and here is the error that I'm seeing: 2010-12-07 12:16:50,281 WARN mapred.JobClient (JobClient.java:configureCommandLineOptions(539)) - Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 2010-12-07

Re: Failure when using insert overwrite after upgrading to Hive 0.6.0

2010-12-07 Thread Ryan LeCompte
Digging even further, here's what I see: NOTE: We have a table in Hive called test_table but this seems to look for default.test_table ? ) 2010-12-07 00:52:24,600 ERROR metadata.Hive (Hive.java:getTable(357)) - NoSuchObjectException(message:default.test_table table not found) at

Re: Ctrl C and Hive ?

2010-12-07 Thread Edward Capriolo
On Tue, Dec 7, 2010 at 3:18 PM, Tali K ncherr...@hotmail.com wrote: 1) When I cancel hive job with Ctrl C, I noticed that java/hive processes still run on some of my nodes. I shutdown hadfoop, and restarted it, but noticed that  2 or 3 java/hadoop processes were still running on each node. So

RE: Ctrl C and Hive ?

2010-12-07 Thread Tali K
Thanks a lot for your quick reply!! Can you also explain also why command hive -e 'select produces output, prints OK, but give me a prompt only after 7-10 min? if I run hive shell, and do queries inside of shell I don't have such a behaviour. I see Ok , and sec after that I'll have

Are there any examples of simple UDAFs?

2010-12-07 Thread Leo Alekseyev
I am trying to write a very simple aggregation function which seems like an overkill for using GenericUDAF as described on the wiki. However, I can't get the code to run. It always throws an exception of the form java.lang.ClassNotFoundException:

Re: Are there any examples of simple UDAFs?

2010-12-07 Thread Tim Robertson
What about the count or max? http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java I've not used UDAFs, but I only

Re: Are there any examples of simple UDAFs?

2010-12-08 Thread Leo Alekseyev
Good point Tim -- one non-generic UDAF example is in org.apache.hadoop.hive.ql.udf.UDAFPercentile. My problem turned out to be classpath related. Remember kids, if classes aren't being found for the UDFs and UDAFs, list jars and delete jar can be your friends! --Leo On Tue, Dec 7, 2010 at 9:39

Re: Failure when using insert overwrite after upgrading to Hive 0.6.0

2010-12-08 Thread Ryan LeCompte
Yes, I also find it very strange. Unfortunately, it's not an ideal workaround since having hive.merge.mapfiles=false means that we have a lot of blocks being utilized that only have a small amount of data in them (much less than the configured HDFS block size). Can you think of any other

Load data into HIVE from HDFS Clarification

2010-12-09 Thread Tali K
When one loads data to hive , if the file is in hdfs, it is moved into the Hive-controlled file system namespace. and deleted from hdfs. Is it possible to change it so , file wouldn't be moved, but copied?

Re: Load data into HIVE from HDFS Clarification

2010-12-09 Thread Edward Capriolo
On Thu, Dec 9, 2010 at 2:53 PM, Tali K ncherr...@hotmail.com wrote: When one  loads data to hive , if the file is in hdfs, it is moved into the Hive-controlled file system namespace. and deleted from hdfs. Is it possible to change it so , file wouldn't be moved, but copied? You can create

unable to load jtds driver for sqlserver

2010-12-09 Thread Viral Bajaria
Hello, I just came across the GenericUDFDBOutput and was able to successfully write data to a mysql db. I tried modifying the connection string to a sql server using jtds library and for some reason it does not seem to work with jtds. Has anyone come across the same issue before ? I did the

Re: Hive HBase intergration scan failing

2010-12-09 Thread vlisovsky
Hi Guys, Wonder if anybody could shed some light on how to reduce the load on HBase cluster when running a full scan. The need is to dump everything I have in HBase and into a Hive table. The HBase data size is around 500g. The job creates 9000 mappers, after about 1000 maps things go south

Re: Hive HBase intergration scan failing

2010-12-09 Thread John Sichi
Try set hbase.client.scanner.caching=5000; Also, check to make sure that you are getting the expected locality so that mappers are running on the same nodes as the region servers they are scanning (assuming that you are running HBase and mapreduce on the same cluster). When I was testing

Hive import issue

2010-12-10 Thread Vivek Mishra
Hi, I am a newbie to hive. When I am trying to import data to HBase via a table managed by Hive. I am getting following errors: mismatched input 'Timestamp' expecting Identifier in column specification mismatched input 'data' expecting Identifier in column specification Remvoing or

Re: Hive HBase intergration scan failing

2010-12-10 Thread vlisovsky
Thanks for the info. Moreover how can we make sure that our regionservers are running with same Datanodes ( locality). Is there a way we can make sure? On Thu, Dec 9, 2010 at 11:09 PM, John Sichi jsi...@fb.com wrote: Try set hbase.client.scanner.caching=5000; Also, check to make sure that

Metastore compatibility

2010-12-10 Thread Steven Wong
Is it safe to share a 0.7 metastore between 0.7 clients/servers and 0.5 clients/servers? Thanks.

Re: Documentation related to DB operations

2010-12-10 Thread Edward Capriolo
On Fri, Dec 10, 2010 at 9:24 PM, Ashutosh Chauhan hashut...@apache.org wrote: It will really help to have the behavior documented for Database in Hive. I thought of doing it myself but then got stumped by location clause. Reading from ticket https://issues.apache.org/jira/browse/HIVE-675 it

RoadMap vs Jira

2010-12-11 Thread Edward Capriolo
In jira we have ~ 600 scheduled issues. These issues range from unconfirmed bugs, general wish list items, very complex additions such as new syntax or new expanding the scope of hive. Almost everything is marked as a MAJOR - BUG, when many things are minor wishes. I believe we should encourage

Re: Hive HBase intergration scan failing

2010-12-11 Thread John Sichi
It's supposed to happen automatically. The JIRA issue below mentions one case where it wasn't, and explains how I detected it and worked around it. To make you're getting locality, look at the task tracer and make sure that for your map tasks, the host used for executing the task matches the

How to use STRUCT data type ?

2010-12-12 Thread Saravanan Rajendran
Hi, I am new to Hive, I am trying to use struct data type. I am able to create table successfully. I would like to know how to give the data format for the strct data type to avail this facility. My table strcture is create table audittrail (status string, page struct name : string, id string

Re: How to use STRUCT data type ?

2010-12-13 Thread Saravanan Rajendran
Hi, I went through the documentation and solved this problem. Sorry for posing this without reading the documentation properly. On Mon, Dec 13, 2010 at 1:22 PM, Saravanan Rajendran sarava...@julysystems.com wrote: Hi, I am new to Hive, I am trying to use struct data type. I am able to

Re: Estimated release date for Hive 0.7?

2010-12-13 Thread Edward Capriolo
On Mon, Dec 13, 2010 at 11:51 AM, Kirk True k...@mustardgrain.com wrote: Hi all, I hunted around a bit, but didn't see any mention of a projected release date for 0.7. Is there one? Thanks, Kirk These are the open issues for 0.7:

Re: Libfb303.jar

2010-12-14 Thread Pavan
Did you try replacing libfb3003.jar in Hadoop with the libfb3003.jar from Hive? And let us know if this doesn't work. ~ Pavan Yara @yarapavan On Tue, Dec 14, 2010 at 4:22 PM, Adarsh Sharma adarsh.sha...@orkash.comwrote: Dear all, I am using Hadoop-0.20.2 and Hadoopdb Hive on a 5 node

hive newbie - importing data into hive

2010-12-14 Thread Sean Curtis
just wondering if i have a pipe delimited file, how i can just import this data into hive: basically i am using the microlens database, which is pipe separated. for example: user id | age | gender | occupation | zip code translates to 123 | 24 | M | worker | 12345 i'd like to just import

Re: hive newbie - importing data into hive

2010-12-14 Thread Bryan Talbot
I'll help by recommending that you get started by looking at the Getting Started Guide. http://wiki.apache.org/hadoop/Hive/GettingStarted -Bryan On Tue, Dec 14, 2010 at 8:23 PM, Sean Curtis sean.cur...@gmail.com wrote: just wondering if i have a pipe delimited file, how i can just import

  1   2   3   4   5   6   7   8   9   10   >