That's right. Hive can use an HBase table as an input format to the
hive query regardless of output format, and can also write the output
to an HBase table regardless of the input format. You can also
supposedly do a join in Hive that uses 1 side of the join from an
HBase table, and the other
Hi all,
I installed Hadoop 0.20.2 and installed hive 0.5.0.
I followed all the instructions on Hive's getting started page for setting
up environment variables like HADOOP_HOME
When I run from command prompt in the hive installation folder as bin/hive
it gives me following exception
Exception
If your query only accesses HBase tables, then yes, Hive does not access any
source data directly from HDFS (although of course it may put intermediate
results in HDFS, e.g. for the result of a join).
However, if your query does something like join a HBase table with a native
Hive table, then
You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive.
From: hdev ml [mailto:hde...@gmail.com]
Sent: Wednesday, October 13, 2010 2:18 PM
To: hive-u...@hadoop.apache.org
Subject: Exception in hive startup
Hi all,
I installed Hadoop 0.20.2 and installed hive 0.5.0.
I followed
This should be documented in README.txt
On Wed, Oct 13, 2010 at 6:14 PM, Steven Wong sw...@netflix.com wrote:
You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive.
*From:* hdev ml [mailto:hde...@gmail.com]
*Sent:* Wednesday, October 13, 2010 2:18 PM
*To:*
On Thu, Oct 14, 2010 at 12:49 AM, Ted Yu yuzhih...@gmail.com wrote:
This should be documented in README.txt
On Wed, Oct 13, 2010 at 6:14 PM, Steven Wong sw...@netflix.com wrote:
You need to run hive_root/build/dist/bin/hive, not hive_root/bin/hive.
From: hdev ml
Hi,
I got some question after deploy hadoop-0.21.0 need to help.
There is a new deploy not update, and i execute start-hdfs.sh,
start-mapred.sh, got the message in the log file below:
2010-10-14 16:59:55,354 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 9000: readAndProcess threw
This month our system will be released, so I want to use newest hive0.6
version, I want to know hive0.6 whether can be released in next week.
Thanks,
LiuLei
Hi,
We're working on building the first release candidate for 0.6. I expect that
the official release will happen soon, but I can't promise that it will
occur it in the next week.
Thanks.
Carl
On Thu, Oct 14, 2010 at 7:30 PM, lei liu liulei...@gmail.com wrote:
This month our system will be
Hi Carl Steinbach ,
Thank you for your reply quickly.
I look the metastore code is updated in 06/Oct/10 (please see
HIVE-1364https://issues.apache.org/jira/browse/HIVE-1364 issues),
could you tell me whether the metasote code in svn will not be updated again
when hive0.6 is released?
Thanks,
Hi, I need help with this scenario. We have a table of events which has
columns date, event (not important for this discussion), and user_id. It is
obviously easy to find number of unique users for each day. I also need to
find number of unique users in the last 30 days for each day. This is also
As far as I know, Hive has no built-in support for sliding-window analytics.
There is an enhancement request here:
https://issues.apache.org/jira/browse/HIVE-896
https://issues.apache.org/jira/browse/HIVE-896Without such support, the
brute force way of doing things is,
SELECT COUNT(DISTINCT
In the multi-insert statement, you cannot put another FROM clause. What you can
do is to put both UDTF in the FROM clause:
FROM foo lateral view someUDTF(foo.a) as t1_a lateral view anotherUDTF(foo.a)
as T2_a
INSERT ...
SELECT a,b,c,count(1), t1_a
..
SELECT a,b,c,count(1), t2_a
..
On Oct 15,
Hi,
I'm writing a UDAF and I'm a little unclear about the PARTIAL1, PARTIAL2,
FINAL and COMPLETE modes.
I've read the extent of the Javadoc ;) and looked at some of the built-in
UDAFs in the Hive source tree and I'm still unclear about the properties of
the input data in each aggregation step.
The output file shows it only have 2 jobs (the mapreduce job and the move
task). This indicates that the plan does not have merge enabled. Merge should
consists of a ConditionalTask and 2 sub tasks (a MR task and a move task). Can
you send the plan of the query?
One thing I noticed is that
Try returning a non-empty string from getDisplayString, I've run across this
issue before.
I use a helper class that looks similar to this Scala function:
override def getDisplayString (children: Array[String]): String =
children.mkString(FunctionName(, , , ))
From: Steven Wong
Oh and we are using CDH2 which I believe is using version 0.4.1
On Wed, Oct 20, 2010 at 5:38 PM, Chris Bates
christopher.andrew.ba...@gmail.com wrote:
Hi all,
I've done a lot of research about this today, but haven't found the
solution. I'm importing a bunch of sequencefiles that were
Hi all,
I cannot access http://hive.apache.org/docs/r0.5.0/ due to 404 Not Found
error.
Can anyone tell me what's happening about it?
Thanks,
Manhee
+1
From: Ashish Thusoo [mailto:athu...@facebook.com]
Sent: Friday, October 22, 2010 2:51 PM
To: user@hive.apache.org
Subject: [VOTE] Bylaws for Apache Hive Project
Hi Folks,
I propose that we adopt the following bylaws for the Apache Hive Project
https://cwiki.apache.org/HIVE/bylaws.html
I knew I was going to miss a pig somewhere... :)
Ashish
Sent from my iPhone
On Oct 22, 2010, at 2:55 PM, John Sichi jsi...@facebook.com wrote:
Hive users etc are encouraged to vote too :)
JVS (gotta love cut-and-paste)
On Oct 22, 2010, at 2:51 PM, Ashish Thusoo wrote:
Hi Folks,
I
Under 'Code change', 'Binding votes says
Lazy approval (not counting the vote of the contributor)
But Lazy Approval says it's an action that that is implicitly allowed except
for -1. I guess the Code change rule wants to say one +1 is required - but
that's not very clear. Maybe Lazy Approval
Good point about knowing which list to vote on. As for the voting new
committers, it would be very hard for me to imagine a nominee being
rejected. I would not call it a formality, but essentially the only
criteria is having x patches. (Its not dancing with the stars :)
Ed
On Friday, October 22,
Hi,
I need to get Hive working on a 0.21.0 Hadoop cluster. Can someone please
let me know how it can be done. I tried HIVE-1612 but it did not work for
me. Am I missing something?
Thanks and Regards,
Sonal
Sonal Goyal | Founder and CEO | Nube Technologies LLP
http://www.nubetech.co |
Hi ray,
https://issues.apache.org/jira/browse/HIVE-1264
You should upgrade Hive to 0.7(trunk). CDH3beta3 includes hadoop security
features.
- Youngwoo
2010/10/25 Ray Duong ray.du...@gmail.com
Hi,
I'm getting the following error message after upgrading to CDH3b3. Does
anyone know how to
You can also run the Hive that's part of CDH3b3 (downloadable at
http://archive.cloudera.com/cdh/3/hive-0.5.0+32.tar.gz), which works
with the security stuff.
-- Philip
On Sun, Oct 24, 2010 at 7:39 PM, 김영우 warwit...@gmail.com wrote:
Hi ray,
https://issues.apache.org/jira/browse/HIVE-1264
Thank Youngwoo,
I checked out the last Hive build from trunk which contains the Hive1264
patch. I was able to get pass last error message and able to query the
Hbase table. However, when I try to copy the data from Hbase into a Hive
table, I get the following error message. Does anyone know
IMHO, there is no straight forward way of doing this in Hadoop except that
you need to install Hadoop components such as MapReduce and HDFS as
different users . This is an ongoing development priority.
The available access related configuration options (before Kerberos V5) are
:
-
On Fri, Oct 29, 2010 at 3:42 PM, John Sichi jsi...@facebook.com wrote:
http://wiki.apache.org/hadoop/Hive/Development/ContributorsMeetings/HiveContributorsMinutes101025
JVS
Carl Steinbach proposed making 0.7.0 a time-based release (rather than
a feature-based release), and that we should
I'm about to investigate the following situation, but I'd appreciate any
insight that can be given.
We have an external table which is comprised of 3 HDFS files.
We then run an INSERT OVERWRITE which is just a SELECT * from the external
table.
The table being overwritten has N buckets.
The issue
Thanks Edward. I'll poke around there.
On Tue, Nov 2, 2010 at 6:40 PM, Edward Capriolo edlinuxg...@gmail.com wrote:
On Tue, Nov 2, 2010 at 12:47 PM, Tim Robertson
timrobertson...@gmail.com wrote:
Hi all,
Is the following a valid UDF please?
When I run it I get the following so I presume
Hive writes to a temporary directory first, and if the UDF fails, the temp.
directory is removed.
The expected final directory is not touched.
-namit
From: Pradeep Kamath [prade...@yahoo-inc.com]
Sent: Tuesday, November 02, 2010 3:26 PM
To:
When I create one table, hive throw below Exception:
2010-11-03 09:42:02,939 ERROR exec.DDLTask
(SessionState.java:printError(277)) - FAILED: Error in metadata:
java.lang.RuntimeExcepti
on: commitTransaction was called but openTransactionCalls = 0. This probably
indicates that there are
Hi,
In MySQL to match '12684041234' to the longest prefix I would do
SELECT num_prefix
FROM nums
WHERE '12684041234' LIKE CONCAT(num_prefix, '%')
AND LENGTH(num_prefix) = (
SELECT MAX(LENGTH(num_prefix))
FROM nums
WHERE '12684041234' LIKE CONCAT(num_prefix, '%')
)
** Table nums has a column
Hi,
I'm taking another pass over the Hypertable Hive storage handler and adding
a bit more functionality to it (including regexp filtering). I was also
planning to upgrade the Hive version it uses from something between 0.4
0.5 to 0.6. In this regard, my preference is to upgrade to 0.6 but I'm
I like the approach to have the properties in the hive-site.xml, as I have
it under version control...
Thanks for the suggestions...
Marcello de Sales
On Thu, Nov 11, 2010 at 8:45 PM, Edward Capriolo edlinuxg...@gmail.comwrote:
On Thu, Nov 11, 2010 at 11:42 PM, Shrijeet Paliwal
It seems that I can't use this with Hadoop 0.18 since the
Hadoop18Shims.getCombineFileInputFormat returns null, and
SemanticAnalyzer.java sets HIVEMERGEMAPREDFILES to false if
CombineFileInputFormat is not supported. Is that right? Maybe I can copy
the Hadoop19Shims implementation of
I copied Hadoop19Shims' implementation of getCombineFileInputFormat
(HIVE-1121) into Hadoop18Shims and it worked, if anyone is interested.
And hopefully we can upgrade our Hadoop version soon :)
On Fri, Nov 12, 2010 at 12:44 PM, Dave Brondsema dbronds...@geek.netwrote:
It seems that I can't
Does it need to be a sequential INT? If not, then a UUID works very well.
Cheers,
Tim
On Tue, Nov 16, 2010 at 8:55 AM, afancy grou...@gmail.com wrote:
Hi, Zhang,
How to integrate this snowflake with Hive? Thanks!
Regards,
afancy
On Mon, Nov 15, 2010 at 10:35 AM, Jeff Zhang
Hi, Tim,
I think sequential integer would be better. Is it difficult to generate a
global integer?
BTW: how to generate UUID in Hive? thanks
Regards,
afancy
On Tue, Nov 16, 2010 at 9:11 AM, Tim Robertson timrobertson...@gmail.comwrote:
Does it need to be a sequential INT? If not, then a UUID
Dear Ted,
Maybe I am missing something, but 'Exception in hive startup' appears
to be a different problem.
When I run install-dir/bin/hive, it starts fine. The problem arises
when I try to create a table once hive is running.
Best,
Hannah
From: Ted Yu
Hi,
Does the INSERT clause have to include the OVERWRITE, which means that the
new data will overwrite the previous data? How to implement the indeed
INSERT operation, instead of OVERWRITE?
BTW: How to implement the DELETE operator? thanks
afancy
As noted here, when writing to HBase, existing rows are overwritten, but old
rows are not deleted.
http://wiki.apache.org/hadoop/Hive/HBaseIntegration#Overwrite
There is not yet any deletion support.
JVS
On Nov 18, 2010, at 1:00 AM, afancy wrote:
Hi,
Does the INSERT clause have to
Hi Ning,
For the dataset I'm experimenting with, the total size of the output
is 2mb, and the files are at most a few kb in size. My
hive.input.format was set to default HiveInputFormat; however, when I
set it to CombineHiveInputFormat, it only made the first stage of the
job use fewer mappers.
I thought I was running Hive with those changes merged in, but to make
sure, I built the latest trunk version. The behavior changed somewhat
(as in, it runs 2 stages instead of 1), but it still generates the
same number of files (# of files generated is equal to the number of
the original
What version of Hadoop are you on?
On Thu, Nov 18, 2010 at 10:48 PM, Leo Alekseyev dnqu...@gmail.com wrote:
I thought I was running Hive with those changes merged in, but to make
sure, I built the latest trunk version. The behavior changed somewhat
(as in, it runs 2 stages instead of 1), but
Hi,
i have a problem selecting tables over a jdbc connection.
when i try to select a partitioned Table i get the following error:
Error in semantic analysis: javax.jdo.JDODataStoreException: Error adding
class org.apache.hadoop.hive.metastore.model.MPartition to list of
persistence-managed
Hi all,
We have a large amount of data in HDFS stored in Avro format. We don't want
to convert it to Hive supported format.
That is why we have developed a custom InputFormat and Deserializer. Our
custom InputFormat does not support file split, because the Avro schema
(which describes the
I'm using Hadoop 0.20.2. Merge jobs (with static partitions) have
worked for me in the past. Again, what's strange here is with the
latest Hive build the merge stage appears to run, but it doesn't
actually merge -- it's a quick map-only job that, near as I can tell,
doesn't do anything.
On Fri,
These are the parameters that control the behavior. (Try to set them
to different values if it does not work in your environment.)
set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat;
set mapred.min.split.size.per.node=10;
set
I can not think this could be the cause.
The problem should be: your files can not be merged. I mean the file
size is bigger than the split size
On Friday, November 19, 2010, Leo Alekseyev dnqu...@gmail.com wrote:
Folks, thanks for your help. I've narrowed the problem down to
compression.
Hello,
Just wanted to let people know I tracked this one down:
It looks like it was not picking up the *hadoop* core-site.xml configuration
file.
- So the variable fs.default.name was never set
- So the warehouse dir became file://[hive.metastore.warehouse.dir] instead of
[hdfs
How many simultaneous users of hive can there be?
Are different invocations of hive by the same user likely to conflict?
I ask as I have millions of files to load into hive and so it makes sense to
set off several simultaneous hive jobs to load this data in, e.g. with commands
like
LOAD DATA
This should be expected. Compressed text files are not splittable so that
CombineHiveInputFormat cannot read multiple files per mapper.
CombinedHiveInputFormat is used when hive.merge.maponly=true. If you set it to
false, we'll use HiveInputFormat and that should be able to merge compressed
I'm trying to create an external table to a pre-existing HBase table
using Hive trunk, like so:
CREATE EXTERNAL TABLE hbase_metrics (key string, value mapstring,int)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH
SERDEPROPERTIES (hbase.columns.mapping = :key,counters:)
Added 1 post at:
http://mevivs.wordpress.com/2010/11/24/hivehbase-integration/
Sharing it if is useful.
Vivek
-Original Message-
From: Vivek Mishra
Sent: Friday, November 19, 2010 10:36 AM
To: user@hive.apache.org
Subject: RE: Hive/HBase integration issue.
Hi,
Just found that, It is
Hi all,
I have a bunch of files, and want to import them into one table each
partition per file. Currently, I have to enter each add partition
statement in cli.
So I wonder whether hive have a batch processing mode, so that I can
put the sql statements in one file, and execute the file using one
Thanks, it works
On Thu, Nov 25, 2010 at 3:49 PM, james warren ja...@rockyou.com wrote:
Try the following:
% hive -f myhive.file
cheers,
-James
On Wed, Nov 24, 2010 at 11:37 PM, Jeff Zhang zjf...@gmail.com wrote:
Hi all,
I have a bunch of files, and want to import them into one table
Hi,
I have set up a 3 node hadoop cluster excluding 1 name node . Currently, i
have installed hive and hive metastore on name node.
But now due to some space issue, i want to move hive out of name node and
install it on different server. How to go for it?
Thanks and Regards
Vaibhav Negi
Carl Steinbach wrote:
Hi Adarsh,
Hibernate will not work with Hive because Hibernate depends on the
ability to execute row-level insert, update and delete operations.
None of these operations are supported by Hive.
Carl
On Tue, Sep 21, 2010 at 3:18 AM, Bennie Schut bsc...@ebuddy.com
Have you looked in the hiveserver logs? I've found that Hive doesn't return
very meaningful error messages to the client, and I have to look in hive's
logs to find out what happened.
On Sun, Nov 28, 2010 at 11:11 PM, shangan shan...@corp.kaixin001.comwrote:
I use hive client through php, it
Hello,
Does anyone have some example code of using building this?
My steps, I'm trying to build a shared library using the thrift api
1. Checked out from SVN (trunk)
2. Made a directory called thrift, into which I place fb303.thrift,
hive_metastore.thrift,hive_service.thrift,queryplan.thrift
I am interested in using the correlation UDAF available in version 0.7 (see
https://issues.apache.org/jira/browse/HIVE-1549) but am using AWS Elastic
Map-Reduce which is currently running version 0.5.
Is there a way to install this function in a older version of Hive?
If not, can anyone suggest
And sorry for the terrible grammar. Should have proofread.
Cheers
J
From: Saptarshi Guha saptar...@revolution-computing.com
Date: November 30, 2010 4:51:35 PM PST
To: user@hive.apache.org
Subject: Deserializing results from ThriftHiveClient::fetch*
Hello
A few questions. Once i get the
Greetings,
The Seattle Scalability Meetup isn't slacking for the holidays. We've
got an awesome lineup for Wed, December 8 at 7pm:
http://www.meetup.com/Seattle-Hadoop-HBase-NoSQL-Meetup/
-Jake Mannix from Twitter will talk about the Twitter Search
infrastructure (with distributed Lucene)
Hi Carl,
I had played around with the HBase thrift api and it was a matter of generating
the cpp files and using them.
I assumed Hive thrift would be something similar - hence my messing around with
that.
But you correctly guessed my intentions : a c++ client. So i'll concentrate on
Hello,
Does anyone get a backtrace when running HiveClientTestC?
On an EC2 instance with Hive (0.6.0), i checked out the trunk
(RELEASE_NOTEST.txt says 0.6.0)
of Hive and compiled the client.
ant compile-cpp -Dthrift.home=/usr/local
Copied the odbc/testdata/ to /tmp and ran
Hello,
This very small code also fails (with similar errors)
#include stdio.h
#include stdlib.h
#include assert.h
#include hiveclient.h
//x.cpp
int main(char **argv, int argc){
fprintf(stderr, Running %s...\n, __FUNCTION__);
char err_buf[MAX_HIVE_ERR_MSG_LEN];
HiveReturn retval;
Hello Fellow Mappers and Reducers,
We are meeting at 7:15 pm on December 2nd at the
University Heights Community Center
5031 University Way NE
Seattle WA 98105
Room #110
The meetings are informal and highly conversational. If you have
questions about Hadoop and map reduce this is a great
place
Hi Satparshi,
It looks like glibc's MALLOC_CHECK feature is broken on x86_64 systems, at
least for some versions.
Some background:
* http://www.novell.com/support/viewContent.do?externalId=3113982sliceId=1
* http://www.pubbs.net/200911/kde/23500-deactivate-malloccheck-on-x8664.html
Setting
Hi Carl,
Excellent - setting to 0/1 worked. The links were informative.
Thanks again for the help.
Regards
Saptarshi
From: Carl Steinbach c...@cloudera.com
Date: December 2, 2010 8:12:47 AM PST
To: user@hive.apache.org
Subject: Re: Backtrace when running build/odbc/test/HiveClientTestC (from
Hi,
Currently I am facing random behavior while trying to create a java client for
Hive hbase integration.
Case: I am trying to create a hive table for existing HBase table. So i have
started hiveserver via /hive -service hiveserver. In logs I can see it is
printing by my sql with CREATE
test:
[junit] Running org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore
[junit] Tests run: 11, Failures: 0, Errors: 0, Time elapsed: 34.192 sec
[junit] BR.recoverFromMismatchedToken
[junit] Running org.apache.hadoop.hive.metastore.TestRemoteHiveMetaStore
[junit]
This happens because you have a slower machine :) In this test
HiveMetaStore is brought up in separate thread and then current thread
sleeps for 5 secs, hoping MetaStore will be up by then. On heavily
loaded machine (or slow machines) 5 seconds may be too less. Change
TestRemoteHiveMetaStore.java
Hi Tali,
Did you run the metastore schema upgrade script? This is a requirement if
you are upgrading from an older version of Hive to version 0.6.
Thanks.
Carl
On Fri, Dec 3, 2010 at 1:31 PM, Tali K ncherr...@hotmail.com wrote:
Hi All,
We installed a new hive distribution - 0.6.
We
You can use collect_set() udaf. (And use lateral view join and explode
if you want operate on the set data.)
On Mon, Dec 6, 2010 at 1:01 PM, Steven Wong sw...@netflix.com wrote:
I have this query to calculate some averages:
select hour, cdn, avg(bitrate) from fact_table group by hour, cdn
I just put the Hive log4j config file on DEBUG, and here is the error that
I'm seeing:
2010-12-07 12:16:50,281 WARN mapred.JobClient
(JobClient.java:configureCommandLineOptions(539)) - Use GenericOptionsParser
for parsing the arguments. Applications should implement Tool for the same.
2010-12-07
Digging even further, here's what I see:
NOTE: We have a table in Hive called test_table but this seems to look for
default.test_table ? )
2010-12-07 00:52:24,600 ERROR metadata.Hive (Hive.java:getTable(357)) -
NoSuchObjectException(message:default.test_table table not found)
at
On Tue, Dec 7, 2010 at 3:18 PM, Tali K ncherr...@hotmail.com wrote:
1) When I cancel hive job with Ctrl C, I noticed that java/hive processes
still run on some of my nodes.
I shutdown hadfoop, and restarted it, but noticed that 2 or 3 java/hadoop
processes were still running on each node.
So
Thanks a lot for your quick reply!!
Can you also explain also why command hive -e 'select
produces output, prints OK, but give me a prompt only after 7-10 min?
if I run hive shell, and do queries inside of shell I don't have such a
behaviour. I see Ok , and sec after that I'll have
I am trying to write a very simple aggregation function which seems
like an overkill for using GenericUDAF as described on the wiki.
However, I can't get the code to run. It always throws an exception
of the form
java.lang.ClassNotFoundException:
What about the count or max?
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFCount.java
http://svn.apache.org/repos/asf/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDAFMax.java
I've not used UDAFs, but I only
Good point Tim -- one non-generic UDAF example is in
org.apache.hadoop.hive.ql.udf.UDAFPercentile.
My problem turned out to be classpath related. Remember kids, if
classes aren't being found for the UDFs and UDAFs, list jars and
delete jar can be your friends!
--Leo
On Tue, Dec 7, 2010 at 9:39
Yes, I also find it very strange.
Unfortunately, it's not an ideal workaround since having
hive.merge.mapfiles=false means that we have a lot of blocks being utilized
that only have a small amount of data in them (much less than the
configured HDFS block size).
Can you think of any other
When one loads data to hive , if the file is in hdfs, it is moved into the
Hive-controlled file system namespace.
and deleted from hdfs. Is it possible to change it so , file wouldn't be moved,
but copied?
On Thu, Dec 9, 2010 at 2:53 PM, Tali K ncherr...@hotmail.com wrote:
When one loads data to hive , if the file is in hdfs, it is moved into the
Hive-controlled file system namespace.
and deleted from hdfs. Is it possible to change it so , file wouldn't be
moved, but copied?
You can create
Hello,
I just came across the GenericUDFDBOutput and was able to successfully write
data to a mysql db.
I tried modifying the connection string to a sql server using jtds library
and for some reason it does not seem to work with jtds.
Has anyone come across the same issue before ?
I did the
Hi Guys,
Wonder if anybody could shed some light on how to reduce the load on HBase
cluster when running a full scan.
The need is to dump everything I have in HBase and into a Hive table. The
HBase data size is around 500g.
The job creates 9000 mappers, after about 1000 maps things go south
Try
set hbase.client.scanner.caching=5000;
Also, check to make sure that you are getting the expected locality so that
mappers are running on the same nodes as the region servers they are scanning
(assuming that you are running HBase and mapreduce on the same cluster). When
I was testing
Hi,
I am a newbie to hive.
When I am trying to import data to HBase via a table managed by Hive. I am
getting following errors:
mismatched input 'Timestamp' expecting Identifier in column specification
mismatched input 'data' expecting Identifier in column specification
Remvoing or
Thanks for the info. Moreover how can we make sure that our regionservers
are running with same Datanodes ( locality). Is there a way we can make
sure?
On Thu, Dec 9, 2010 at 11:09 PM, John Sichi jsi...@fb.com wrote:
Try
set hbase.client.scanner.caching=5000;
Also, check to make sure that
Is it safe to share a 0.7 metastore between 0.7 clients/servers and 0.5
clients/servers?
Thanks.
On Fri, Dec 10, 2010 at 9:24 PM, Ashutosh Chauhan hashut...@apache.org wrote:
It will really help to have the behavior documented for Database in
Hive. I thought of doing it myself but then got stumped by location
clause. Reading from ticket
https://issues.apache.org/jira/browse/HIVE-675 it
In jira we have ~ 600 scheduled issues. These issues range from
unconfirmed bugs, general wish list items, very complex additions such
as new syntax or new expanding the scope of hive. Almost everything is
marked as a MAJOR - BUG, when many things are minor wishes.
I believe we should encourage
It's supposed to happen automatically. The JIRA issue below mentions one case
where it wasn't, and explains how I detected it and worked around it. To make
you're getting locality, look at the task tracer and make sure that for your
map tasks, the host used for executing the task matches the
Hi,
I am new to Hive, I am trying to use struct data type. I am able to create
table successfully. I would like to know how to give the data format for the
strct data type to avail this facility.
My table strcture is
create table audittrail (status string, page struct name : string,
id string
Hi,
I went through the documentation and solved this problem. Sorry for posing
this without reading the documentation properly.
On Mon, Dec 13, 2010 at 1:22 PM, Saravanan Rajendran
sarava...@julysystems.com wrote:
Hi,
I am new to Hive, I am trying to use struct data type. I am able to
On Mon, Dec 13, 2010 at 11:51 AM, Kirk True k...@mustardgrain.com wrote:
Hi all,
I hunted around a bit, but didn't see any mention of a projected release
date for 0.7. Is there one?
Thanks,
Kirk
These are the open issues for 0.7:
Did you try replacing libfb3003.jar in Hadoop with the libfb3003.jar from
Hive?
And let us know if this doesn't work.
~
Pavan Yara
@yarapavan
On Tue, Dec 14, 2010 at 4:22 PM, Adarsh Sharma adarsh.sha...@orkash.comwrote:
Dear all,
I am using Hadoop-0.20.2 and Hadoopdb Hive on a 5 node
just wondering if i have a pipe delimited file, how i can just import this data
into hive:
basically i am using the microlens database, which is pipe separated. for
example:
user id | age | gender | occupation | zip code
translates to
123 | 24 | M | worker | 12345
i'd like to just import
I'll help by recommending that you get started by looking at the Getting
Started Guide.
http://wiki.apache.org/hadoop/Hive/GettingStarted
-Bryan
On Tue, Dec 14, 2010 at 8:23 PM, Sean Curtis sean.cur...@gmail.com wrote:
just wondering if i have a pipe delimited file, how i can just import
1 - 100 of 19456 matches
Mail list logo