Hi Wojciech,
Vaibhav is correct. There's a configuration problem in the copy of
hive-default.xml that ships with CDH3u1 which sets
hive.input.format=CombineHiveInputFormat, but leaves mapred.max.split.size
undefined. You can fix this problem by setting mapred.max.split.size in
hive-default.xml to
Hi,
I am adding the log information for a reduce task. I am running hadoop
in standalone mode.
2011-07-28 19:16:42,621 ERROR
org.apache.hadoop.hive.ql.stats.jdbc.JDBCStatsPublisher: Error during JDBC
connection to jdbc:derby:;databaseName=TempStatsStore;create=true.
java.lang.ClassNotFoundExc
Hi,
I have a table, which has close to a billion rows.. I am trying to
create an index for the table, when I do the alter command, I always end up
with map-reduce jobs with errors. The same runs fine for small tables
though, I also notice that the number of reducers are set to 24, even if set
Hey ,
Is any1 using google snappy i tried it but didnt get success.
If there any1 is using it please tell me the procedure to use it.
--
With Regards
Vikas Srivastava
DWH & Analytics Team
Mob:+91 9560885900
One97 | Let's get talking !
If you are using CombineHiveInputFormat it might be the case that all files are
being combined into one large split and hence 1 mapper gets created.
If that is the case you can set the max split size in hive-default.xml config
file to create more splits and hence more map tasks:
mapred.max.s
On Thu, Jul 28, 2011 at 9:23 AM, Wojciech Langiewicz
wrote:
> Hello,
> I'm having isssue running Hive jobs after updating from Hive 0.5 to Hive
> 0.7 (from CDHb4 to CDHu1).
>
> No matter what query I'm running Hive is always using one mapper.
> I have tried different queries with various sizes of
Hello,
I'm having isssue running Hive jobs after updating from Hive 0.5 to Hive
0.7 (from CDHb4 to CDHu1).
No matter what query I'm running Hive is always using one mapper.
I have tried different queries with various sizes of input and ones with
many reducers or no reducers.
For version 0.5