this is the error on hadoop job

2013-01-23 12:15:44,884 INFO org.apache.hadoop.mapred.ReduceTask:
Failed to fetch map-output from attempt_201301231151_0002_m_000001_0
even after MAX_FETCH_RETRIES_PER_MAP retries...  or it is a read
error,  reporting to the JobTracker
2013-01-23 12:15:44,885 FATAL org.apache.hadoop.mapred.ReduceTask:
Shuffle failed with too many fetch failures and insufficient
progress!Killing task attempt_201301231151_0002_r_000000_0.

2013-01-23 12:15:45,220 FATAL org.apache.hadoop.mapred.Task: Failed to
contact the tasktracker
org.apache.hadoop.ipc.RemoteException: java.io.IOException:
JvmValidate Failed. Ignoring request from task:
attempt_201301231151_0002_r_000000_0, with JvmId:
jvm_201301231151_0002_r_1079250852


so something is a mess either your network went down or nodes went down

hive tries to get the same task log from the host (savitha-vitualbox)
and it can't figure out what that host is.





On Wed, Jan 23, 2013 at 1:28 PM, venkatramanan <venkatraman...@smartek21.com
> wrote:

>  No, all the nodes are up and running. i dont know, when hive takes the
> other nodes "HOST NAME" thats the error i guess..
>
> revert me if am wrong
>
>
> On Wednesday 23 January 2013 01:07 PM, Nitin Pawar wrote:
>
> when you ran the query, did the VM shutdown ?
>
>
> On Wed, Jan 23, 2013 at 12:57 PM, venkatramanan <
> venkatraman...@smartek21.com> wrote:
>
>>  Hi,
>>
>> I got the following error while executing the "select count(1) from
>> tweettrend;"
>>
>> Below are the exact log msg from the jobtracker Web Interface
>>
>> *Hive Cli Error:*
>>
>> Exception in thread "Thread-21" java.lang.RuntimeException: Error while
>> reading from task log url
>>     at
>> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getStackTraces(TaskLogProcessor.java:240)
>>     at
>> org.apache.hadoop.hive.ql.exec.JobDebugger.showJobFailDebugInfo(JobDebugger.java:227)
>>     at org.apache.hadoop.hive.ql.exec.JobDebugger.run(JobDebugger.java:92)
>>     at java.lang.Thread.run(Thread.java:722)
>> Caused by: java.net.UnknownHostException: savitha-VirtualBox
>>     at
>> java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:178)
>>     at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:391)
>>     at java.net.Socket.connect(Socket.java:579)
>>     at java.net.Socket.connect(Socket.java:528)
>>     at sun.net.NetworkClient.doConnect(NetworkClient.java:180)
>>     at sun.net.www.http.HttpClient.openServer(HttpClient.java:378)
>>     at sun.net.www.http.HttpClient.openServer(HttpClient.java:473)
>>     at sun.net.www.http.HttpClient.<init>(HttpClient.java:203)
>>     at sun.net.www.http.HttpClient.New(HttpClient.java:290)
>>     at sun.net.www.http.HttpClient.New(HttpClient.java:306)
>>     at
>> sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:995)
>>     at
>> sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:931)
>>     at
>> sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:849)
>>     at
>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1299)
>>     at java.net.URL.openStream(URL.java:1037)
>>     at
>> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getStackTraces(TaskLogProcessor.java:192)
>>     ... 3 more
>> FAILED: Execution Error, return code 2 from
>> org.apache.hadoop.hive.ql.exec.MapRedTask
>> MapReduce Jobs Launched:
>> Job 0: Map: 2  Reduce: 1   Cumulative CPU: 9.0 sec   HDFS Read: 408671053
>> HDFS Write: 0 FAIL
>> Total MapReduce CPU Time Spent: 9 seconds 0 msec
>>
>> *syslog logs*
>>
>> utCopier.copyOutput(ReduceTask.java:1394)
>>      at 
>> org.apache.hadoop.mapred.ReduceTask$ReduceCopier$MapOutputCopier.run(ReduceTask.java:1326)
>>
>> 2013-01-23 12:15:44,884 INFO org.apache.hadoop.mapred.ReduceTask: Task 
>> attempt_201301231151_0002_r_000000_0: Failed fetch #10 from 
>> attempt_201301231151_0002_m_000001_0
>> 2013-01-23 12:15:44,884 INFO org.apache.hadoop.mapred.ReduceTask: Failed to 
>> fetch map-output from attempt_201301231151_0002_m_000001_0 even after 
>> MAX_FETCH_RETRIES_PER_MAP retries...  or it is a read error,  reporting to 
>> the JobTracker
>> 2013-01-23 12:15:44,885 FATAL org.apache.hadoop.mapred.ReduceTask: Shuffle 
>> failed with too many fetch failures and insufficient progress!Killing task 
>> attempt_201301231151_0002_r_000000_0.
>> 2013-01-23 12:15:44,889 WARN org.apache.hadoop.mapred.ReduceTask: 
>> attempt_201301231151_0002_r_000000_0 adding host savitha-VirtualBox to 
>> penalty box, next contact in 137 seconds
>> 2013-01-23 12:15:44,889 INFO org.apache.hadoop.mapred.ReduceTask: 
>> attempt_201301231151_0002_r_000000_0: Got 1 map-outputs from previous 
>> failures
>> 2013-01-23 12:15:45,218 FATAL org.apache.hadoop.mapred.Task: 
>> attempt_201301231151_0002_r_000000_0 GetMapEventsThread Ignoring exception : 
>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: JvmValidate 
>> Failed. Ignoring request from task: attempt_201301231151_0002_r_000000_0, 
>> with JvmId: jvm_201301231151_0002_r_1079250852
>>      at 
>> org.apache.hadoop.mapred.TaskTracker.validateJVM(TaskTracker.java:3278)
>>      at 
>> org.apache.hadoop.mapred.TaskTracker.getMapCompletionEvents(TaskTracker.java:3537)
>>      at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source)
>>      at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:601)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>      at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>>
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1070)
>>      at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>      at $Proxy1.getMapCompletionEvents(Unknown Source)
>>      at 
>> org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.getMapCompletionEvents(ReduceTask.java:2846)
>>      at 
>> org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.run(ReduceTask.java:2810)
>>
>> 2013-01-23 12:15:45,220 FATAL org.apache.hadoop.mapred.Task: Failed to 
>> contact the tasktracker
>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: JvmValidate 
>> Failed. Ignoring request from task: attempt_201301231151_0002_r_000000_0, 
>> with JvmId: jvm_201301231151_0002_r_1079250852
>>      at 
>> org.apache.hadoop.mapred.TaskTracker.validateJVM(TaskTracker.java:3278)
>>      at 
>> org.apache.hadoop.mapred.TaskTracker.fatalError(TaskTracker.java:3520)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:601)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>      at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>>
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1070)
>>      at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>      at $Proxy1.fatalError(Unknown Source)
>>      at org.apache.hadoop.mapred.Task.reportFatalError(Task.java:298)
>>      at 
>> org.apache.hadoop.mapred.ReduceTask$ReduceCopier$GetMapEventsThread.run(ReduceTask.java:2829)
>>
>> thanks,
>> Venkat
>>
>> -------- Original Message --------  Subject: Re: Configure Hive in
>> Cluster  Date: Thu, 17 Jan 2013 17:23:03 +0530  From: venkatramanan
>> <venkatraman...@smartek21.com> <venkatraman...@smartek21.com>  Reply-To:
>> <user@hive.apache.org> <user@hive.apache.org>  To: 
>> <user@hive.apache.org><user@hive.apache.org>
>>
>>
>> Can you suggest me the mandatory hive parameters and clustering
>> configuration steps
>>
>> On Thursday 17 January 2013 12:56 PM, Nitin Pawar wrote:
>>
>> looks like a very small cluster with very limited memory to run mapreduce
>> jobs also number of map/reduce slots on nodes are less so at a time only
>> one map is running.
>>
>>  but still 15 min is a lot of time for 600MB memory
>>
>>
>> On Thu, Jan 17, 2013 at 12:47 PM, venkatramanan <
>> venkatraman...@smartek21.com> wrote:
>>
>>>  Below details are the cluster configuration
>>>
>>> Configured Capacity         : 82.8 GB
>>> DFS Used                          : 1.16 GB
>>> Non DFS Used                  : 31.95 GB
>>> DFS Remaining                : 49.69 GB
>>> DFS Used%                      : 1.4 %
>>> DFS Remaining%              : 60.01 %
>>> Live Nodes <http://localhost:50070/dfsnodelist.jsp?whatNodes=LIVE>
>>>                   : 2
>>> Dead Nodes <http://localhost:50070/dfsnodelist.jsp?whatNodes=DEAD>
>>>                 : 0
>>> Decommissioning 
>>> Nodes<http://localhost:50070/dfsnodelist.jsp?whatNodes=DECOMMISSIONING>: 0
>>> Number of Under-Replicated Blocks : 0
>>>
>>> My Select Query is:
>>>
>>> "select * from tweet where Id = 810;"
>>>
>>> This query takes 15 min to complete
>>>
>>>
>>>
>>> On Thursday 17 January 2013 12:29 PM, Nitin Pawar wrote:
>>>
>>> how many number of nodes you have for select query?
>>> whats your select query?
>>>
>>>  if its just a select * from table then it does not run any mapreduce
>>> job
>>>  so its just taking time to show data on your screen if you are using
>>> that query
>>>
>>>
>>> On Thu, Jan 17, 2013 at 12:24 PM, venkatramanan <
>>> venkatraman...@smartek21.com> wrote:
>>>
>>>>  I didnt set any hive parameters and my total table size is 610 MB only
>>>>
>>>>
>>>>
>>>> On Thursday 17 January 2013 12:11 PM, Nitin Pawar wrote:
>>>>
>>>> a bit more details on size of table and select query will help
>>>> also did you set any hive parameters ?
>>>>
>>>>
>>>> On Thu, Jan 17, 2013 at 12:12 PM, venkatramanan <
>>>> venkatraman...@smartek21.com> wrote:
>>>>
>>>>>  Hi All,
>>>>>
>>>>> Am Newbie in apache hive. I have create a table and thats points to
>>>>> the HDFS Folder path and its takes 15 min to execute the simple "*
>>>>> select*" stmt, Can anyone suggest me for a best practices and
>>>>> performance improvement on hive.
>>>>>
>>>>> Thanks in Advance
>>>>>
>>>>> Venkat
>>>>>
>>>>
>>>>
>>>>
>>>>  --
>>>> Nitin Pawar
>>>>
>>>>
>>>>
>>>
>>>
>>>  --
>>> Nitin Pawar
>>>
>>>
>>>
>>
>>
>>  --
>> Nitin Pawar
>>
>>
>>
>>
>>
>
>
>  --
> Nitin Pawar
>
>
>
> --
>
>


-- 
Nitin Pawar

Reply via email to