I get the following ClosedByInterruptException often - but not always - when 
running a query with hive.exec.parallel=true. It seems to happen only when 2 MR 
jobs are being launched in parallel. I doubt I'm the first person to have seen 
this error in this scenario, but googling didn't help me. Any idea?

I am using a pre-release version of Hive 0.7.

Ended Job = job_201106080050_0253
Launching Job 2 out of 3
Launching Job 3 out of 3
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
java.io.IOException: Call to ip-10-78-97-240.ec2.internal/10.78.97.240:9001 
failed on local exception: java.nio.channels.ClosedByInterruptException
                at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
                at org.apache.hadoop.ipc.Client.call(Client.java:743)
                at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
                at org.apache.hadoop.mapred.$Proxy9.getNewJobId(Unknown Source)
                at 
org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:804)
                at 
org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:777)
                at 
org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:660)
                at 
org.apache.hadoop.hive.ql.exec.MapRedTask.execute(MapRedTask.java:121)
                at 
org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:128)
                at 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:55)
                at 
org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:47)
Caused by: java.nio.channels.ClosedByInterruptException
                at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:184)
                at 
sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:341)
                at 
org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
                at 
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
                at 
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
                at 
org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
                at 
java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
                at 
java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
                at java.io.DataOutputStream.flush(DataOutputStream.java:106)
                at 
org.apache.hadoop.ipc.Client$Connection.sendParam(Client.java:480)
                at org.apache.hadoop.ipc.Client.call(Client.java:721)
                ... 9 more
Job Submission failed with exception 'java.io.IOException(Call to 
ip-10-78-97-240.ec2.internal/10.78.97.240:9001 failed on local exception: 
java.nio.channels.ClosedByInterruptException)'
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.MapRedTask

Thanks.
Steven

Reply via email to