Vikas

Can you be more specific. What to check for in Hive logs.

Thanks

Regards

Anand

Sent from my iPhone

> On 15-May-2015, at 2:41 pm, Vikas Parashar <[email protected]> wrote:
> 
> Hi Anand, 
> 
> It seems your namenode is working fine. I can't see any "safemode" related 
> logs in your namenode file. Kindly check it hive logs as well.
> 
>> On Fri, May 15, 2015 at 12:40 PM, Anand Murali <[email protected]> wrote:
>> Vikas:
>> 
>> Please find attached. At this time I would like to tell you that with the 
>> current installation, I am able to run mapreduce jobs and pig scripts 
>> without any installation errors. So please, any suggestions made should not 
>> break and cascade other installations.
>> 
>> Thanks
>> 
>> Regards,
>>  
>> Anand Murali  
>> 11/7, 'Anand Vihar', Kandasamy St, Mylapore
>> Chennai - 600 004, India
>> Ph: (044)- 28474593/ 43526162 (voicemail)
>> 
>> 
>> 
>> On Friday, May 15, 2015 12:31 PM, Kiran Dangeti <[email protected]> 
>> wrote:
>> 
>> 
>> Anand,
>> Sometimes it will error out due some resources are not available. So stop 
>> and start the hadoop cluster and see
>> On May 15, 2015 12:24 PM, "Anand Murali" <[email protected]> wrote:
>> Dear All:
>> 
>> I am running Hadoop-2.6 (pseudo mode) on Ubuntu 15.04, and trying to connect 
>> Hive to it after installation. I run . .hadoop as start-up script which 
>> contain environment variables setting. Find below
>> 
>> . ,hadoop
>> export HADOOP_HOME=/home/anand_vihar/hadoop-2.6.0
>> export JAVA_HOME=/home/anand_vihar/jdk1.7.0_75/
>> export HADOOP\_PREFIX=/home/anand_vihar/hadoop-2.6.0
>> export HADOOP_INSTALL=/home/anand_vihar/hadoop-2.6.0
>> export PIG_HOME=/home/anand_vihar/pig-0.14.0
>> export PIG_INSTALL=/home/anand_vihar/pig-0.14.0
>> export PIG_CLASSPATH=/home/anand_vihar/hadoop-2.6.0/etc/hadoop/
>> export HIVE_HOME=/home/anand_vihar/hive-1.1.0
>> export HIVE_INSTALL=/home/anand_vihar/hive-1.1.0
>> export 
>> PATH=$PATH:$HADOOP_INSTALL/bin:$HADOOP_INSTALL/sbin:$HADOOP_HOME:$JAVA_HOME:$PIG_INSTALL/bin:$PIG_CLASSPATH:$HIVE_HOME:$HIVE_INSTALL/bin
>> echo $HADOOP_HOME
>> echo $JAVA_HOME
>> echo $HADOOP_INSTALL
>> echo $PIG_HOME
>> echo $PIG_INSTALL
>> echo $PIG_CLASSPATH
>> echo $HIVE_HOME
>> echo $PATH
>> 
>> 
>> Error
>> 
>> anand_vihar@Latitude-E5540:~$ hive
>> 
>> Logging initialized using configuration in 
>> jar:file:/home/anand_vihar/hive-1.1.0/lib/hive-common-1.1.0.jar!/hive-log4j.properties
>> SLF4J: Class path contains multiple SLF4J bindings.
>> SLF4J: Found binding in 
>> [jar:file:/home/anand_vihar/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in 
>> [jar:file:/home/anand_vihar/hive-1.1.0/lib/hive-jdbc-1.1.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
>> explanation.
>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> Exception in thread "main" java.lang.RuntimeException: 
>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.SafeModeException):
>>  Cannot create directory 
>> /tmp/hive/anand_vihar/a9eb2cf7-9890-4ec3-af6c-ae0c40d9e9d7. Name node is in 
>> safe mode.
>> The reported blocks 2 has reached the threshold 0.9990 of total blocks 2. 
>> The number of live datanodes 1 has reached the minimum number 0. In safe 
>> mode extension. Safe mode will be turned off automatically in 6 seconds.
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:4216)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:4191)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:813)
>>     at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:600)
>>     at 
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>>     at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
>>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
>>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
>>     at java.security.AccessController.doPrivileged(Native Method)
>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>     at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
>> 
>>     at 
>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:472)
>>     at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:671)
>>     at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>     at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>     at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>     at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>> Caused by: 
>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.SafeModeException):
>>  Cannot create directory 
>> /tmp/hive/anand_vihar/a9eb2cf7-9890-4ec3-af6c-ae0c40d9e9d7. Name node is in 
>> safe mode.
>> The reported blocks 2 has reached the threshold 0.9990 of total blocks 2. 
>> The number of live datanodes 1 has reached the minimum number 0. In safe 
>> mode extension. Safe mode will be turned off automatically in 6 seconds.
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1364)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:4216)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:4191)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:813)
>>     at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:600)
>>     at 
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>>     at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
>>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
>>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
>>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
>>     at java.security.AccessController.doPrivileged(Native Method)
>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>     at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
>> 
>>     at org.apache.hadoop.ipc.Client.call(Client.java:1468)
>>     at org.apache.hadoop.ipc.Client.call(Client.java:1399)
>>     at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
>>     at com.sun.proxy.$Proxy13.mkdirs(Unknown Source)
>>     at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>     at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>     at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
>>     at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>>     at com.sun.proxy.$Proxy14.mkdirs(Unknown Source)
>>     at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2753)
>>     at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2724)
>>     at 
>> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
>>     at 
>> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
>>     at 
>> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>>     at 
>> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
>>     at 
>> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
>>     at 
>> org.apache.hadoop.hive.ql.session.SessionState.createPath(SessionState.java:584)
>>     at 
>> org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:526)
>>     at 
>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:458)
>>     ... 8 more
>> 
>> Can somebody advise.
>> 
>> Thanks
>>  
>> Anand Murali  
>> 11/7, 'Anand Vihar', Kandasamy St, Mylapore
>> Chennai - 600 004, India
>> Ph: (044)- 28474593/ 43526162 (voicemail)
> 

Reply via email to