Hi,

Now it looks like hdfs permission problem.

org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
Permission denied: user=root, access=WRITE, inode="/user":hdfs:supergroup:
drwxr-xr-x

You'll need permission to access '/user' directory for 'root' user. ('root'
is not a super user in your hdfs).

Hope this helps.

Best,
moon

On Mon, Jul 13, 2015 at 9:36 AM Vadla, Karthik <karthik.va...@intel.com>
wrote:

>  Hi Moon,
>
>
>
> Yes it did set master property and exported Hadoop config
>
>
>
> Master= yarn-client
>
> export HADOOP_CONF_DIR =/etc/Hadoop/conf      (in zeppelin-env.sh file)
>
>
>
> It is throwing below error.
>
>
>
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
> Permission denied: user=root, access=WRITE,
> inode="/user":hdfs:supergroup:drwxr-xr-x
>
>        at
> org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6596)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6578)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6530)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:4334)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:4304)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:4277)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:852)
>
>        at
> org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.mkdirs(AuthorizationProviderProxyClientProtocol.java:321)
>
>        at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:601)
>
>        at
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>
>        at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
>
>        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
>
>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
>
>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
>
>        at java.security.AccessController.doPrivileged(Native Method)
>
>        at javax.security.auth.Subject.doAs(Subject.java:415)
>
>        at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>
>        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2038)
>
>
>
>        at org.apache.hadoop.ipc.Client.call(Client.java:1468)
>
>        at org.apache.hadoop.ipc.Client.call(Client.java:1399)
>
>        at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
>
>        at com.sun.proxy.$Proxy14.mkdirs(Unknown Source)
>
>        at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:539)
>
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
>        at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>
>        at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>
>        at java.lang.reflect.Method.invoke(Method.java:483)
>
>        at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
>
>        at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
>
>        at com.sun.proxy.$Proxy15.mkdirs(Unknown Source)
>
>        at
> org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2760)
>
>        at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2731)
>
>        at
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:870)
>
>        at
> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:866)
>
>        at
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
>
>        at
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:866)
>
>        at
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:859)
>
>        at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1817)
>
>        at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:597)
>
>        at
> org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:224)
>
>        at
> org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:384)
>
>        at
> org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:102)
>
>        at
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58)
>
>        at
> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:141)
>
>        at org.apache.spark.SparkContext.<init>(SparkContext.scala:381)
>
>        at
> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:301)
>
>        at
> org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:146)
>
>        at
> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:423)
>
>        at
> org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
>
>        at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
>
>        at
> org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:353)
>
>        at
> org.apache.zeppelin.spark.PySparkInterpreter.getJavaSparkContext(PySparkInterpreter.java:374)
>
>        at
> org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:140)
>
>        at
> org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
>
>        at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
>
>        at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
>
>        at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:275)
>
>        at org.apache.zeppelin.scheduler.Job.run(Job.java:170)
>
>        at
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118)
>
>        at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>
>        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
>        at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>
>        at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>
>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
>        at java.lang.Thread.run(Thread.java:745)
>
>
>
>
>
>
>
> Thanks
>
> Karthik
>
> *From:* moon soo Lee [mailto:m...@apache.org]
> *Sent:* Sunday, July 12, 2015 9:05 AM
> *To:* users@zeppelin.incubator.apache.org
> *Subject:* Re: Yarn configuration on Zeppelin
>
>
>
> Hi,
>
>
>
> Did you set 'master' property to 'yarn-client' in 'Interpreter' menu?
>
> You'll also need export HADOOP_CONF_DIR in bin/zeppelin-env.sh file.
>
>
>
> Hope this helps.
>
>
>
> Thanks,
>
> moon
>
>
>
> On Fri, Jul 10, 2015 at 1:26 PM Vadla, Karthik <karthik.va...@intel.com>
> wrote:
>
>  Hi All,
>
>
>
> I have built my zeppelin binaries with yarn profile. With below command in
> maven
>
> *mvn clean package -Pspark-1.3 -Ppyspark -Dhadoop.version=2.6.0-cdh5.4.2
> -Phadoop-2.6 -Pyarn –DskipTests*
>
>
>
> I have enabled *hive-on-spark* option in Cloudera manager and copied
> *hive-site.xml* to my *zeppelin conf/* folder.
>
> But still I’m not able to see any queries ran on spark with zeppelin
> notebook in my Yarn  ResourceManager Web UI (master)
> <http://master.trinity2.cluster.gao-nova:8088/> .
>
>
>
> Do I need to do any specific configuration. ?
>
>
>
> Reading some previous post I got some idea that zeppelin is using hive
> server-2 . Can anyone help me where I can find configuration folder and
> what files I need to copy.
>
>
>
> Appreciate your help
>
>
>
> Thanks
>
> Karthik Vadla
>
>

Reply via email to