On 23 Feb 2016, at 08:22, Arunkumar Pillai
<[email protected]<mailto:[email protected]>> wrote:
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
at
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:129)
... 33 more
Caused by: java.lang.UnsatisfiedLinkError:
org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V
at
org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native Method)
at
org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49)
at
org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback.<init>(JniBasedUnixGroupsMappingWithFallback.java:39)
looks supiciously like there's a mismatch between the libhadoop.so on your path
and the Hadoop version. Is there a Hadoop 2.6 installed on the same system?
you could try to skip that bit of JNI code by switching to the shell:
spark.hadoop.security.group.mapping
org.apache.hadoop.security.ShellBasedUnixGroupsMapping
...but that will just postpone the problem
Best to find all copies of libhadoop.so on your filesystem, and make sure the
one that gets loaded is the Hadoop 2.7 one