I don't see any version flag for /usr/bin/jar, but I think I see the problem now; the openjdk version is 7, but javac -version gives 1.6.0_34; so spark was compiled with java 6 despite the system using jre 1.7. Thanks for the sanity check! Now I just need to find out why javac is downgraded on the system..
On 2/24/15, Sean Owen <so...@cloudera.com> wrote: > So you mean that the script is checking for this error, and takes it > as a sign that you compiled with java 6. > > Your command seems to confirm that reading the assembly jar does fail > on your system though. What version does the jar command show? are you > sure you don't have JRE 7 but JDK 6 installed? > > On Tue, Feb 24, 2015 at 11:02 PM, Mike Hynes <91m...@gmail.com> wrote: >> ./bin/compute-classpath.sh fails with error: >> >> $> jar -tf >> assembly/target/scala-2.10/spark-assembly-1.3.0-SNAPSHOT-hadoop1.0.4.jar >> nonexistent/class/path >> java.util.zip.ZipException: invalid CEN header (bad signature) >> at java.util.zip.ZipFile.open(Native Method) >> at java.util.zip.ZipFile.<init>(ZipFile.java:132) >> at java.util.zip.ZipFile.<init>(ZipFile.java:93) >> at sun.tools.jar.Main.list(Main.java:997) >> at sun.tools.jar.Main.run(Main.java:242) >> at sun.tools.jar.Main.main(Main.java:1167) >> >> However, I both compiled the distribution and am running spark with Java >> 1.7; >> $ java -version >> java version "1.7.0_75" >> OpenJDK Runtime Environment (IcedTea 2.5.4) >> (7u75-2.5.4-1~trusty1) >> OpenJDK 64-Bit Server VM (build 24.75-b04, mixed mode) >> on a system running Ubuntu: >> $ uname -srpov >> Linux 3.13.0-44-generic #73-Ubuntu SMP Tue Dec 16 00:22:43 UTC 2014 >> x86_64 GNU/Linux >> $ uname -srpo >> Linux 3.13.0-44-generic x86_64 GNU/Linux >> >> This problem was reproduced on Arch Linux: >> >> $ uname -srpo >> Linux 3.18.5-1-ARCH x86_64 GNU/Linux >> with >> $ java -version >> java version "1.7.0_75" >> OpenJDK Runtime Environment (IcedTea 2.5.4) (Arch Linux build >> 7.u75_2.5.4-1-x86_64) >> OpenJDK 64-Bit Server VM (build 24.75-b04, mixed mode) >> >> In both of these cases, the problem is not the java versioning; >> neither system even has a java 6 installation. This seems like a false >> positive to me in compute-classpath.sh. >> >> When I comment out the relevant lines in compute-classpath.sh, the >> scripts start-{master,slaves,...}.sh all run fine, and I have no >> problem launching applications. >> >> Could someone please offer some insight into this issue? >> >> Thanks, >> Mike >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> For additional commands, e-mail: dev-h...@spark.apache.org >> > -- Thanks, Mike --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org