Hi, Using the spark-shell, I can't sc.parallelize to get an RDD.
Looks like a bug. scala> sc.parallelize(Array("a","s","d")) java.lang.NullPointerException at <init>(<console>:17) at <init>(<console>:22) at <init>(<console>:24) at <init>(<console>:26) at <init>(<console>:28) at <init>(<console>:30) at <init>(<console>:32) at <init>(<console>:34) at <init>(<console>:36) at .<init>(<console>:40) at .<clinit>(<console>) at .<init>(<console>:11) at .<clinit>(<console>) at $export(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:629) at spark.repl.SparkIMain$Request$$anonfun$10.apply(SparkIMain.scala:890) at scala.tools.nsc.interpreter.Line$$anonfun$1.apply$mcV$sp(Line.scala:43) at scala.tools.nsc.io.package$$anon$2.run(package.scala:25) at java.lang.Thread.run(Thread.java:744)