Hi everybody!I have tried all solution i have found in users posts related to 
newAPIHadoopFile on the mailing list but with no success .Can anybody help me 
with the following problem?

I'm new into Spark and i'm trying to make a simple read into hdfs with a custom 
inputformat. but i'am having this problem :
[INFO] --- exec-maven-plugin:1.2.1:java (default-cli) @ simple-project 
---log4j:WARN No appenders could be found for logger 
(akka.event.slf4j.Slf4jEventHandler).log4j:WARN Please initialize the log4j 
system properly.log4j:WARN See 
http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
info.[WARNING]java.lang.reflect.InvocationTargetException        at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)   
     at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)        at 
org.codehaus.mojo.exec.ExecJavaMojo$1.run(ExecJavaMojo.java:297)        at 
java.lang.Thread.run(Thread.java:724)Caused by: 
org.apache.spark.SparkException: Job aborted: Task 0.0:0 failed more than 0 
times; aborting job java.io.IOException: File Data Format not coherent with 
configuration parameters        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:827)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:825)
        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)  
      at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)     
   at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:825)      
  at 
org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:440)    
    at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:502)
        at 
org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:157)[INFO]
 ------------------------------------------------------------------------[INFO] 
BUILD FAILURE[INFO] 
------------------------------------------------------------------------[INFO] 
Total time: 4.367s[INFO] Finished at: Wed Jan 22 14:42:00 CET 2014[INFO] Final 
Memory: 25M/439M[INFO] 
------------------------------------------------------------------------

my driver :package spark.test;/*** SimpleApp.java ***/import 
java.io.IOException;
import org.apache.hadoop.conf.Configuration;import 
org.apache.hadoop.io.Text;import org.apache.hadoop.mapreduce.Job;import 
org.apache.spark.api.java.JavaPairRDD;import 
org.apache.spark.api.java.JavaSparkContext;
public class SimpleApp {        public static void main(String[] args) throws 
IOException {                             String NameNodeFolder = 
"hdfs://localhost:8020/user/cjaber/";                           String logFile 
= NameNodeFolder+"input/CSVfile.csv"; // Should be some file on your system     
         JavaSparkContext sc = new JavaSparkContext("local", "Simple App",      
                         "$YOUR_SPARK_HOME", new 
String[]{"target/simple-project-1.0.jar"});                                     
        Configuration conf = new Configuration();               Job job= new 
Job(conf);         job.setInputFormatClass(SignalInputFormat.class);            
                   JavaPairRDD<Text,SignalWritable> data = 
sc.newAPIHadoopFile(logFile,SignalInputFormat.class, Text.class, 
SignalWritable.class,job.getConfiguration());
                System.out.println(data.count());                               
        }}

Best Regards;Chadi
                                                                                
  

Reply via email to