Hubert STEFANI created KYLIN-3644:
-------------------------------------

             Summary: NumberFormatExcetion on null values when building cube 
with Spark
                 Key: KYLIN-3644
                 URL: https://issues.apache.org/jira/browse/KYLIN-3644
             Project: Kylin
          Issue Type: Bug
          Components: Spark Engine
    Affects Versions: v2.5.0
            Reporter: Hubert STEFANI
         Attachments: 01_overview_table.jpg, 02_dimension_cube.jpg, 
03_measure_cube.jpg

We encounter an error any time we try to build a cube with the following steps :
 * upload a csv on AWS S3 with following characteristics : the column on which 
the measure will be defined has some null values (Cf. attachment)
 * create a hive table with spark
 * create a model on  top of this table,
 * create a cube with a SUM measure
 * chose Spark as Engine
 * Launch build

Result : The build process fails at '{color:#4383b4}#7 Step Name: {color}Build 
Cube with Spark' with the following error :

 

""""""

18/10/23 09:25:39 INFO scheduler.DAGScheduler: Job 0 failed: 
saveAsNewAPIHadoopDataset at SparkCubingByLayer.java:253, took 7,277136 s

Exception in thread "main" java.lang.RuntimeException: error execute 
org.apache.kylin.engine.spark.SparkCubingByLayer. Root cause: Job aborted due 
to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost 
task 0.3 in stage 0.0 (TID 4, ip-172-31-35-113.eu-west-1.compute.internal, 
executor 4): java.lang.NumberFormatException: For input string: "\N"

    at sun.misc.FloatingDecimal.readJavaFormatString(FloatingDecimal.java:2043)

    at sun.misc.FloatingDecimal.parseDouble(FloatingDecimal.java:110)

    at java.lang.Double.parseDouble(Double.java:538)

    at 
org.apache.kylin.measure.basic.DoubleIngester.valueOf(DoubleIngester.java:38)

    at 
org.apache.kylin.measure.basic.DoubleIngester.valueOf(DoubleIngester.java:28)

    at 
org.apache.kylin.engine.mr.common.BaseCuboidBuilder.buildValueOf(BaseCuboidBuilder.java:162)

    at 
org.apache.kylin.engine.mr.common.BaseCuboidBuilder.buildValueObjects(BaseCuboidBuilder.java:127)

    at 
org.apache.kylin.engine.spark.SparkCubingByLayer$EncodeBaseCuboid.call(SparkCubingByLayer.java:297)

    at 
org.apache.kylin.engine.spark.SparkCubingByLayer$EncodeBaseCuboid.call(SparkCubingByLayer.java:257)

    at 
org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)

    at 
org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)

"""""

Note 1: the build  process is OK when run with Map/Reduce Engine.

Note 2: the error doesn't seem to be related to AWS environment.

 

Sample of csv :

ID;CATEGORIE;TEL;MONTANT;MAGASIN;MATRICULE;VILLE;

970;161;6-98-6-6-42;838.47034;Magasin_19;Client_Matricule_28;MARSEILLE;
971;89;62-15-2-64-86;;;Client_Matricule_1;LYON;
972;87;17-64-97-74-42;;;Client_Matricule_105;ORBEC;
973;174;79-33-90-0-55;;Magasin_7;Client_Matricule_55;AJACCIO;
974;172;89-95-71-6-49;141.64174;Magasin_9;Client_Matricule_105;BASTIA;
975;83;7-27-95-28-7;897.28204;;Client_Matricule_199;AJACCIO;
976;170;67-72-18-29-34;164.07967;Magasin_3;Client_Matricule_137;LILLE;
977;130;14-69-4-23-27;1928.9557;Magasin_1;Client_Matricule_17;NOMNOM;
978;43;55-91-84-98-49;891.2691;Magasin_0;Client_Matricule_22;NOMNOM;
979;117;98-96-0-54-39;1636.3994;Magasin_9;Client_Matricule_142;MARSEILLE;
980;163;37-55-76-53-38;;;Client_Matricule_64;NEWYORK;
981;106;32-40-6-46-15;;Magasin_2;Client_Matricule_158;NOMNOM;
982;56;95-60-83-89-90;;;Client_Matricule_102;NOMNOM;
983;168;21-56-62-0-58;;;Client_Matricule_160;NOMNOM;
984;154;92-67-37-94-60;;;Client_Matricule_137;PARIS;

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to