there should be a hive flat table after step 1, can you figure out how many rows exist in the flat table , using hive shell?
On Fri, Jul 24, 2015 at 2:29 PM, Vineet Mishra <[email protected]> wrote: > kylin.hdfs.working.dir=/tmp > > Well usually the cube building runs smoothly without any such unconditional > issue, its this time when I am having a lookup table with join on a > particular column I am landing into the issue. > > Any quick resolve would be highly appreciated! > > Thanks, > > On Fri, Jul 24, 2015 at 11:55 AM, Maliakkal Padmanabhan, Aroop < > [email protected]> wrote: > > > Have seen these issues after a cleanup of hdfs:///tmp . > > What is the value of kylin.hdfs.working.dir in your cluster ? > > > > Thanks, > > /Aroop > > > > > > > > On 7/24/15, 11:21 AM, "Vineet Mishra" <[email protected]> wrote: > > > > >Let me mention the Lookup is Hierarchical. > > > > > >Thanks, > > > > > >On Fri, Jul 24, 2015 at 11:12 AM, Vineet Mishra <[email protected] > > > > >wrote: > > > > > >> Hi Li and Shi, > > >> > > >> I am building a cube with some lookup table in between and getting > > >> exception at third step of cube build i.e Build Dimension Dictionary > > >>with > > >> exception saying > > >> > > >> java.io.FileNotFoundException: File does not exist: > > >> > > > >>/tmp/kylin-5a2ea405-24a2-45ed-958e-2a7fddd8cc97/sc_o2s_metrics_verified12 > > >>3455/fact_distinct_columns/SC > > >> at > > >> > > > >>org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSys > > >>tem.java:1093) > > >> at > > >> > > > >>org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSys > > >>tem.java:1085) > > >> at > > >> > > > >>org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolve > > >>r.java:81) > > >> at > > >> > > > >>org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFil > > >>eSystem.java:1085) > > >> at > > >>org.apache.kylin.dict.lookup.FileTable.getSignature(FileTable.java:62) > > >> at > > >> > > > >>org.apache.kylin.dict.DictionaryManager.buildDictionary(DictionaryManager > > >>.java:164) > > >> at > > >>org.apache.kylin.cube.CubeManager.buildDictionary(CubeManager.java:154) > > >> at > > >> > > > >>org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(Dictionar > > >>yGeneratorCLI.java:53) > > >> at > > >> > > > >>org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(Dictionar > > >>yGeneratorCLI.java:42) > > >> at > > >> > > > >>org.apache.kylin.job.hadoop.dict.CreateDictionaryJob.run(CreateDictionary > > >>Job.java:53) > > >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > >> at > > >> > > > >>org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecu > > >>table.java:63) > > >> at > > >> > > > >>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecuta > > >>ble.java:107) > > >> at > > >> > > > >>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultCha > > >>inedExecutable.java:50) > > >> at > > >> > > > >>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecuta > > >>ble.java:107) > > >> at > > >> > > > >>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Defau > > >>ltScheduler.java:132) > > >> at > > >> > > > >>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java > > >>:1145) > > >> at > > >> > > > >>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.jav > > >>a:615) > > >> at java.lang.Thread.run(Thread.java:745) > > >> > > >> It will be great if you could help out in getting this issue resolved. > > >> > > >> URGENT CALL! > > >> > > >> Thanks, > > >> > > > > > -- Regards, *Bin Mahone | 马洪宾* Apache Kylin: http://kylin.io Github: https://github.com/binmahone
