Hi: Recently, in the process of transforming kylin, I found a problem. My kylin's version is 2.6.2, the building engine used is spark . When building the dimensional dictionary, kylin server will start a shell process separately to start the spark program. After spark creates the dictionary for each field, it stores the dictionary information on hdfs, and then updates the cubename.json file under the /cube path on hdfs. But in the process of creating HTable later, the information of tablename.json will be read again, and then updated. The problem is that the creation of Htable is in the process of kylin server, so the ResourceStore instance is different from the spark program.In the process of creating a dictionary, spark only updates the information on hdfs, but does not update metadata in other places, such as mysql. The ResourceStore instance in kylin server is in the form of JDBC, resulting in the contents of the jdbc json file and the file on hdfs are different.Will this be a problem?
Thanks.
