[20/49] carbondata git commit: [CARBONDATA-2025] Unify all path construction through CarbonTablePath static method

2018-02-28 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/eec8f18a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
--
diff --git 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
index c8a6b1d..184bf1b 100644
--- 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
+++ 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
@@ -28,7 +28,7 @@ import 
org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil
 import org.apache.carbondata.core.constants.CarbonCommonConstants
 import org.apache.carbondata.core.datastore.impl.FileFactory
 import org.apache.carbondata.core.statusmanager.{SegmentStatus, 
SegmentStatusManager}
-import org.apache.carbondata.core.util.path.CarbonStorePath
+import org.apache.carbondata.core.util.path.CarbonTablePath
 import org.apache.carbondata.events.OperationContext
 import org.apache.carbondata.processing.loading.model.CarbonLoadModel
 import org.apache.carbondata.processing.merger.{CarbonDataMergerUtil, 
CompactionType}
@@ -74,7 +74,7 @@ class AggregateDataMapCompactor(carbonLoadModel: 
CarbonLoadModel,
   "true")
 loadCommand.processData(sqlContext.sparkSession)
 val newLoadMetaDataDetails = SegmentStatusManager.readLoadMetadata(
-  carbonTable.getMetaDataFilepath, uuid)
+  carbonTable.getMetadataPath, uuid)
 val updatedLoadMetaDataDetails = newLoadMetaDataDetails collect {
   case load if loadMetaDataDetails.contains(load) =>
 load.setMergedLoadName(mergedLoadName)
@@ -83,11 +83,8 @@ class AggregateDataMapCompactor(carbonLoadModel: 
CarbonLoadModel,
 load
   case other => other
 }
-val carbonTablePath = CarbonStorePath
-  
.getCarbonTablePath(carbonLoadModel.getCarbonDataLoadSchema.getCarbonTable
-.getAbsoluteTableIdentifier)
-SegmentStatusManager
-  
.writeLoadDetailsIntoFile(carbonTablePath.getTableStatusFilePathWithUUID(uuid),
+SegmentStatusManager.writeLoadDetailsIntoFile(
+  CarbonTablePath.getTableStatusFilePathWithUUID(uuid),
 updatedLoadMetaDataDetails)
 
carbonLoadModel.setLoadMetadataDetails(updatedLoadMetaDataDetails.toList.asJava)
   } finally {

http://git-wip-us.apache.org/repos/asf/carbondata/blob/eec8f18a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
--
diff --git 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
index 6f08154..8d3110a 100644
--- 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
+++ 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
@@ -59,7 +59,7 @@ import org.apache.carbondata.core.mutate.CarbonUpdateUtil
 import org.apache.carbondata.core.scan.partition.PartitionUtil
 import org.apache.carbondata.core.statusmanager.{LoadMetadataDetails, 
SegmentStatus, SegmentStatusManager}
 import org.apache.carbondata.core.util.{ByteUtil, CarbonProperties, CarbonUtil}
-import org.apache.carbondata.core.util.path.{CarbonStorePath, CarbonTablePath}
+import org.apache.carbondata.core.util.path.CarbonTablePath
 import org.apache.carbondata.events.{OperationContext, OperationListenerBus}
 import org.apache.carbondata.processing.exception.DataLoadingException
 import org.apache.carbondata.processing.loading.FailureCauses
@@ -72,7 +72,7 @@ import 
org.apache.carbondata.processing.merger.{CarbonCompactionUtil, CarbonData
 import org.apache.carbondata.processing.util.{CarbonDataProcessorUtil, 
CarbonLoaderUtil}
 import org.apache.carbondata.spark.{DataLoadResultImpl, PartitionFactory, _}
 import org.apache.carbondata.spark.load._
-import org.apache.carbondata.spark.util.{CarbonScalaUtil, CommonUtil, 
DataLoadingUtil, Util}
+import org.apache.carbondata.spark.util.{CarbonScalaUtil, CommonUtil, Util}
 
 /**
  * This is the factory class which can create different RDD depends on user 
needs.
@@ -127,7 +127,7 @@ object CarbonDataRDDFactory {
   LOGGER.error("Not able to acquire the compaction lock for table " +
   s"${ carbonLoadModel.getDatabaseName }.${ 
carbonLoadModel.getTableName }")
   CarbonCompactionUtil
-  .createCompactionRequiredFile(carbonTable.getMetaDataFilepath, 
compactionType)
+  .createCompactionRequiredFile(carbonTable.getMetadataPath, 
compactionType)
   // throw exception only in case of DDL trigger.
   if 

[20/49] carbondata git commit: [CARBONDATA-2025] Unify all path construction through CarbonTablePath static method

2018-02-26 Thread jackylk
http://git-wip-us.apache.org/repos/asf/carbondata/blob/4d453d4b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
--
diff --git 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
index c8a6b1d..184bf1b 100644
--- 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
+++ 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/AggregateDataMapCompactor.scala
@@ -28,7 +28,7 @@ import 
org.apache.spark.sql.execution.command.preaaggregate.PreAggregateUtil
 import org.apache.carbondata.core.constants.CarbonCommonConstants
 import org.apache.carbondata.core.datastore.impl.FileFactory
 import org.apache.carbondata.core.statusmanager.{SegmentStatus, 
SegmentStatusManager}
-import org.apache.carbondata.core.util.path.CarbonStorePath
+import org.apache.carbondata.core.util.path.CarbonTablePath
 import org.apache.carbondata.events.OperationContext
 import org.apache.carbondata.processing.loading.model.CarbonLoadModel
 import org.apache.carbondata.processing.merger.{CarbonDataMergerUtil, 
CompactionType}
@@ -74,7 +74,7 @@ class AggregateDataMapCompactor(carbonLoadModel: 
CarbonLoadModel,
   "true")
 loadCommand.processData(sqlContext.sparkSession)
 val newLoadMetaDataDetails = SegmentStatusManager.readLoadMetadata(
-  carbonTable.getMetaDataFilepath, uuid)
+  carbonTable.getMetadataPath, uuid)
 val updatedLoadMetaDataDetails = newLoadMetaDataDetails collect {
   case load if loadMetaDataDetails.contains(load) =>
 load.setMergedLoadName(mergedLoadName)
@@ -83,11 +83,8 @@ class AggregateDataMapCompactor(carbonLoadModel: 
CarbonLoadModel,
 load
   case other => other
 }
-val carbonTablePath = CarbonStorePath
-  
.getCarbonTablePath(carbonLoadModel.getCarbonDataLoadSchema.getCarbonTable
-.getAbsoluteTableIdentifier)
-SegmentStatusManager
-  
.writeLoadDetailsIntoFile(carbonTablePath.getTableStatusFilePathWithUUID(uuid),
+SegmentStatusManager.writeLoadDetailsIntoFile(
+  CarbonTablePath.getTableStatusFilePathWithUUID(uuid),
 updatedLoadMetaDataDetails)
 
carbonLoadModel.setLoadMetadataDetails(updatedLoadMetaDataDetails.toList.asJava)
   } finally {

http://git-wip-us.apache.org/repos/asf/carbondata/blob/4d453d4b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
--
diff --git 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
index 6f08154..8d3110a 100644
--- 
a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
+++ 
b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
@@ -59,7 +59,7 @@ import org.apache.carbondata.core.mutate.CarbonUpdateUtil
 import org.apache.carbondata.core.scan.partition.PartitionUtil
 import org.apache.carbondata.core.statusmanager.{LoadMetadataDetails, 
SegmentStatus, SegmentStatusManager}
 import org.apache.carbondata.core.util.{ByteUtil, CarbonProperties, CarbonUtil}
-import org.apache.carbondata.core.util.path.{CarbonStorePath, CarbonTablePath}
+import org.apache.carbondata.core.util.path.CarbonTablePath
 import org.apache.carbondata.events.{OperationContext, OperationListenerBus}
 import org.apache.carbondata.processing.exception.DataLoadingException
 import org.apache.carbondata.processing.loading.FailureCauses
@@ -72,7 +72,7 @@ import 
org.apache.carbondata.processing.merger.{CarbonCompactionUtil, CarbonData
 import org.apache.carbondata.processing.util.{CarbonDataProcessorUtil, 
CarbonLoaderUtil}
 import org.apache.carbondata.spark.{DataLoadResultImpl, PartitionFactory, _}
 import org.apache.carbondata.spark.load._
-import org.apache.carbondata.spark.util.{CarbonScalaUtil, CommonUtil, 
DataLoadingUtil, Util}
+import org.apache.carbondata.spark.util.{CarbonScalaUtil, CommonUtil, Util}
 
 /**
  * This is the factory class which can create different RDD depends on user 
needs.
@@ -127,7 +127,7 @@ object CarbonDataRDDFactory {
   LOGGER.error("Not able to acquire the compaction lock for table " +
   s"${ carbonLoadModel.getDatabaseName }.${ 
carbonLoadModel.getTableName }")
   CarbonCompactionUtil
-  .createCompactionRequiredFile(carbonTable.getMetaDataFilepath, 
compactionType)
+  .createCompactionRequiredFile(carbonTable.getMetadataPath, 
compactionType)
   // throw exception only in case of DDL trigger.
   if