http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableAddColumnCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableAddColumnCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableAddColumnCommand.scala index 22ff5c4..1f1e7bd 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableAddColumnCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableAddColumnCommand.scala @@ -25,7 +25,8 @@ import org.apache.spark.sql.hive.CarbonSessionCatalog import org.apache.spark.util.AlterTableUtil import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException -import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.features.TableOperation import org.apache.carbondata.core.locks.{ICarbonLock, LockUsage} import org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl @@ -39,11 +40,11 @@ private[sql] case class CarbonAlterTableAddColumnCommand( extends MetadataCommand { override def processMetadata(sparkSession: SparkSession): Seq[Row] = { - val LOGGER: LogService = LogServiceFactory.getLogService(this.getClass.getCanonicalName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) val tableName = alterTableAddColumnsModel.tableName val dbName = alterTableAddColumnsModel.databaseName .getOrElse(sparkSession.catalog.currentDatabase) - LOGGER.audit(s"Alter table add columns request has been received for $dbName.$tableName") + Audit.log(LOGGER, s"Alter table add columns request has been received for $dbName.$tableName") val locksToBeAcquired = List(LockUsage.METADATA_LOCK, LockUsage.COMPACTION_LOCK) var locks = List.empty[ICarbonLock] var timeStamp = 0L @@ -104,10 +105,10 @@ private[sql] case class CarbonAlterTableAddColumnCommand( carbonTable, alterTableAddColumnsModel) OperationListenerBus.getInstance.fireEvent(alterTablePostExecutionEvent, operationContext) LOGGER.info(s"Alter table for add columns is successful for table $dbName.$tableName") - LOGGER.audit(s"Alter table for add columns is successful for table $dbName.$tableName") + Audit.log(LOGGER, s"Alter table for add columns is successful for table $dbName.$tableName") } catch { case e: Exception => - LOGGER.error(e, "Alter table add columns failed") + LOGGER.error("Alter table add columns failed", e) if (newCols.nonEmpty) { LOGGER.info("Cleaning up the dictionary files as alter table add operation failed") new AlterTableDropColumnRDD(sparkSession,
http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDataTypeChangeCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDataTypeChangeCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDataTypeChangeCommand.scala index 9ce79e9..716b9c9 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDataTypeChangeCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDataTypeChangeCommand.scala @@ -25,7 +25,8 @@ import org.apache.spark.sql.hive.CarbonSessionCatalog import org.apache.spark.util.AlterTableUtil import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException -import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.features.TableOperation import org.apache.carbondata.core.locks.{ICarbonLock, LockUsage} import org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl @@ -40,11 +41,12 @@ private[sql] case class CarbonAlterTableDataTypeChangeCommand( extends MetadataCommand { override def processMetadata(sparkSession: SparkSession): Seq[Row] = { - val LOGGER: LogService = LogServiceFactory.getLogService(this.getClass.getCanonicalName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) val tableName = alterTableDataTypeChangeModel.tableName val dbName = alterTableDataTypeChangeModel.databaseName .getOrElse(sparkSession.catalog.currentDatabase) - LOGGER.audit(s"Alter table change data type request has been received for $dbName.$tableName") + Audit.log(LOGGER, + s"Alter table change data type request has been received for $dbName.$tableName") val locksToBeAcquired = List(LockUsage.METADATA_LOCK, LockUsage.COMPACTION_LOCK) var locks = List.empty[ICarbonLock] // get the latest carbon table and check for column existence @@ -68,7 +70,7 @@ private[sql] case class CarbonAlterTableDataTypeChangeCommand( val columnName = alterTableDataTypeChangeModel.columnName val carbonColumns = carbonTable.getCreateOrderColumn(tableName).asScala.filter(!_.isInvisible) if (!carbonColumns.exists(_.getColName.equalsIgnoreCase(columnName))) { - LOGGER.audit(s"Alter table change data type request has failed. " + + Audit.log(LOGGER, s"Alter table change data type request has failed. " + s"Column $columnName does not exist") throwMetadataException(dbName, tableName, s"Column does not exist: $columnName") } @@ -76,7 +78,7 @@ private[sql] case class CarbonAlterTableDataTypeChangeCommand( if (carbonColumn.size == 1) { validateColumnDataType(alterTableDataTypeChangeModel.dataTypeInfo, carbonColumn.head) } else { - LOGGER.audit(s"Alter table change data type request has failed. " + + Audit.log(LOGGER, s"Alter table change data type request has failed. " + s"Column $columnName is invalid") throwMetadataException(dbName, tableName, s"Invalid Column: $columnName") } @@ -116,7 +118,8 @@ private[sql] case class CarbonAlterTableDataTypeChangeCommand( alterTableDataTypeChangeModel) OperationListenerBus.getInstance.fireEvent(alterTablePostExecutionEvent, operationContext) LOGGER.info(s"Alter table for data type change is successful for table $dbName.$tableName") - LOGGER.audit(s"Alter table for data type change is successful for table $dbName.$tableName") + Audit.log(LOGGER, + s"Alter table for data type change is successful for table $dbName.$tableName") } catch { case e: Exception => LOGGER.error("Alter table change datatype failed : " + e.getMessage) http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDropColumnCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDropColumnCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDropColumnCommand.scala index 1dbe28c..d601ed6 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDropColumnCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableDropColumnCommand.scala @@ -26,14 +26,13 @@ import org.apache.spark.sql.hive.CarbonSessionCatalog import org.apache.spark.util.AlterTableUtil import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException -import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.features.TableOperation import org.apache.carbondata.core.locks.{ICarbonLock, LockUsage} import org.apache.carbondata.core.metadata.converter.ThriftWrapperSchemaConverterImpl import org.apache.carbondata.core.metadata.encoder.Encoding import org.apache.carbondata.core.metadata.schema.table.CarbonTable -import org.apache.carbondata.core.util.CarbonUtil -import org.apache.carbondata.core.util.path.CarbonTablePath import org.apache.carbondata.events.{AlterTableDropColumnPostEvent, AlterTableDropColumnPreEvent, OperationContext, OperationListenerBus} import org.apache.carbondata.format.SchemaEvolutionEntry import org.apache.carbondata.spark.rdd.AlterTableDropColumnRDD @@ -43,11 +42,11 @@ private[sql] case class CarbonAlterTableDropColumnCommand( extends MetadataCommand { override def processMetadata(sparkSession: SparkSession): Seq[Row] = { - val LOGGER: LogService = LogServiceFactory.getLogService(this.getClass.getCanonicalName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) val tableName = alterTableDropColumnModel.tableName val dbName = alterTableDropColumnModel.databaseName .getOrElse(sparkSession.catalog.currentDatabase) - LOGGER.audit(s"Alter table drop columns request has been received for $dbName.$tableName") + Audit.log(LOGGER, s"Alter table drop columns request has been received for $dbName.$tableName") var locks = List.empty[ICarbonLock] var timeStamp = 0L val locksToBeAcquired = List(LockUsage.METADATA_LOCK, LockUsage.COMPACTION_LOCK) @@ -162,7 +161,7 @@ private[sql] case class CarbonAlterTableDropColumnCommand( OperationListenerBus.getInstance().fireEvent(alterTableDropColumnPostEvent, operationContext) LOGGER.info(s"Alter table for drop columns is successful for table $dbName.$tableName") - LOGGER.audit(s"Alter table for drop columns is successful for table $dbName.$tableName") + Audit.log(LOGGER, s"Alter table for drop columns is successful for table $dbName.$tableName") } catch { case e: Exception => LOGGER.error("Alter table drop columns failed : " + e.getMessage) http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableRenameCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableRenameCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableRenameCommand.scala index f1e17bd..a1c68a3 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableRenameCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/schema/CarbonAlterTableRenameCommand.scala @@ -27,12 +27,11 @@ import org.apache.spark.sql.hive.{CarbonRelation, CarbonSessionCatalog} import org.apache.spark.util.AlterTableUtil import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException -import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.datamap.DataMapStoreManager -import org.apache.carbondata.core.datastore.impl.FileFactory import org.apache.carbondata.core.exception.ConcurrentOperationException import org.apache.carbondata.core.features.TableOperation -import org.apache.carbondata.core.locks.{ICarbonLock, LockUsage} import org.apache.carbondata.core.metadata.CarbonTableIdentifier import org.apache.carbondata.core.metadata.schema.table.{CarbonTable, DataMapSchema} import org.apache.carbondata.core.statusmanager.SegmentStatusManager @@ -44,7 +43,7 @@ private[sql] case class CarbonAlterTableRenameCommand( extends MetadataCommand { override def processMetadata(sparkSession: SparkSession): Seq[Nothing] = { - val LOGGER: LogService = LogServiceFactory.getLogService(this.getClass.getCanonicalName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) val oldTableIdentifier = alterTableRenameModel.oldTableIdentifier val newTableIdentifier = alterTableRenameModel.newTableIdentifier val oldDatabaseName = oldTableIdentifier.database @@ -61,14 +60,14 @@ private[sql] case class CarbonAlterTableRenameCommand( } val oldTableName = oldTableIdentifier.table.toLowerCase val newTableName = newTableIdentifier.table.toLowerCase - LOGGER.audit(s"Rename table request has been received for $oldDatabaseName.$oldTableName") + Audit.log(LOGGER, s"Rename table request has been received for $oldDatabaseName.$oldTableName") LOGGER.info(s"Rename table request has been received for $oldDatabaseName.$oldTableName") val metastore = CarbonEnv.getInstance(sparkSession).carbonMetastore val relation: CarbonRelation = metastore.lookupRelation(oldTableIdentifier.database, oldTableName)(sparkSession) .asInstanceOf[CarbonRelation] if (relation == null) { - LOGGER.audit(s"Rename table request has failed. " + + Audit.log(LOGGER, s"Rename table request has failed. " + s"Table $oldDatabaseName.$oldTableName does not exist") throwMetadataException(oldDatabaseName, oldTableName, "Table does not exist") } @@ -163,13 +162,13 @@ private[sql] case class CarbonAlterTableRenameCommand( OperationListenerBus.getInstance().fireEvent(alterTableRenamePostEvent, operationContext) sparkSession.catalog.refreshTable(newIdentifier.quotedString) - LOGGER.audit(s"Table $oldTableName has been successfully renamed to $newTableName") + Audit.log(LOGGER, s"Table $oldTableName has been successfully renamed to $newTableName") LOGGER.info(s"Table $oldTableName has been successfully renamed to $newTableName") } catch { case e: ConcurrentOperationException => throw e case e: Exception => - LOGGER.error(e, "Rename table failed: " + e.getMessage) + LOGGER.error("Rename table failed: " + e.getMessage, e) if (carbonTable != null) { AlterTableUtil.revertRenameTableChanges( newTableName, http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableAsSelectCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableAsSelectCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableAsSelectCommand.scala index 19c265d..3252f1d 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableAsSelectCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableAsSelectCommand.scala @@ -24,7 +24,9 @@ import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan import org.apache.spark.sql.execution.command.AtomicRunnableCommand import org.apache.spark.sql.execution.command.management.CarbonInsertIntoCommand +import org.apache.carbondata.api.CarbonStore.LOGGER import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.metadata.schema.table.TableInfo /** @@ -53,12 +55,12 @@ case class CarbonCreateTableAsSelectCommand( databaseOpt = Some(tableInfo.getDatabaseName) } val dbName = CarbonEnv.getDatabaseName(databaseOpt)(sparkSession) - LOGGER.audit(s"Request received for CTAS for $dbName.$tableName") + Audit.log(LOGGER, s"Request received for CTAS for $dbName.$tableName") // check if table already exists if (sparkSession.sessionState.catalog.listTables(dbName) .exists(_.table.equalsIgnoreCase(tableName))) { if (!ifNotExistsSet) { - LOGGER.audit( + Audit.log(LOGGER, s"Table creation with Database name [$dbName] and Table name [$tableName] failed. " + s"Table [$tableName] already exists under database [$dbName]") throw new TableAlreadyExistsException(dbName, tableName) @@ -96,7 +98,7 @@ case class CarbonCreateTableAsSelectCommand( val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) loadCommand.processData(sparkSession) val carbonTable = loadCommand.relation.carbonTable - LOGGER.audit(s"CTAS operation completed successfully for " + + Audit.log(LOGGER, s"CTAS operation completed successfully for " + s"${carbonTable.getDatabaseName}.${carbonTable.getTableName}") } Seq.empty http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableCommand.scala index 42ea0bd..5d039bf 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonCreateTableCommand.scala @@ -25,7 +25,9 @@ import org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException import org.apache.spark.sql.execution.SQLExecution.EXECUTION_ID_KEY import org.apache.spark.sql.execution.command.MetadataCommand +import org.apache.carbondata.api.CarbonStore.LOGGER import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.constants.CarbonCommonConstants import org.apache.carbondata.core.datastore.compression.CompressorFactory import org.apache.carbondata.core.datastore.impl.FileFactory @@ -60,12 +62,12 @@ case class CarbonCreateTableCommand( // set dbName and tableUnique Name in the table info tableInfo.setDatabaseName(dbName) tableInfo.setTableUniqueName(CarbonTable.buildUniqueName(dbName, tableName)) - LOGGER.audit(s"Creating Table with Database name [$dbName] and Table name [$tableName]") + Audit.log(LOGGER, s"Creating Table with Database name [$dbName] and Table name [$tableName]") val isTransactionalTable = tableInfo.isTransactionalTable if (sparkSession.sessionState.catalog.listTables(dbName) .exists(_.table.equalsIgnoreCase(tableName))) { if (!ifNotExistsSet) { - LOGGER.audit( + Audit.log(LOGGER, s"Table creation with Database name [$dbName] and Table name [$tableName] failed. " + s"Table [$tableName] already exists under database [$dbName]") throw new TableAlreadyExistsException(dbName, tableName) @@ -178,15 +180,15 @@ case class CarbonCreateTableCommand( case _: Exception => // No operation } val msg = s"Create table'$tableName' in database '$dbName' failed" - LOGGER.audit(msg.concat(", ").concat(e.getMessage)) - LOGGER.error(e, msg) + Audit.log(LOGGER, msg.concat(", ").concat(e.getMessage)) + LOGGER.error(msg, e) throwMetadataException(dbName, tableName, msg.concat(", ").concat(e.getMessage)) } } val createTablePostExecutionEvent: CreateTablePostExecutionEvent = CreateTablePostExecutionEvent(sparkSession, tableIdentifier) OperationListenerBus.getInstance.fireEvent(createTablePostExecutionEvent, operationContext) - LOGGER.audit(s"Table created with Database name [$dbName] and Table name [$tableName]") + Audit.log(LOGGER, s"Table created with Database name [$dbName] and Table name [$tableName]") } Seq.empty } http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonDropTableCommand.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonDropTableCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonDropTableCommand.scala index e4b298f..e8cb689 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonDropTableCommand.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/table/CarbonDropTableCommand.scala @@ -26,11 +26,11 @@ import org.apache.spark.sql.catalyst.analysis.NoSuchTableException import org.apache.spark.sql.execution.command.AtomicRunnableCommand import org.apache.spark.sql.execution.command.datamap.CarbonDropDataMapCommand -import org.apache.carbondata.common.logging.{LogService, LogServiceFactory} +import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.cache.dictionary.ManageDictionaryAndBTree import org.apache.carbondata.core.constants.CarbonCommonConstants import org.apache.carbondata.core.datamap.DataMapStoreManager -import org.apache.carbondata.core.datamap.status.DataMapStatusManager import org.apache.carbondata.core.datastore.impl.FileFactory import org.apache.carbondata.core.exception.ConcurrentOperationException import org.apache.carbondata.core.locks.{CarbonLockFactory, CarbonLockUtil, ICarbonLock, LockUsage} @@ -51,7 +51,7 @@ case class CarbonDropTableCommand( var childDropDataMapCommands : Seq[CarbonDropDataMapCommand] = Seq.empty override def processMetadata(sparkSession: SparkSession): Seq[Row] = { - val LOGGER: LogService = LogServiceFactory.getLogService(this.getClass.getCanonicalName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getCanonicalName) val dbName = databaseNameOp.getOrElse(sparkSession.catalog.currentDatabase) val carbonLocks: scala.collection.mutable.ListBuffer[ICarbonLock] = ListBuffer() @@ -71,7 +71,7 @@ case class CarbonDropTableCommand( if (SegmentStatusManager.isLoadInProgressInTable(carbonTable)) { throw new ConcurrentOperationException(carbonTable, "loading", "drop table") } - LOGGER.audit(s"Deleting table [$tableName] under database [$dbName]") + Audit.log(LOGGER, s"Deleting table [$tableName] under database [$dbName]") if (carbonTable.isStreamingSink) { // streaming table should acquire streaming.lock carbonLocks += CarbonLockUtil.getLockObject(identifier, LockUsage.STREAMING_LOCK) @@ -142,7 +142,7 @@ case class CarbonDropTableCommand( ifExistsSet, sparkSession) OperationListenerBus.getInstance.fireEvent(dropTablePostEvent, operationContext) - LOGGER.audit(s"Deleted table [$tableName] under database [$dbName]") + Audit.log(LOGGER, s"Deleted table [$tableName] under database [$dbName]") } catch { case ex: NoSuchTableException => @@ -153,7 +153,7 @@ case class CarbonDropTableCommand( throw ex case ex: Exception => val msg = s"Dropping table $dbName.$tableName failed: ${ex.getMessage}" - LOGGER.error(ex, msg) + LOGGER.error(msg, ex) throwMetadataException(dbName, tableName, msg) } finally { if (carbonLocks.nonEmpty) { http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/execution/strategy/DDLStrategy.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/strategy/DDLStrategy.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/strategy/DDLStrategy.scala index f9046f0..6e183b2 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/strategy/DDLStrategy.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/strategy/DDLStrategy.scala @@ -54,8 +54,7 @@ object MatchCreateDataSourceTable { } class DDLStrategy(sparkSession: SparkSession) extends SparkStrategy { - val LOGGER: LogService = - LogServiceFactory.getLogService(this.getClass.getName) + val LOGGER = LogServiceFactory.getLogService(this.getClass.getName) def apply(plan: LogicalPlan): Seq[SparkPlan] = { plan match { case LoadDataCommand(identifier, path, isLocal, isOverwrite, partition) http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala index 1622724..5165526 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/sql/parser/CarbonSpark2SqlParser.scala @@ -37,7 +37,9 @@ import org.apache.spark.sql.execution.command.stream.{CarbonCreateStreamCommand, import org.apache.spark.sql.util.CarbonException import org.apache.spark.util.CarbonReflectionUtils +import org.apache.carbondata.api.CarbonStore.LOGGER import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.constants.CarbonCommonConstants import org.apache.carbondata.spark.CarbonOption import org.apache.carbondata.spark.util.{CarbonScalaUtil, CommonUtil} @@ -530,7 +532,7 @@ class CarbonSpark2SqlParser extends CarbonDDLSqlParser { if (name.startsWith("default.value.") && fields.count(p => p.column.equalsIgnoreCase(colName)) == 1) { LOGGER.error(s"Duplicate default value exist for new column: ${ colName }") - LOGGER.audit( + Audit.log(LOGGER, s"Validation failed for Create/Alter Table Operation " + s"for ${ table }. " + s"Duplicate default value exist for new column: ${ colName }") http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/integration/spark2/src/main/scala/org/apache/spark/util/AlterTableUtil.scala ---------------------------------------------------------------------- diff --git a/integration/spark2/src/main/scala/org/apache/spark/util/AlterTableUtil.scala b/integration/spark2/src/main/scala/org/apache/spark/util/AlterTableUtil.scala index cab9de5..20cffa7 100644 --- a/integration/spark2/src/main/scala/org/apache/spark/util/AlterTableUtil.scala +++ b/integration/spark2/src/main/scala/org/apache/spark/util/AlterTableUtil.scala @@ -29,8 +29,10 @@ import org.apache.spark.sql.catalyst.TableIdentifier import org.apache.spark.sql.hive.{CarbonRelation, CarbonSessionCatalog} import org.apache.spark.sql.hive.HiveExternalCatalog._ +import org.apache.carbondata.api.CarbonStore.LOGGER import org.apache.carbondata.common.exceptions.sql.MalformedCarbonCommandException import org.apache.carbondata.common.logging.LogServiceFactory +import org.apache.carbondata.common.logging.impl.Audit import org.apache.carbondata.core.constants.CarbonCommonConstants import org.apache.carbondata.core.datamap.DataMapStoreManager import org.apache.carbondata.core.datastore.block.SegmentPropertiesAndSchemaHolder @@ -68,7 +70,7 @@ object AlterTableUtil { .lookupRelation(Option(dbName), tableName)(sparkSession) .asInstanceOf[CarbonRelation] if (relation == null) { - LOGGER.audit(s"Alter table request has failed. " + + Audit.log(LOGGER, s"Alter table request has failed. " + s"Table $dbName.$tableName does not exist") sys.error(s"Table $dbName.$tableName does not exist") } @@ -292,7 +294,7 @@ object AlterTableUtil { (sparkSession: SparkSession, catalog: CarbonSessionCatalog): Unit = { val tableName = tableIdentifier.table val dbName = tableIdentifier.database.getOrElse(sparkSession.catalog.currentDatabase) - LOGGER.audit(s"Alter table newProperties request has been received for $dbName.$tableName") + Audit.log(LOGGER, s"Alter table newProperties request has been received for $dbName.$tableName") val locksToBeAcquired = List(LockUsage.METADATA_LOCK, LockUsage.COMPACTION_LOCK) var locks = List.empty[ICarbonLock] try { @@ -378,10 +380,10 @@ object AlterTableUtil { propKeys, set) LOGGER.info(s"Alter table newProperties is successful for table $dbName.$tableName") - LOGGER.audit(s"Alter table newProperties is successful for table $dbName.$tableName") + Audit.log(LOGGER, s"Alter table newProperties is successful for table $dbName.$tableName") } catch { case e: Exception => - LOGGER.error(e, "Alter table newProperties failed") + LOGGER.error("Alter table newProperties failed", e) sys.error(s"Alter table newProperties operation failed: ${e.getMessage}") } finally { // release lock after command execution completion http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/datamap/DataMapWriterListener.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/datamap/DataMapWriterListener.java b/processing/src/main/java/org/apache/carbondata/processing/datamap/DataMapWriterListener.java index 55a251d..e88c422 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/datamap/DataMapWriterListener.java +++ b/processing/src/main/java/org/apache/carbondata/processing/datamap/DataMapWriterListener.java @@ -24,7 +24,6 @@ import java.util.Map; import java.util.Set; import java.util.concurrent.ConcurrentHashMap; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datamap.DataMapMeta; import org.apache.carbondata.core.datamap.DataMapStoreManager; @@ -39,12 +38,14 @@ import org.apache.carbondata.core.metadata.schema.table.CarbonTable; import org.apache.carbondata.core.metadata.schema.table.column.CarbonColumn; import org.apache.carbondata.processing.store.TablePage; +import org.apache.log4j.Logger; + /** * It is for writing DataMap for one table */ public class DataMapWriterListener { - private static final LogService LOG = LogServiceFactory.getLogService( + private static final Logger LOG = LogServiceFactory.getLogService( DataMapWriterListener.class.getCanonicalName()); // list indexed column -> list of data map writer @@ -68,7 +69,7 @@ public class DataMapWriterListener { try { tableIndices = DataMapStoreManager.getInstance().getAllDataMap(carbonTable); } catch (IOException e) { - LOG.error(e, "Error while retrieving datamaps"); + LOG.error("Error while retrieving datamaps", e); throw new RuntimeException(e); } if (tableIndices != null) { http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java b/processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java index 4e2b871..c12ed1c 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java +++ b/processing/src/main/java/org/apache/carbondata/processing/datatypes/PrimitiveDataType.java @@ -27,7 +27,6 @@ import java.util.ArrayList; import java.util.List; import java.util.Map; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.cache.Cache; import org.apache.carbondata.core.cache.CacheProvider; @@ -61,12 +60,14 @@ import org.apache.carbondata.processing.loading.dictionary.PreCreatedDictionary; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * Primitive DataType stateless object used in data loading */ public class PrimitiveDataType implements GenericDataType<Object> { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(PrimitiveDataType.class.getName()); /** http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/AbstractDataLoadProcessorStep.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/AbstractDataLoadProcessorStep.java b/processing/src/main/java/org/apache/carbondata/processing/loading/AbstractDataLoadProcessorStep.java index 69f79f8..b93f21d 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/AbstractDataLoadProcessorStep.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/AbstractDataLoadProcessorStep.java @@ -21,12 +21,13 @@ import java.io.IOException; import java.util.Iterator; import java.util.concurrent.atomic.AtomicLong; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.processing.datamap.DataMapWriterListener; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; import org.apache.carbondata.processing.loading.row.CarbonRowBatch; +import org.apache.log4j.Logger; + /** * This base abstract class for data loading. * It can do transformation jobs as per the implementation. @@ -38,7 +39,7 @@ import org.apache.carbondata.processing.loading.row.CarbonRowBatch; */ public abstract class AbstractDataLoadProcessorStep { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(AbstractDataLoadProcessorStep.class.getName()); protected CarbonDataLoadConfiguration configuration; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/BadRecordsLogger.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/BadRecordsLogger.java b/processing/src/main/java/org/apache/carbondata/processing/loading/BadRecordsLogger.java index 58ec0d5..e0dcd26 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/BadRecordsLogger.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/BadRecordsLogger.java @@ -27,7 +27,6 @@ import java.nio.charset.Charset; import java.util.Map; import java.util.concurrent.ConcurrentHashMap; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.impl.FileFactory; @@ -35,12 +34,14 @@ import org.apache.carbondata.core.datastore.impl.FileFactory.FileType; import org.apache.carbondata.core.util.CarbonUtil; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; +import org.apache.log4j.Logger; + public class BadRecordsLogger { /** * Comment for <code>LOGGER</code> */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(BadRecordsLogger.class.getName()); /** * Which holds the key and if any bad rec found to check from API to update http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadExecutor.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadExecutor.java b/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadExecutor.java index fc5c41f..6550afe 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadExecutor.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadExecutor.java @@ -18,7 +18,6 @@ package org.apache.carbondata.processing.loading; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.metadata.CarbonTableIdentifier; import org.apache.carbondata.processing.loading.exception.BadRecordFoundException; @@ -27,12 +26,14 @@ import org.apache.carbondata.processing.loading.exception.NoRetryException; import org.apache.carbondata.processing.loading.model.CarbonLoadModel; import org.apache.carbondata.processing.util.CarbonBadRecordUtil; +import org.apache.log4j.Logger; + /** * It executes the data load. */ public class DataLoadExecutor { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(DataLoadExecutor.class.getName()); private AbstractDataLoadProcessorStep loadProcessorStep; @@ -62,7 +63,7 @@ public class DataLoadExecutor { throw e; } } catch (Exception e) { - LOGGER.error(e, "Data Loading failed for table " + loadModel.getTableName()); + LOGGER.error("Data Loading failed for table " + loadModel.getTableName(), e); throw new CarbonDataLoadingException( "Data Loading failed for table " + loadModel.getTableName(), e); } http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadProcessBuilder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadProcessBuilder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadProcessBuilder.java index f89bc2f..f89a4e7 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadProcessBuilder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/DataLoadProcessBuilder.java @@ -22,7 +22,6 @@ import java.util.ArrayList; import java.util.List; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.constants.CarbonLoadOptionConstants; @@ -51,12 +50,13 @@ import org.apache.carbondata.processing.loading.steps.SortProcessorStepImpl; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; import org.apache.commons.lang3.StringUtils; +import org.apache.log4j.Logger; /** * It builds the pipe line of steps for loading data to carbon. */ public final class DataLoadProcessBuilder { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(DataLoadProcessBuilder.class.getName()); public AbstractDataLoadProcessorStep build(CarbonLoadModel loadModel, String[] storeLocation, http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/TableProcessingOperations.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/TableProcessingOperations.java b/processing/src/main/java/org/apache/carbondata/processing/loading/TableProcessingOperations.java index 5bed8b1..89da224 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/TableProcessingOperations.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/TableProcessingOperations.java @@ -23,7 +23,6 @@ import java.util.concurrent.Callable; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.filesystem.CarbonFile; import org.apache.carbondata.core.datastore.filesystem.CarbonFileFilter; @@ -40,9 +39,10 @@ import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; import org.apache.carbondata.processing.util.CarbonLoaderUtil; import org.apache.commons.lang3.StringUtils; +import org.apache.log4j.Logger; public class TableProcessingOperations { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(CarbonLoaderUtil.class.getName()); /** @@ -136,7 +136,7 @@ public class TableProcessingOperations { try { CarbonUtil.deleteFoldersAndFiles(new File(loc)); } catch (IOException | InterruptedException e) { - LOGGER.error(e, "Failed to delete local data load folder location: " + loc); + LOGGER.error("Failed to delete local data load folder location: " + loc, e); } } LOGGER.info( http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/MeasureFieldConverterImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/MeasureFieldConverterImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/MeasureFieldConverterImpl.java index 212037b..41d171b 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/MeasureFieldConverterImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/MeasureFieldConverterImpl.java @@ -16,7 +16,6 @@ */ package org.apache.carbondata.processing.loading.converter.impl; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.row.CarbonRow; @@ -28,12 +27,14 @@ import org.apache.carbondata.processing.loading.converter.FieldConverter; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * Converter for measure */ public class MeasureFieldConverterImpl implements FieldConverter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(MeasureFieldConverterImpl.class.getName()); private int index; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/RowConverterImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/RowConverterImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/RowConverterImpl.java index 2d4e167..458b3ab 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/RowConverterImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/converter/impl/RowConverterImpl.java @@ -27,7 +27,6 @@ import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.row.CarbonRow; import org.apache.carbondata.core.dictionary.client.DictionaryClient; @@ -44,13 +43,15 @@ import org.apache.carbondata.processing.loading.converter.RowConverter; import org.apache.carbondata.processing.loading.exception.BadRecordFoundException; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; +import org.apache.log4j.Logger; + /** * It converts the complete row if necessary, dictionary columns are encoded with dictionary values * and nondictionary values are converted to binary. */ public class RowConverterImpl implements RowConverter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(RowConverterImpl.class.getName()); private CarbonDataLoadConfiguration configuration; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVInputFormat.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVInputFormat.java b/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVInputFormat.java index 86c71a6..f01aea8 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVInputFormat.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVInputFormat.java @@ -22,7 +22,6 @@ import java.io.InputStreamReader; import java.io.Reader; import java.nio.charset.Charset; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.util.CarbonProperties; @@ -52,6 +51,7 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hadoop.mapreduce.lib.input.FileInputFormat; import org.apache.hadoop.mapreduce.lib.input.FileSplit; import org.apache.hadoop.util.LineReader; +import org.apache.log4j.Logger; /** * An {@link org.apache.hadoop.mapreduce.InputFormat} for csv files. Files are broken into lines. @@ -77,7 +77,7 @@ public class CSVInputFormat extends FileInputFormat<NullWritable, StringArrayWri public static final int DEFAULT_MAX_NUMBER_OF_COLUMNS_FOR_PARSING = 2000; public static final int THRESHOLD_MAX_NUMBER_OF_COLUMNS_FOR_PARSING = 20000; - private static LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(CSVInputFormat.class.toString()); http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVRecordReaderIterator.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVRecordReaderIterator.java b/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVRecordReaderIterator.java index 24ef9c1..d963502 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVRecordReaderIterator.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/csvinput/CSVRecordReaderIterator.java @@ -19,13 +19,11 @@ package org.apache.carbondata.processing.loading.csvinput; import java.io.IOException; - import org.apache.carbondata.common.CarbonIterator; import org.apache.carbondata.processing.loading.exception.CarbonDataLoadingException; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; import com.univocity.parsers.common.TextParsingException; - import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.mapreduce.InputSplit; import org.apache.hadoop.mapreduce.RecordReader; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/model/CarbonLoadModelBuilder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/model/CarbonLoadModelBuilder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/model/CarbonLoadModelBuilder.java index ddd54a4..7fecb12 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/model/CarbonLoadModelBuilder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/model/CarbonLoadModelBuilder.java @@ -29,7 +29,6 @@ import org.apache.carbondata.common.Strings; import org.apache.carbondata.common.annotations.InterfaceAudience; import org.apache.carbondata.common.constants.LoggerAction; import org.apache.carbondata.common.exceptions.sql.InvalidLoadOptionException; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.compression.CompressorFactory; @@ -45,13 +44,14 @@ import org.apache.carbondata.processing.util.TableOptionConstant; import org.apache.commons.lang.StringUtils; import org.apache.hadoop.conf.Configuration; +import org.apache.log4j.Logger; /** * Builder for {@link CarbonLoadModel} */ @InterfaceAudience.Internal public class CarbonLoadModelBuilder { - private static final LogService LOGGER = LogServiceFactory.getLogService( + private static final Logger LOGGER = LogServiceFactory.getLogService( CarbonLoadModelBuilder.class.getName()); private CarbonTable table; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/model/LoadOption.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/model/LoadOption.java b/processing/src/main/java/org/apache/carbondata/processing/loading/model/LoadOption.java index 98cd90d..a1dee27 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/model/LoadOption.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/model/LoadOption.java @@ -27,7 +27,6 @@ import java.util.Map; import org.apache.carbondata.common.Maps; import org.apache.carbondata.common.annotations.InterfaceAudience; import org.apache.carbondata.common.exceptions.sql.InvalidLoadOptionException; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.constants.CarbonLoadOptionConstants; @@ -39,6 +38,7 @@ import org.apache.carbondata.processing.util.CarbonLoaderUtil; import org.apache.commons.lang.StringUtils; import org.apache.hadoop.conf.Configuration; +import org.apache.log4j.Logger; /** * Provide utilities to populate loading options @@ -46,7 +46,7 @@ import org.apache.hadoop.conf.Configuration; @InterfaceAudience.Internal public class LoadOption { - private static LogService LOG = LogServiceFactory.getLogService(LoadOption.class.getName()); + private static final Logger LOG = LogServiceFactory.getLogService(LoadOption.class.getName()); /** * Based on the input options, fill and return data loading options with default value http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/partition/impl/RangePartitionerImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/partition/impl/RangePartitionerImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/partition/impl/RangePartitionerImpl.java index d59ad02..fd4e037 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/partition/impl/RangePartitionerImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/partition/impl/RangePartitionerImpl.java @@ -21,14 +21,15 @@ import java.util.Arrays; import java.util.Comparator; import org.apache.carbondata.common.annotations.InterfaceAudience; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.row.CarbonRow; import org.apache.carbondata.processing.loading.partition.Partitioner; +import org.apache.log4j.Logger; + @InterfaceAudience.Internal public class RangePartitionerImpl implements Partitioner<CarbonRow> { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(RangePartitionerImpl.class.getName()); private CarbonRow[] rangeBounds; private Comparator<CarbonRow> comparator; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/SorterFactory.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/SorterFactory.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/SorterFactory.java index b74b393..aad9083 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/SorterFactory.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/SorterFactory.java @@ -19,7 +19,6 @@ package org.apache.carbondata.processing.loading.sort; import java.util.concurrent.atomic.AtomicLong; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.util.CarbonProperties; @@ -31,9 +30,11 @@ import org.apache.carbondata.processing.loading.sort.impl.UnsafeParallelReadMerg import org.apache.carbondata.processing.loading.sort.impl.UnsafeParallelReadMergeSorterWithColumnRangeImpl; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + public class SorterFactory { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(SorterFactory.class.getName()); public static Sorter createSorter(CarbonDataLoadConfiguration configuration, AtomicLong counter) { http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterImpl.java index 74e1594..f0920ee 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterImpl.java @@ -24,7 +24,6 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicLong; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; @@ -42,6 +41,8 @@ import org.apache.carbondata.processing.sort.sortdata.SortIntermediateFileMerger import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * It parallely reads data from array of iterates and do merge sort. * First it sorts the data and write to temp files. These temp files will be merge sorted to get @@ -49,7 +50,7 @@ import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; */ public class ParallelReadMergeSorterImpl extends AbstractMergeSorter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(ParallelReadMergeSorterImpl.class.getName()); private SortParameters sortParameters; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterWithColumnRangeImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterWithColumnRangeImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterWithColumnRangeImpl.java index 5419e05..3b767aa 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterWithColumnRangeImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/ParallelReadMergeSorterWithColumnRangeImpl.java @@ -26,7 +26,6 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicLong; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; @@ -44,6 +43,8 @@ import org.apache.carbondata.processing.sort.sortdata.SortIntermediateFileMerger import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * It parallely reads data from array of iterates and do merge sort. * First it sorts the data and write to temp files. These temp files will be merge sorted to get @@ -52,7 +53,7 @@ import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; * bucketing,sort_column_bounds, it sorts each range of data separately and write to temp files. */ public class ParallelReadMergeSorterWithColumnRangeImpl extends AbstractMergeSorter { - private static final LogService LOGGER = LogServiceFactory.getLogService( + private static final Logger LOGGER = LogServiceFactory.getLogService( ParallelReadMergeSorterWithColumnRangeImpl.class.getName()); private SortParameters originSortParameters; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeBatchParallelReadMergeSorterImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeBatchParallelReadMergeSorterImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeBatchParallelReadMergeSorterImpl.java index 1b1d383..9cb67df 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeBatchParallelReadMergeSorterImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeBatchParallelReadMergeSorterImpl.java @@ -28,7 +28,6 @@ import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicLong; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; @@ -47,13 +46,15 @@ import org.apache.carbondata.processing.sort.exception.CarbonSortKeyAndGroupByEx import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * It parallely reads data from array of iterates and do merge sort. * It sorts data in batches and send to the next step. */ public class UnsafeBatchParallelReadMergeSorterImpl extends AbstractMergeSorter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeBatchParallelReadMergeSorterImpl.class.getName()); private SortParameters sortParameters; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterImpl.java index afa30c0..6e11ca6 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterImpl.java @@ -24,7 +24,6 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicLong; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; import org.apache.carbondata.core.datastore.row.CarbonRow; @@ -41,6 +40,8 @@ import org.apache.carbondata.processing.loading.sort.unsafe.merger.UnsafeSingleT import org.apache.carbondata.processing.sort.exception.CarbonSortKeyAndGroupByException; import org.apache.carbondata.processing.sort.sortdata.SortParameters; +import org.apache.log4j.Logger; + /** * It parallely reads data from array of iterates and do merge sort. * First it sorts the data and write to temp files. These temp files will be merge sorted to get @@ -48,7 +49,7 @@ import org.apache.carbondata.processing.sort.sortdata.SortParameters; */ public class UnsafeParallelReadMergeSorterImpl extends AbstractMergeSorter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeParallelReadMergeSorterImpl.class.getName()); private SortParameters sortParameters; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterWithColumnRangeImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterWithColumnRangeImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterWithColumnRangeImpl.java index 5766105..a8ec05c 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterWithColumnRangeImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/impl/UnsafeParallelReadMergeSorterWithColumnRangeImpl.java @@ -27,7 +27,6 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicLong; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.row.CarbonRow; @@ -45,6 +44,7 @@ import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; import org.apache.commons.lang3.StringUtils; +import org.apache.log4j.Logger; /** * It parallely reads data from array of iterates and do merge sort. @@ -55,7 +55,7 @@ import org.apache.commons.lang3.StringUtils; */ public class UnsafeParallelReadMergeSorterWithColumnRangeImpl extends AbstractMergeSorter { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService( UnsafeParallelReadMergeSorterWithColumnRangeImpl.class.getName()); http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/UnsafeSortDataRows.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/UnsafeSortDataRows.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/UnsafeSortDataRows.java index 9c23dde..e199c89 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/UnsafeSortDataRows.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/UnsafeSortDataRows.java @@ -27,7 +27,6 @@ import java.util.concurrent.Executors; import java.util.concurrent.Semaphore; import java.util.concurrent.TimeUnit; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.impl.FileFactory; @@ -53,11 +52,13 @@ import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.sort.sortdata.TableFieldStat; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + public class UnsafeSortDataRows { /** * LOGGER */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeSortDataRows.class.getName()); /** * threadStatusObserver http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java index b805d37..896af60 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeFinalMergePageHolder.java @@ -17,7 +17,6 @@ package org.apache.carbondata.processing.loading.sort.unsafe.holder; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.metadata.datatype.DataType; import org.apache.carbondata.processing.loading.row.IntermediateSortTempRow; @@ -25,9 +24,11 @@ import org.apache.carbondata.processing.loading.sort.unsafe.UnsafeCarbonRowPage; import org.apache.carbondata.processing.loading.sort.unsafe.merger.UnsafeInMemoryIntermediateDataMerger; import org.apache.carbondata.processing.sort.sortdata.IntermediateSortTempRowComparator; +import org.apache.log4j.Logger; + public class UnsafeFinalMergePageHolder implements SortTempChunkHolder { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeFinalMergePageHolder.class.getName()); private int counter; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java index baa9e71..e5680de 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryHolder.java @@ -17,15 +17,16 @@ package org.apache.carbondata.processing.loading.sort.unsafe.holder; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.processing.loading.row.IntermediateSortTempRow; import org.apache.carbondata.processing.loading.sort.unsafe.UnsafeCarbonRowPage; import org.apache.carbondata.processing.sort.sortdata.IntermediateSortTempRowComparator; +import org.apache.log4j.Logger; + public class UnsafeInmemoryHolder implements SortTempChunkHolder { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeInmemoryHolder.class.getName()); private int counter; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryMergeHolder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryMergeHolder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryMergeHolder.java index f8689d9..b47a5d5 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryMergeHolder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeInmemoryMergeHolder.java @@ -16,17 +16,18 @@ */ package org.apache.carbondata.processing.loading.sort.unsafe.holder; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.processing.loading.sort.unsafe.UnsafeCarbonRowPage; import org.apache.carbondata.processing.loading.sort.unsafe.comparator.UnsafeRowComparator; +import org.apache.log4j.Logger; + /** * It is used for merging unsafe inmemory intermediate data */ public class UnsafeInmemoryMergeHolder implements Comparable<UnsafeInmemoryMergeHolder> { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeInmemoryMergeHolder.class.getName()); private int counter; http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeSortTempFileChunkHolder.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeSortTempFileChunkHolder.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeSortTempFileChunkHolder.java index a991d4c..4a97b20 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeSortTempFileChunkHolder.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/holder/UnsafeSortTempFileChunkHolder.java @@ -27,7 +27,6 @@ import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.impl.FileFactory; @@ -40,12 +39,14 @@ import org.apache.carbondata.processing.sort.sortdata.IntermediateSortTempRowCom import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.sort.sortdata.TableFieldStat; +import org.apache.log4j.Logger; + public class UnsafeSortTempFileChunkHolder implements SortTempChunkHolder { /** * LOGGER */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeSortTempFileChunkHolder.class.getName()); /** http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeInMemoryIntermediateDataMerger.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeInMemoryIntermediateDataMerger.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeInMemoryIntermediateDataMerger.java index a65de16..8a4503b 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeInMemoryIntermediateDataMerger.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeInMemoryIntermediateDataMerger.java @@ -26,7 +26,6 @@ import java.util.PriorityQueue; import java.util.Random; import java.util.concurrent.Callable; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.datastore.impl.FileFactory; @@ -39,11 +38,13 @@ import org.apache.carbondata.processing.loading.sort.unsafe.holder.UnsafeInmemor import org.apache.carbondata.processing.sort.exception.CarbonSortKeyAndGroupByException; import org.apache.carbondata.processing.sort.sortdata.SortParameters; +import org.apache.log4j.Logger; + public class UnsafeInMemoryIntermediateDataMerger implements Callable<Void> { /** * LOGGER */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeInMemoryIntermediateDataMerger.class.getName()); /** @@ -117,7 +118,7 @@ public class UnsafeInMemoryIntermediateDataMerger implements Callable<Void> { + ", containing rows: " + totalSize); } } catch (Exception e) { - LOGGER.error(e, "Problem while intermediate merging"); + LOGGER.error("Problem while intermediate merging", e); throw e; } finally { if (spillDisk) { http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateFileMerger.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateFileMerger.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateFileMerger.java index 0a12eda..041544b 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateFileMerger.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateFileMerger.java @@ -26,7 +26,6 @@ import java.util.NoSuchElementException; import java.util.PriorityQueue; import java.util.concurrent.Callable; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.impl.FileFactory; import org.apache.carbondata.core.util.CarbonUtil; @@ -38,11 +37,13 @@ import org.apache.carbondata.processing.sort.exception.CarbonSortKeyAndGroupByEx import org.apache.carbondata.processing.sort.sortdata.SortParameters; import org.apache.carbondata.processing.sort.sortdata.TableFieldStat; +import org.apache.log4j.Logger; + public class UnsafeIntermediateFileMerger implements Callable<Void> { /** * LOGGER */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeIntermediateFileMerger.class.getName()); /** @@ -105,7 +106,7 @@ public class UnsafeIntermediateFileMerger implements Callable<Void> { LOGGER.info("============================== Intermediate Merge of " + fileConterConst + " Sort Temp Files Cost Time: " + intermediateMergeCostTime + "(s)"); } catch (Exception e) { - LOGGER.error(e, "Problem while intermediate merging"); + LOGGER.error("Problem while intermediate merging", e); clear(); throwable = e; } finally { @@ -114,7 +115,7 @@ public class UnsafeIntermediateFileMerger implements Callable<Void> { try { finish(); } catch (CarbonSortKeyAndGroupByException e) { - LOGGER.error(e, "Problem while deleting the merge file"); + LOGGER.error("Problem while deleting the merge file", e); throwable = e; } } else { http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateMerger.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateMerger.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateMerger.java index 18f35d3..f972f0c 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateMerger.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeIntermediateMerger.java @@ -27,7 +27,6 @@ import java.util.concurrent.Executors; import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.constants.CarbonCommonConstants; import org.apache.carbondata.core.memory.UnsafeSortMemoryManager; @@ -37,12 +36,14 @@ import org.apache.carbondata.processing.loading.sort.unsafe.UnsafeCarbonRowPage; import org.apache.carbondata.processing.sort.exception.CarbonSortKeyAndGroupByException; import org.apache.carbondata.processing.sort.sortdata.SortParameters; +import org.apache.log4j.Logger; + /** * It does mergesort intermediate files to big file. */ public class UnsafeIntermediateMerger { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeIntermediateMerger.class.getName()); /** @@ -211,7 +212,7 @@ public class UnsafeIntermediateMerger { try { mergerTask.get(i).get(); } catch (InterruptedException | ExecutionException e) { - LOGGER.error(e, e.getMessage()); + LOGGER.error(e.getMessage(), e); throw new CarbonSortKeyAndGroupByException(e.getMessage(), e); } } http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeSingleThreadFinalSortFilesMerger.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeSingleThreadFinalSortFilesMerger.java b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeSingleThreadFinalSortFilesMerger.java index 2dd2f31..7e36389 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeSingleThreadFinalSortFilesMerger.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/sort/unsafe/merger/UnsafeSingleThreadFinalSortFilesMerger.java @@ -28,7 +28,6 @@ import java.util.NoSuchElementException; import java.util.PriorityQueue; import org.apache.carbondata.common.CarbonIterator; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; import org.apache.carbondata.processing.loading.row.IntermediateSortTempRow; @@ -40,11 +39,13 @@ import org.apache.carbondata.processing.loading.sort.unsafe.holder.UnsafeInmemor import org.apache.carbondata.processing.loading.sort.unsafe.holder.UnsafeSortTempFileChunkHolder; import org.apache.carbondata.processing.sort.sortdata.SortParameters; +import org.apache.log4j.Logger; + public class UnsafeSingleThreadFinalSortFilesMerger extends CarbonIterator<Object[]> { /** * LOGGER */ - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(UnsafeSingleThreadFinalSortFilesMerger.class.getName()); /** http://git-wip-us.apache.org/repos/asf/carbondata/blob/06adb5a0/processing/src/main/java/org/apache/carbondata/processing/loading/steps/CarbonRowDataWriterProcessorStepImpl.java ---------------------------------------------------------------------- diff --git a/processing/src/main/java/org/apache/carbondata/processing/loading/steps/CarbonRowDataWriterProcessorStepImpl.java b/processing/src/main/java/org/apache/carbondata/processing/loading/steps/CarbonRowDataWriterProcessorStepImpl.java index 7683bbc..ae42df7 100644 --- a/processing/src/main/java/org/apache/carbondata/processing/loading/steps/CarbonRowDataWriterProcessorStepImpl.java +++ b/processing/src/main/java/org/apache/carbondata/processing/loading/steps/CarbonRowDataWriterProcessorStepImpl.java @@ -25,7 +25,6 @@ import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.Future; -import org.apache.carbondata.common.logging.LogService; import org.apache.carbondata.common.logging.LogServiceFactory; import org.apache.carbondata.core.datastore.exception.CarbonDataWriterException; import org.apache.carbondata.core.datastore.row.CarbonRow; @@ -49,13 +48,15 @@ import org.apache.carbondata.processing.store.CarbonFactHandler; import org.apache.carbondata.processing.store.CarbonFactHandlerFactory; import org.apache.carbondata.processing.util.CarbonDataProcessorUtil; +import org.apache.log4j.Logger; + /** * It reads data from sorted files which are generated in previous sort step. * And it writes data to carbondata file. It also generates mdk key while writing to carbondata file */ public class CarbonRowDataWriterProcessorStepImpl extends AbstractDataLoadProcessorStep { - private static final LogService LOGGER = + private static final Logger LOGGER = LogServiceFactory.getLogService(CarbonRowDataWriterProcessorStepImpl.class.getName()); private int dimensionWithComplexCount; @@ -146,11 +147,11 @@ public class CarbonRowDataWriterProcessorStepImpl extends AbstractDataLoadProces } } } catch (CarbonDataWriterException e) { - LOGGER.error(e, "Failed for table: " + tableName + " in DataWriterProcessorStepImpl"); + LOGGER.error("Failed for table: " + tableName + " in DataWriterProcessorStepImpl", e); throw new CarbonDataLoadingException( "Error while initializing data handler : " + e.getMessage()); } catch (Exception e) { - LOGGER.error(e, "Failed for table: " + tableName + " in DataWriterProcessorStepImpl"); + LOGGER.error("Failed for table: " + tableName + " in DataWriterProcessorStepImpl", e); if (e instanceof BadRecordFoundException) { throw new BadRecordFoundException(e.getMessage(), e); } @@ -198,7 +199,7 @@ public class CarbonRowDataWriterProcessorStepImpl extends AbstractDataLoadProces } catch (Exception e) { // if throw exception from here dataHandler will not be closed. // so just holding exception and later throwing exception - LOGGER.error(e, "Failed for table: " + tableName + " in finishing data handler"); + LOGGER.error("Failed for table: " + tableName + " in finishing data handler", e); exception = new CarbonDataWriterException( "Failed for table: " + tableName + " in finishing data handler", e); } @@ -229,10 +230,10 @@ public class CarbonRowDataWriterProcessorStepImpl extends AbstractDataLoadProces try { dataHandler.closeHandler(); } catch (CarbonDataWriterException e) { - LOGGER.error(e, e.getMessage()); + LOGGER.error(e.getMessage(), e); throw new CarbonDataLoadingException(e.getMessage()); } catch (Exception e) { - LOGGER.error(e, e.getMessage()); + LOGGER.error(e.getMessage(), e); throw new CarbonDataLoadingException("There is an unexpected error: " + e.getMessage()); } }
