http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupHandler.java deleted file mode 100644 index f764f18..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupHandler.java +++ /dev/null @@ -1,744 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.backup; - -import java.io.IOException; -import java.util.ArrayList; -import java.util.HashMap; -import java.util.List; -import java.util.concurrent.Callable; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileStatus; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.backup.BackupManifest.BackupImage; -import org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.Connection; -import org.apache.hadoop.hbase.client.ConnectionFactory; -import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; -import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; -import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -import org.apache.hadoop.hbase.util.FSUtils; -import org.apache.zookeeper.KeeperException.NoNodeException; - -/** - * A Handler to carry the operations of backup progress - */ [email protected] [email protected] -public class BackupHandler implements Callable<Object> { - private static final Log LOG = LogFactory.getLog(BackupHandler.class); - - // backup phase - // for overall backup (for table list, some table may go online, while some may go offline) - protected static enum BACKUPPHASE { - REQUEST, SNAPSHOT, PREPARE_INCREMENTAL, SNAPSHOTCOPY, INCREMENTAL_COPY, STORE_MANIFEST; - } - - // backup status flag - protected static enum BACKUPSTATUS { - WAITING, ONGOING, COMPLETE, FAILED, CANCELLED; - } - - protected BackupContext backupContext; - private BackupManager backupManager; - private Configuration conf; - - public BackupHandler(BackupContext backupContext, - BackupManager backupManager, Configuration conf) { - this.backupContext = backupContext; - this.backupManager = backupManager; - this.conf = conf; - } - - public BackupContext getBackupContext() { - return backupContext; - } - - @Override - public Object call() throws Exception { - try { - // overall backup begin - this.beginBackup(backupContext); - HashMap<String, String> newTimestamps = null; - boolean fromExistingSnapshot = false; // supported by future jira - // handle full or incremental backup for table or table list - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - String savedStartCode = null; - HBaseAdmin hbadmin = null; - Connection conn = null; - boolean firstBackup = false; - // do snapshot for full table backup, if backing up from existing snapshot, then skip the - // step of taking snapshot - if (fromExistingSnapshot) { - LOG.error("Backup from existing snapshot, so skip the snapshot step. "); - LOG.error("This feature will be supported by a future jira"); - throw new RuntimeException("Backup from existing snapshot is not supported"); - } else { - try { - savedStartCode = backupManager.readBackupStartCode(); - firstBackup = savedStartCode == null; - if (firstBackup) { - // This is our first backup. Let's put some marker on ZK so that we can hold the logs - // while we do the backup. - backupManager.writeBackupStartCode("0"); - } - // We roll log here before we do the snapshot. It is possible there is duplicate data - // in the log that is already in the snapshot. But if we do it after the snapshot, we - // could have data loss. - // A better approach is to do the roll log on each RS in the same global procedure as - // the snapshot. - LOG.info("Execute roll log procedure for full backup ..."); - conn = ConnectionFactory.createConnection(conf); - hbadmin = (HBaseAdmin) conn.getAdmin(); - hbadmin.execProcedure(LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_SIGNATURE, - LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_NAME, new HashMap<String, String>()); - newTimestamps = backupManager.readRegionServerLastLogRollResult(); - if (firstBackup) { - // Updates registered log files - // We record ALL old WAL files as registered, because - // this is a first full backup in the system and these - // files are not needed for next incremental backup - List<String> logFiles = BackupUtil.getWALFilesOlderThan(conf, newTimestamps); - backupManager.recordWALFiles(logFiles); - } - this.snapshotForFullBackup(backupContext); - } catch (BackupException e) { - // fail the overall backup and return - this.failBackup(backupContext, e, "Unexpected BackupException : "); - return null; - } finally { - if (hbadmin != null) { - hbadmin.close(); - } - if (conn != null) { - conn.close(); - } - } - } - - // update the faked progress currently for snapshot done - this.updateProgress("10.0%", 0); - // do snapshot copy - try { - this.snapshotCopy(backupContext); - } catch (Exception e) { - // fail the overall backup and return - this.failBackup(backupContext, e, "Unexpected BackupException : "); - return null; - } - // Updates incremental backup table set - backupManager.addIncrementalBackupTableSet(backupContext.getTables()); - - } else if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - LOG.info("For incremental backup, current table set is " - + backupManager.getIncrementalBackupTableSet()); - // do incremental table backup preparation - backupContext.setPhase(BACKUPPHASE.PREPARE_INCREMENTAL); - // avoid action if has been cancelled - if (backupContext.isCancelled()) { - return null; - } - try { - IncrementalBackupManager incrBackupManager = new IncrementalBackupManager(backupManager); - - newTimestamps = incrBackupManager.getIncrBackupLogFileList(backupContext); - } catch (Exception e) { - // fail the overall backup and return - this.failBackup(backupContext, e, "Unexpected Exception : "); - return null; - } - // update the faked progress currently for incremental preparation done - this.updateProgress("10.0%", 0); - - // do incremental copy - try { - // copy out the table and region info files for each table - BackupUtil.copyTableRegionInfo(backupContext, conf); - this.incrementalCopy(backupContext); - // Save list of WAL files copied - backupManager.recordWALFiles(backupContext.getIncrBackupFileList()); - } catch (Exception e) { - // fail the overall backup and return - this.failBackup(backupContext, e, "Unexpected exception doing incremental copy : "); - return null; - } - - } else { - LOG.error("Unsupport backup type: " + backupContext.getType()); - } - - // set overall backup status: complete. Here we make sure to complete the backup. After this - // checkpoint, even if entering cancel process, will let the backup finished - backupContext.setFlag(BACKUPSTATUS.COMPLETE); - - if (!fromExistingSnapshot) { - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - // Set the previousTimestampMap which is before this current log roll to the manifest. - HashMap<String, HashMap<String, String>> previousTimestampMap = - backupManager.readLogTimestampMap(); - backupContext.setIncrTimestampMap(previousTimestampMap); - } - // The table list in backupContext is good for both full backup and incremental backup. - // For incremental backup, it contains the incremental backup table set. - - backupManager.writeRegionServerLogTimestamp(backupContext.getTables(), newTimestamps); - - HashMap<String, HashMap<String, String>> newTableSetTimestampMap = - backupManager.readLogTimestampMap(); - - String newStartCode = - BackupUtil.getMinValue(BackupUtil.getRSLogTimestampMins(newTableSetTimestampMap)); - backupManager.writeBackupStartCode(newStartCode); - } - - // backup complete - this.completeBackup(backupContext); - } catch (Exception e) { - // even during completing backup (#completeBackup(backupContext)), exception may occur, or - // exception occur during other process, fail the backup finally - this.failBackup(backupContext, e, "Error caught during backup progress: "); - } - return null; - } - - /** - * Begin the overall backup. - * @param backupContext backup context - * @throws IOException exception - */ - private void beginBackup(BackupContext backupContext) throws IOException { - - // set the start timestamp of the overall backup - long startTs = EnvironmentEdgeManager.currentTime(); - backupContext.setStartTs(startTs); - // set overall backup status: ongoing - backupContext.setFlag(BACKUPSTATUS.ONGOING); - LOG.info("Backup " + backupContext.getBackupId() + " starts at " + startTs + "."); - - backupManager.updateBackupStatus(backupContext); - LOG.debug("Backup session " + backupContext.getBackupId() + " has been started."); - - } - - /** - * Snapshot for full table backup. - * @param backupContext backup context - * @throws IOException exception - */ - private void snapshotForFullBackup(BackupContext backupContext) throws IOException { - - LOG.info("HBase snapshot full backup for " + backupContext.getBackupId()); - - // avoid action if has been cancelled - if (backupContext.isCancelled()) { - return; - } - - HBaseAdmin hbadmin = null; - Connection conn = null; - - // we do HBase snapshot for tables in the table list one by one currently - for (String table : backupContext.getTables()) { - - // avoid action if it has been cancelled - if (backupContext.isCancelled()) { - return; - } - - HBaseProtos.SnapshotDescription backupSnapshot; - try { - // wrap a SnapshotDescription for offline/online snapshot - backupSnapshot = this.wrapSnapshotDescription(table); - - // set the snapshot name in BackupStatus of this table - backupContext.setSnapshotName(table, backupSnapshot.getName()); - - // Kick off snapshot for backup - conn = ConnectionFactory.createConnection(conf); - hbadmin = (HBaseAdmin) conn.getAdmin(); - hbadmin.snapshot(backupSnapshot); - - if (LOG.isDebugEnabled() == false) { - // In DEBUG mode we log message already. - // This is not to duplicate that message. - LOG.info("Snapshot has been launched, waiting to finish ..."); - } - - } catch (Exception e) { - LOG.error("Snapshot failed to create " + getMessage(e)); - - // currently, we fail the overall backup if any table in the list failed, so throw the - // exception out for overall backup failing - throw new BackupException("Backup snapshot failed on table " + table, e); - } finally { - if (hbadmin != null) { - hbadmin.close(); - } - if (conn != null) { - conn.close(); - } - } - - // set the snapshot name in BackupStatus of this table, only after snapshot success. - backupContext.setSnapshotName(table, backupSnapshot.getName()); - - } // for each table in the backup table list - - } - - /** - * Fail the overall backup. - * @param backupContext backup context - * @param e exception - * @throws Exception exception - */ - private void failBackup(BackupContext backupContext, Exception e, String msg) throws Exception { - - LOG.error(msg + getMessage(e)); - - // If this is a cancel exception, then we've already cleaned. - - if (this.backupContext.getFlag().equals(BACKUPSTATUS.CANCELLED)) { - return; - } - - // set the failure timestamp of the overall backup - backupContext.setEndTs(EnvironmentEdgeManager.currentTime()); - - // set failure message - backupContext.setFailedMsg(e.getMessage()); - - // set overall backup status: failed - backupContext.setFlag(BACKUPSTATUS.FAILED); - - // compose the backup failed data - String backupFailedData = - "BackupId=" + backupContext.getBackupId() + ",startts=" + backupContext.getStartTs() - + ",failedts=" + backupContext.getEndTs() + ",failedphase=" + backupContext.getPhase() - + ",failedmessage=" + backupContext.getFailedMsg(); - LOG.error(backupFailedData); - - backupManager.updateBackupStatus(backupContext); - - // if full backup, then delete HBase snapshots if there already have snapshots taken - // and also clean up export snapshot log files if exist - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - if (!backupContext.fromExistingSnapshot()) { - this.deleteSnapshot(backupContext); - } - this.cleanupExportSnapshotLog(); - } /* - * else { // support incremental backup code in future jira // TODO. See HBASE-14124 } - */ - - // clean up the uncompleted data at target directory if the ongoing backup has already entered - // the copy phase - // For incremental backup, DistCp logs will be cleaned with the targetDir. - this.cleanupTargetDir(); - - LOG.info("Backup " + backupContext.getBackupId() + " failed."); - } - - /** - * Update the ongoing back token znode with new progress. - * @param newProgress progress - * @param bytesCopied bytes copied - * @throws NoNodeException exception - */ - public void updateProgress(String newProgress, long bytesCopied) throws IOException { - - // compose the new backup progress data, using fake number for now - String backupProgressData = newProgress; - - backupContext.setProgress(newProgress); - backupManager.updateBackupStatus(backupContext); - LOG.debug("Backup progress data \"" + backupProgressData - + "\" has been updated to hbase:backup for " + backupContext.getBackupId()); - } - - /** - * Complete the overall backup. - * @param backupContext backup context - * @throws Exception exception - */ - private void completeBackup(BackupContext backupContext) throws Exception { - - // set the complete timestamp of the overall backup - backupContext.setEndTs(EnvironmentEdgeManager.currentTime()); - // set overall backup status: complete - backupContext.setFlag(BACKUPSTATUS.COMPLETE); - // add and store the manifest for the backup - this.addManifest(backupContext); - - // after major steps done and manifest persisted, do convert if needed for incremental backup - /* in-fly convert code here, provided by future jira */ - LOG.debug("in-fly convert code here, provided by future jira"); - - // compose the backup complete data - String backupCompleteData = - this.obtainBackupMetaDataStr(backupContext) + ",startts=" + backupContext.getStartTs() - + ",completets=" + backupContext.getEndTs() + ",bytescopied=" - + backupContext.getTotalBytesCopied(); - if (LOG.isDebugEnabled()) { - LOG.debug("Backup " + backupContext.getBackupId() + " finished: " + backupCompleteData); - } - backupManager.updateBackupStatus(backupContext); - - // when full backup is done: - // - delete HBase snapshot - // - clean up directories with prefix "exportSnapshot-", which are generated when exporting - // snapshots - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - if (!backupContext.fromExistingSnapshot()) { - this.deleteSnapshot(backupContext); - } - this.cleanupExportSnapshotLog(); - } else if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - this.cleanupDistCpLog(); - } else { - LOG.error(" other backup types have not been implemented yet"); - } - - LOG.info("Backup " + backupContext.getBackupId() + " completed."); - } - - /** - * Get backup request meta data dir as string. - * @param backupContext backup context - * @return meta data dir - */ - private String obtainBackupMetaDataStr(BackupContext backupContext) { - StringBuffer sb = new StringBuffer(); - sb.append("type=" + backupContext.getType() + ",tablelist="); - for (String table : backupContext.getTables()) { - sb.append(table + ";"); - } - if (sb.lastIndexOf(";") > 0) { - sb.delete(sb.lastIndexOf(";"), sb.lastIndexOf(";") + 1); - } - sb.append(",targetRootDir=" + backupContext.getTargetRootDir()); - if (backupContext.fromExistingSnapshot()) { - sb.append(",snapshot=" + backupContext.getExistingSnapshot()); - } - return sb.toString(); - } - - /** - * Do snapshot copy. - * @param backupContext backup context - * @throws Exception exception - */ - private void snapshotCopy(BackupContext backupContext) throws Exception { - - LOG.info("Snapshot copy is starting."); - - // set overall backup phase: snapshot_copy - backupContext.setPhase(BACKUPPHASE.SNAPSHOTCOPY); - - // avoid action if has been cancelled - if (backupContext.isCancelled()) { - return; - } - - // call ExportSnapshot to copy files based on hbase snapshot for backup - // ExportSnapshot only support single snapshot export, need loop for multiple tables case - BackupCopyService copyService = BackupRestoreServiceFactory.getBackupCopyService(conf); - - // number of snapshots matches number of tables - float numOfSnapshots = backupContext.getSnapshotNames().size(); - - LOG.debug("There are " + (int) numOfSnapshots + " snapshots to be copied."); - - for (String table : backupContext.getTables()) { - // Currently we simply set the sub copy tasks by counting the table snapshot number, we can - // calculate the real files' size for the percentage in the future. - // TODO this below - // backupCopier.setSubTaskPercntgInWholeTask(1f / numOfSnapshots); - int res = 0; - String[] args = new String[4]; - args[0] = "-snapshot"; - args[1] = backupContext.getSnapshotName(table); - args[2] = "-copy-to"; - args[3] = backupContext.getBackupStatus(table).getTargetDir(); - - LOG.debug("Copy snapshot " + args[1] + " to " + args[3]); - res = copyService.copy(this, conf, BackupCopyService.Type.FULL, args); - // if one snapshot export failed, do not continue for remained snapshots - if (res != 0) { - - LOG.error("Exporting Snapshot " + args[1] + " failed with return code: " + res + "."); - - throw new IOException("Failed of exporting snapshot " + args[1] + " to " + args[3] - + " with reason code " + res); - } - - LOG.info("Snapshot copy " + args[1] + " finished."); - } - } - - /** - * Wrap a SnapshotDescription for a target table. - * @param table table - * @return a SnapshotDescription especially for backup. - */ - private SnapshotDescription wrapSnapshotDescription(String table) { - // Mock a SnapshotDescription from backupContext to call SnapshotManager function, - // Name it in the format "snapshot_<timestamp>_<table>" - HBaseProtos.SnapshotDescription.Builder builder = HBaseProtos.SnapshotDescription.newBuilder(); - builder.setTable(table); - TableName tableName = TableName.valueOf(table); - builder.setName("snapshot_" + Long.toString(EnvironmentEdgeManager.currentTime()) + "_" - + tableName.getNamespaceAsString() + "_" + tableName.getQualifierAsString()); - HBaseProtos.SnapshotDescription backupSnapshot = builder.build(); - - LOG.debug("Wrapped a SnapshotDescription " + backupSnapshot.getName() - + " from backupContext to request snapshot for backup."); - - return backupSnapshot; - } - - /** - * Delete HBase snapshot for backup. - * @param backupCtx backup context - * @throws Exception exception - */ - private void deleteSnapshot(BackupContext backupCtx) throws IOException { - - LOG.debug("Trying to delete snapshot for full backup."); - Connection conn = null; - Admin admin = null; - try { - conn = ConnectionFactory.createConnection(conf); - admin = conn.getAdmin(); - for (String snapshotName : backupCtx.getSnapshotNames()) { - if (snapshotName == null) { - continue; - } - LOG.debug("Trying to delete snapshot: " + snapshotName); - admin.deleteSnapshot(snapshotName); - LOG.debug("Deleting the snapshot " + snapshotName + " for backup " - + backupCtx.getBackupId() + " succeeded."); - } - } finally { - if (admin != null) { - admin.close(); - } - if (conn != null) { - conn.close(); - } - } - } - - /** - * Clean up directories with prefix "exportSnapshot-", which are generated when exporting - * snapshots. - * @throws IOException exception - */ - private void cleanupExportSnapshotLog() throws IOException { - FileSystem fs = FSUtils.getCurrentFileSystem(conf); - Path stagingDir = - new Path(conf.get(BackupRestoreConstants.CONF_STAGING_ROOT, fs.getWorkingDirectory() - .toString())); - FileStatus[] files = FSUtils.listStatus(fs, stagingDir); - if (files == null) { - return; - } - for (FileStatus file : files) { - if (file.getPath().getName().startsWith("exportSnapshot-")) { - LOG.debug("Delete log files of exporting snapshot: " + file.getPath().getName()); - if (FSUtils.delete(fs, file.getPath(), true) == false) { - LOG.warn("Can not delete " + file.getPath()); - } - } - } - } - - /** - * Clean up directories with prefix "_distcp_logs-", which are generated when DistCp copying - * hlogs. - * @throws IOException exception - */ - private void cleanupDistCpLog() throws IOException { - Path rootPath = new Path(backupContext.getHLogTargetDir()).getParent(); - FileSystem fs = FileSystem.get(rootPath.toUri(), conf); - FileStatus[] files = FSUtils.listStatus(fs, rootPath); - if (files == null) { - return; - } - for (FileStatus file : files) { - if (file.getPath().getName().startsWith("_distcp_logs")) { - LOG.debug("Delete log files of DistCp: " + file.getPath().getName()); - FSUtils.delete(fs, file.getPath(), true); - } - } - } - - /** - * Clean up the uncompleted data at target directory if the ongoing backup has already entered the - * copy phase. - */ - private void cleanupTargetDir() { - try { - // clean up the uncompleted data at target directory if the ongoing backup has already entered - // the copy phase - LOG.debug("Trying to cleanup up target dir. Current backup phase: " - + backupContext.getPhase()); - if (backupContext.getPhase().equals(BACKUPPHASE.SNAPSHOTCOPY) - || backupContext.getPhase().equals(BACKUPPHASE.INCREMENTAL_COPY) - || backupContext.getPhase().equals(BACKUPPHASE.STORE_MANIFEST)) { - FileSystem outputFs = - FileSystem.get(new Path(backupContext.getTargetRootDir()).toUri(), conf); - - // now treat one backup as a transaction, clean up data that has been partially copied at - // table level - for (String table : backupContext.getTables()) { - Path targetDirPath = - new Path(HBackupFileSystem.getTableBackupDir(backupContext.getTargetRootDir(), - backupContext.getBackupId(), table)); - if (outputFs.delete(targetDirPath, true)) { - LOG.info("Cleaning up uncompleted backup data at " + targetDirPath.toString() - + " done."); - } else { - LOG.info("No data has been copied to " + targetDirPath.toString() + "."); - } - - Path tableDir = targetDirPath.getParent(); - FileStatus[] backups = FSUtils.listStatus(outputFs, tableDir); - if (backups == null || backups.length == 0) { - outputFs.delete(tableDir, true); - LOG.debug(tableDir.toString() + " is empty, remove it."); - } - } - } - - } catch (IOException e1) { - LOG.error("Cleaning up uncompleted backup data of " + backupContext.getBackupId() + " at " - + backupContext.getTargetRootDir() + " failed due to " + e1.getMessage() + "."); - } - } - - /** - * Add manifest for the current backup. The manifest is stored - * within the table backup directory. - * @param backupContext The current backup context - * @throws IOException exception - * @throws BackupException exception - */ - private void addManifest(BackupContext backupContext) throws IOException, BackupException { - // set the overall backup phase : store manifest - backupContext.setPhase(BACKUPPHASE.STORE_MANIFEST); - - // avoid action if has been cancelled - if (backupContext.isCancelled()) { - return; - } - - BackupManifest manifest; - boolean fromExistingSnapshot = false; // to be implemented in future jira - - // Since we have each table's backup in its own directory structure, - // we'll store its manifest with the table directory. - for (String table : backupContext.getTables()) { - manifest = new BackupManifest(backupContext, table); - if (fromExistingSnapshot) { - // mark backing up from existing snapshot in manifest, so that later, dependency analysis - // can skip this backup image - LOG.debug("backup using existing snapshot will be supported in future jira"); - } else { - ArrayList<BackupImage> ancestors = this.backupManager.getAncestors(backupContext, table); - for (BackupImage image : ancestors) { - manifest.addDependentImage(image); - } - } - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - // We'll store the log timestamps for this table only in its manifest. - HashMap<String, HashMap<String, String>> tableTimestampMap = - new HashMap<String, HashMap<String, String>>(); - tableTimestampMap.put(table, backupContext.getIncrTimestampMap().get(table)); - manifest.setIncrTimestampMap(tableTimestampMap); - } - manifest.store(conf); - } - - // For incremental backup, we store a overall manifest in - // <backup-root-dir>/WALs/<backup-id> - // This is used when created the next incremental backup - if (backupContext.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - manifest = new BackupManifest(backupContext); - // set the table region server start and end timestamps for incremental backup - manifest.setIncrTimestampMap(backupContext.getIncrTimestampMap()); - ArrayList<BackupImage> ancestors = this.backupManager.getAncestors(backupContext); - for (BackupImage image : ancestors) { - manifest.addDependentImage(image); - } - manifest.store(conf); - } - } - - /** - * Do incremental copy. - * @param backupContext backup context - */ - private void incrementalCopy(BackupContext backupContext) throws Exception { - - LOG.info("Incremental copy is starting."); - - // set overall backup phase: incremental_copy - backupContext.setPhase(BACKUPPHASE.INCREMENTAL_COPY); - - // avoid action if has been cancelled - if (backupContext.isCancelled()) { - return; - } - - // get incremental backup file list and prepare parms for DistCp - List<String> incrBackupFileList = backupContext.getIncrBackupFileList(); - String[] strArr = incrBackupFileList.toArray(new String[incrBackupFileList.size() + 1]); - strArr[strArr.length - 1] = backupContext.getHLogTargetDir(); - - BackupCopyService copyService = BackupRestoreServiceFactory.getBackupCopyService(conf); - int res = copyService.copy(this, conf, BackupCopyService.Type.INCREMENTAL, strArr); - - if (res != 0) { - LOG.error("Copy incremental log files failed with return code: " + res + "."); - throw new IOException("Failed of Hadoop Distributed Copy from " + incrBackupFileList + " to " - + backupContext.getHLogTargetDir()); - } - LOG.info("Incremental copy from " + incrBackupFileList + " to " - + backupContext.getHLogTargetDir() + " finished."); - - } - - private String getMessage(Exception e) { - String msg = e.getMessage(); - if (msg == null || msg.equals("")) { - msg = e.getClass().getName(); - } - return msg; - } -}
http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManager.java deleted file mode 100644 index fdb3c46..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManager.java +++ /dev/null @@ -1,488 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.backup; - -import com.google.common.util.concurrent.ThreadFactoryBuilder; - -import java.io.IOException; -import java.util.ArrayList; -import java.util.HashMap; -import java.util.List; -import java.util.Set; -import java.util.concurrent.CancellationException; -import java.util.concurrent.ExecutionException; -import java.util.concurrent.ExecutorService; -import java.util.concurrent.Future; -import java.util.concurrent.LinkedBlockingQueue; -import java.util.concurrent.ThreadPoolExecutor; -import java.util.concurrent.TimeUnit; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.HTableDescriptor; -import org.apache.hadoop.hbase.backup.BackupHandler.BACKUPSTATUS; -import org.apache.hadoop.hbase.backup.BackupManifest.BackupImage; -import org.apache.hadoop.hbase.backup.BackupUtil.BackupCompleteData; -import org.apache.hadoop.hbase.backup.master.BackupLogCleaner; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.client.Connection; -import org.apache.hadoop.hbase.client.ConnectionFactory; -import org.apache.hadoop.hbase.client.HBaseAdmin; - - - -/** - * Handles backup requests on server-side, creates backup context records in hbase:backup - * to keep track backup. The timestamps kept in hbase:backup table will be used for future - * incremental backup. Creates BackupContext and DispatchRequest. - * - */ [email protected] [email protected] -public class BackupManager { - private static final Log LOG = LogFactory.getLog(BackupManager.class); - - private Configuration conf = null; - private BackupContext backupContext = null; - private ExecutorService pool = null; - - private boolean backupComplete = false; - - private BackupSystemTable systemTable; - - /** - * Backup manager constructor. - * @param conf configuration - * @throws IOException exception - */ - public BackupManager(Configuration conf) throws IOException { - if (!conf.getBoolean(HConstants.BACKUP_ENABLE_KEY, HConstants.BACKUP_ENABLE_DEFAULT)) { - throw new BackupException("HBase backup is not enabled. Check your " + - HConstants.BACKUP_ENABLE_KEY + " setting."); - } - this.conf = conf; - this.systemTable = BackupSystemTable.getTable(conf); - Runtime.getRuntime().addShutdownHook(new ExitHandler()); - } - - /** - * This method modifies the master's configuration in order to inject backup-related features - * @param conf configuration - */ - public static void decorateMasterConfiguration(Configuration conf) { - if (!isBackupEnabled(conf)) { - return; - } - String plugins = conf.get(HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS); - String cleanerClass = BackupLogCleaner.class.getCanonicalName(); - if (!plugins.contains(cleanerClass)) { - conf.set(HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS, plugins + "," + cleanerClass); - } - if (LOG.isDebugEnabled()) { - LOG.debug("Added log cleaner: " + cleanerClass); - } - } - - private static boolean isBackupEnabled(Configuration conf) { - return conf.getBoolean(HConstants.BACKUP_ENABLE_KEY, HConstants.BACKUP_ENABLE_DEFAULT); - } - - private class ExitHandler extends Thread { - public ExitHandler() { - super("Backup Manager Exit Handler"); - } - - public void run() { - - if (backupContext != null && !backupComplete) { - - // program exit and backup is not complete, then mark as cancelled to avoid submitted backup - // handler's taking further action - backupContext.markCancel(); - - LOG.debug("Backup is cancelled due to force program exiting."); - try { - cancelBackup(backupContext.getBackupId()); - } catch (Exception e) { - String msg = e.getMessage(); - if (msg == null || msg.equals("")) { - msg = e.getClass().getName(); - } - LOG.error("Failed to cancel backup " + backupContext.getBackupId() + " due to " + msg); - } - } - - exit(); - } - } - - /** - * Cancel the ongoing backup via backup id. - * @param backupId The id of the ongoing backup to be cancelled - * @throws Exception exception - */ - private void cancelBackup(String backupId) throws Exception { - // TODO: will be implemented in Phase 2: HBASE-14125 - LOG.debug("Try to cancel the backup " + backupId + ". the feature is NOT implemented yet"); - - } - - /** - * Stop all the work of backup. - */ - public void exit() { - - // currently, we shutdown now for all ongoing back handlers, we may need to do something like - // record the failed list somewhere later - if (this.pool != null) { - this.pool.shutdownNow(); - } - - } - - /** - * Create a BackupContext based on input backup request. - * @param backupId backup id - * @param type type - * @param tablelist table list - * @param targetRootDir root dir - * @param snapshot snapshot name - * @return BackupContext context - * @throws BackupException exception - */ - protected BackupContext createBackupContext(String backupId, String type, String tablelist, - String targetRootDir, String snapshot) throws BackupException { - - if (targetRootDir == null) { - throw new BackupException("Wrong backup request parameter: target backup root directory"); - } - - if (type.equals(BackupRestoreConstants.BACKUP_TYPE_FULL) && tablelist == null) { - // If table list is null for full backup, which means backup all tables. Then fill the table - // list with all user tables from meta. It no table available, throw the request exception. - - HTableDescriptor[] htds = null; - try (Connection conn = ConnectionFactory.createConnection(conf); - HBaseAdmin hbadmin = (HBaseAdmin) conn.getAdmin()) { - - htds = hbadmin.listTables(); - } catch (Exception e) { - throw new BackupException(e); - } - - if (htds == null) { - throw new BackupException("No table exists for full backup of all tables."); - } else { - StringBuilder sb = new StringBuilder(); - for (HTableDescriptor hTableDescriptor : htds) { - sb.append(hTableDescriptor.getNameAsString() - + BackupRestoreConstants.TABLENAME_DELIMITER_IN_COMMAND); - } - sb.deleteCharAt(sb.lastIndexOf(BackupRestoreConstants.TABLENAME_DELIMITER_IN_COMMAND)); - tablelist = sb.toString(); - - LOG.info("Full backup all the tables available in the cluster: " + tablelist); - } - } - - // there are one or more tables in the table list - return new BackupContext(backupId, type, - tablelist.split(BackupRestoreConstants.TABLENAME_DELIMITER_IN_COMMAND), targetRootDir, - snapshot); - - } - - /** - * Check if any ongoing backup. Currently, we only reply on checking status in hbase:backup. We - * need to consider to handle the case of orphan records in the future. Otherwise, all the coming - * request will fail. - * @return the ongoing backup id if on going backup exists, otherwise null - * @throws IOException exception - */ - private String getOngoingBackupId() throws IOException { - - ArrayList<BackupContext> sessions = systemTable.getBackupContexts(BACKUPSTATUS.ONGOING); - if (sessions.size() == 0) { - return null; - } - return sessions.get(0).getBackupId(); - } - - /** - * Start the backup manager service. - * @throws IOException exception - */ - public void initialize() throws IOException { - String ongoingBackupId = this.getOngoingBackupId(); - if (ongoingBackupId != null) { - LOG.info("There is a ongoing backup " + ongoingBackupId - + ". Can not launch new backup until no ongoing backup remains."); - throw new BackupException("There is ongoing backup."); - } - - // Initialize thread pools - int nrThreads = this.conf.getInt("hbase.backup.threads.max", 1); - ThreadFactoryBuilder builder = new ThreadFactoryBuilder(); - builder.setNameFormat("BackupHandler-%1$d"); - this.pool = - new ThreadPoolExecutor(nrThreads, nrThreads, 60, TimeUnit.SECONDS, - new LinkedBlockingQueue<Runnable>(), builder.build()); - ((ThreadPoolExecutor) pool).allowCoreThreadTimeOut(true); - } - - /** - * Dispatch and handle a backup request. - * @param backupContext backup context - * @throws BackupException exception - */ - public void dispatchRequest(BackupContext backupContext) throws BackupException { - - this.backupContext = backupContext; - - LOG.info("Got a backup request: " + "Type: " + backupContext.getType() + "; Tables: " - + backupContext.getTableListAsString() + "; TargetRootDir: " - + backupContext.getTargetRootDir()); - - // dispatch the request to a backup handler and put it handler map - - BackupHandler handler = new BackupHandler(this.backupContext, this, conf); - Future<Object> future = this.pool.submit(handler); - // wait for the execution to complete - try { - future.get(); - } catch (InterruptedException e) { - throw new BackupException(e); - } catch (CancellationException e) { - throw new BackupException(e); - } catch (ExecutionException e) { - throw new BackupException(e); - } - - // mark the backup complete for exit handler's processing - backupComplete = true; - - LOG.info("Backup request " + backupContext.getBackupId() + " has been executed."); - } - - /** - * Get direct ancestors of the current backup. - * @param backupCtx The backup context for the current backup - * @return The ancestors for the current backup - * @throws IOException exception - * @throws BackupException exception - */ - protected ArrayList<BackupImage> getAncestors(BackupContext backupCtx) throws IOException, - BackupException { - LOG.debug("Getting the direct ancestors of the current backup ..."); - - ArrayList<BackupImage> ancestors = new ArrayList<BackupImage>(); - - // full backup does not have ancestor - if (backupCtx.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - LOG.debug("Current backup is a full backup, no direct ancestor for it."); - return ancestors; - } - - // get all backup history list in descending order - - ArrayList<BackupCompleteData> allHistoryList = getBackupHistory(); - for (BackupCompleteData backup : allHistoryList) { - BackupImage image = - new BackupImage(backup.getBackupToken(), backup.getType(), backup.getBackupRootPath(), - backup.getTableList(), Long.parseLong(backup.getStartTime()), Long.parseLong(backup - .getEndTime())); - - // add the full backup image as an ancestor until the last incremental backup - if (backup.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - - // backup image from existing snapshot does not involve in dependency - if (backup.fromExistingSnapshot()) { - continue; - } - // check the backup image coverage, if previous image could be covered by the newer ones, - // then no need to add - if (!BackupManifest.canCoverImage(ancestors, image)) { - ancestors.add(image); - } - } else { - // found last incremental backup, if previously added full backup ancestor images can cover - // it, then this incremental ancestor is not the dependent of the current incremental - // backup, that is to say, this is the backup scope boundary of current table set. - // Otherwise, this incremental backup ancestor is the dependent ancestor of the ongoing - // incremental backup - if (BackupManifest.canCoverImage(ancestors, image)) { - LOG.debug("Met the backup boundary of the current table set. " - + "The root full backup images for the current backup scope:"); - for (BackupImage image1 : ancestors) { - LOG.debug(" BackupId: " + image1.getBackupId() + ", Backup directory: " - + image1.getRootDir()); - } - } else { - Path logBackupPath = - HBackupFileSystem.getLogBackupPath(backup.getBackupRootPath(), - backup.getBackupToken()); - LOG.debug("Current backup has an incremental backup ancestor, " - + "touching its image manifest in " + logBackupPath.toString() - + " to construct the dependency."); - - BackupManifest lastIncrImgManifest = new BackupManifest(conf, logBackupPath); - BackupImage lastIncrImage = lastIncrImgManifest.getBackupImage(); - ancestors.add(lastIncrImage); - - LOG.debug("Last dependent incremental backup image information:"); - LOG.debug(" Token: " + lastIncrImage.getBackupId()); - LOG.debug(" Backup directory: " + lastIncrImage.getRootDir()); - } - } - } - LOG.debug("Got " + ancestors.size() + " ancestors for the current backup."); - return ancestors; - } - - /** - * Get the direct ancestors of this backup for one table involved. - * @param backupContext backup context - * @param table table - * @return backupImages on the dependency list - * @throws BackupException exception - * @throws IOException exception - */ - protected ArrayList<BackupImage> getAncestors(BackupContext backupContext, String table) - throws BackupException, IOException { - ArrayList<BackupImage> ancestors = getAncestors(backupContext); - ArrayList<BackupImage> tableAncestors = new ArrayList<BackupImage>(); - for (BackupImage image : ancestors) { - if (image.hasTable(table)) { - tableAncestors.add(image); - if (image.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - break; - } - } - } - return tableAncestors; - } - - /** - * hbase:backup operations - */ - - /** - * Updates status (state) of a backup session in a persistent store - * @param context context - * @throws IOException exception - */ - public void updateBackupStatus(BackupContext context) throws IOException { - systemTable.updateBackupStatus(context); - } - - /** - * Read the last backup start code (timestamp) of last successful backup. Will return null - * if there is no startcode stored in hbase:backup or the value is of length 0. These two - * cases indicate there is no successful backup completed so far. - * @return the timestamp of a last successful backup - * @throws IOException exception - */ - public String readBackupStartCode() throws IOException { - return systemTable.readBackupStartCode(); - } - - /** - * Write the start code (timestamp) to hbase:backup. If passed in null, then write 0 byte. - * @param startCode start code - * @throws IOException exception - */ - public void writeBackupStartCode(String startCode) throws IOException { - systemTable.writeBackupStartCode(startCode); - } - - /** - * Get the RS log information after the last log roll from hbase:backup. - * @return RS log info - * @throws IOException exception - */ - public HashMap<String, String> readRegionServerLastLogRollResult() throws IOException { - return systemTable.readRegionServerLastLogRollResult(); - } - - /** - * Get all completed backup information (in desc order by time) - * @return history info of BackupCompleteData - * @throws IOException exception - */ - public ArrayList<BackupCompleteData> getBackupHistory() throws IOException { - return systemTable.getBackupHistory(); - } - - /** - * Write the current timestamps for each regionserver to hbase:backup after a successful full or - * incremental backup. Each table may have a different set of log timestamps. The saved timestamp - * is of the last log file that was backed up already. - * @param tables tables - * @throws IOException exception - */ - public void writeRegionServerLogTimestamp(Set<String> tables, - HashMap<String, String> newTimestamps) throws IOException { - systemTable.writeRegionServerLogTimestamp(tables, newTimestamps); - } - - /** - * Read the timestamp for each region server log after the last successful backup. Each table has - * its own set of the timestamps. The info is stored for each table as a concatinated string on ZK - * under //hbase//backup//incr//tablelogtimestamp//table_name - * @return the timestamp for each region server. key: tableName value: - * RegionServer,PreviousTimeStamp - * @throws IOException exception - */ - public HashMap<String, HashMap<String, String>> readLogTimestampMap() throws IOException { - return systemTable.readLogTimestampMap(); - } - - /** - * Return the current tables covered by incremental backup. - * @return set of tableNames - * @throws IOException exception - */ - public Set<String> getIncrementalBackupTableSet() throws IOException { - return systemTable.getIncrementalBackupTableSet(); - } - - /** - * Adds set of tables to overall incremental backup table set - * @param tables tables - * @throws IOException exception - */ - public void addIncrementalBackupTableSet(Set<String> tables) throws IOException { - systemTable.addIncrementalBackupTableSet(tables); - } - - /** - * Saves list of WAL files after incremental backup operation. These files will be stored until - * TTL expiration and are used by Backup Log Cleaner plugin to determine which WAL files can be - * safely purged. - */ - - public void recordWALFiles(List<String> files) throws IOException { - systemTable.addWALFiles(files, backupContext.getBackupId()); - } - -} http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManifest.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManifest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManifest.java deleted file mode 100644 index f41540b..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupManifest.java +++ /dev/null @@ -1,814 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.backup; - -import java.io.File; -import java.io.IOException; -import java.util.ArrayList; -import java.util.Collections; -import java.util.HashMap; -import java.util.LinkedList; -import java.util.Map; -import java.util.Map.Entry; -import java.util.Properties; -import java.util.TreeMap; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FSDataInputStream; -import org.apache.hadoop.fs.FSDataOutputStream; -import org.apache.hadoop.fs.FileStatus; -import org.apache.hadoop.fs.FileSystem; -import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.hbase.util.FSUtils; -import org.codehaus.jackson.JsonGenerationException; -import org.codehaus.jackson.JsonParseException; -import org.codehaus.jackson.map.JsonMappingException; -import org.codehaus.jackson.map.ObjectMapper; - -/** - * Backup manifest Contains all the meta data of a backup image. The manifest info will be bundled - * as manifest file together with data. So that each backup image will contain all the info needed - * for restore. - */ [email protected] [email protected] -public class BackupManifest { - - private static final Log LOG = LogFactory.getLog(BackupManifest.class); - - // manifest file name - public static final String FILE_NAME = ".backup.manifest"; - - // manifest file version, current is 1.0 - public static final String MANIFEST_VERSION = "1.0"; - - // tags of fields for manifest file - public static final String TAG_VERSION = "Manifest-Version"; - public static final String TAG_BACKUPID = "Backup-Id"; - public static final String TAG_BACKUPTYPE = "Backup-Type"; - public static final String TAG_TABLESET = "Table-Set"; - public static final String TAG_STARTTS = "Start-Timestamp"; - public static final String TAG_COMPLETETS = "Complete-Timestamp"; - public static final String TAG_TABLEBYTES = "Total-Table-Bytes"; - public static final String TAG_LOGBYTES = "Total-Log-Bytes"; - public static final String TAG_INCRTIMERANGE = "Incremental-Time-Range"; - public static final String TAG_DEPENDENCY = "Dependency"; - public static final String TAG_IMAGESTATE = "Image-State"; - public static final String TAG_COMPACTION = "Compaction"; - - public static final String ERROR_DEPENDENCY = "DEPENDENCY_ERROR"; - - public static final int DELETE_SUCCESS = 0; - public static final int DELETE_FAILED = -1; - - // currently only one state, will have CONVERTED, and MERGED in future JIRA - public static final String IMAGE_STATE_ORIG = "ORIGINAL"; - public static final String IMAGE_STATE_CONVERT = "CONVERTED"; - public static final String IMAGE_STATE_MERGE = "MERGED"; - public static final String IMAGE_STATE_CONVERT_MERGE = "CONVERTED,MERGED"; - - // backup image, the dependency graph is made up by series of backup images - - public static class BackupImage implements Comparable<BackupImage> { - - private String backupId; - private String type; - private String rootDir; - private String tableSet; - private long startTs; - private long completeTs; - private ArrayList<BackupImage> ancestors; - - public BackupImage() { - super(); - } - - public BackupImage(String backupId, String type, String rootDir, String tableSet, long startTs, - long completeTs) { - this.backupId = backupId; - this.type = type; - this.rootDir = rootDir; - this.tableSet = tableSet; - this.startTs = startTs; - this.completeTs = completeTs; - } - - public String getBackupId() { - return backupId; - } - - public void setBackupId(String backupId) { - this.backupId = backupId; - } - - public String getType() { - return type; - } - - public void setType(String type) { - this.type = type; - } - - public String getRootDir() { - return rootDir; - } - - public void setRootDir(String rootDir) { - this.rootDir = rootDir; - } - - public String getTableSet() { - return tableSet; - } - - public void setTableSet(String tableSet) { - this.tableSet = tableSet; - } - - public long getStartTs() { - return startTs; - } - - public void setStartTs(long startTs) { - this.startTs = startTs; - } - - public long getCompleteTs() { - return completeTs; - } - - public void setCompleteTs(long completeTs) { - this.completeTs = completeTs; - } - - public ArrayList<BackupImage> getAncestors() { - if (this.ancestors == null) { - this.ancestors = new ArrayList<BackupImage>(); - } - return this.ancestors; - } - - public void addAncestor(BackupImage backupImage) { - this.getAncestors().add(backupImage); - } - - public boolean hasAncestor(String token) { - for (BackupImage image : this.getAncestors()) { - if (image.getBackupId().equals(token)) { - return true; - } - } - return false; - } - - public boolean hasTable(String table) { - String[] tables = this.getTableSet().split(";"); - for (String t : tables) { - if (t.equals(table)) { - return true; - } - } - return false; - } - - @Override - public int compareTo(BackupImage other) { - String thisBackupId = this.getBackupId(); - String otherBackupId = other.getBackupId(); - Long thisTS = new Long(thisBackupId.substring(thisBackupId.lastIndexOf("_") + 1)); - Long otherTS = new Long(otherBackupId.substring(otherBackupId.lastIndexOf("_") + 1)); - return thisTS.compareTo(otherTS); - } - } - - // manifest version - private String version = MANIFEST_VERSION; - - // hadoop hbase configuration - protected Configuration config = null; - - // backup root directory - private String rootDir = null; - - // backup image directory - private String tableBackupDir = null; - - // backup log directory if this is an incremental backup - private String logBackupDir = null; - - // backup token - private String token; - - // backup type, full or incremental - private String type; - - // the table set for the backup - private ArrayList<String> tableSet; - - // actual start timestamp of the backup process - private long startTs; - - // actual complete timestamp of the backup process - private long completeTs; - - // total bytes for table backup image - private long tableBytes; - - // total bytes for the backed-up logs for incremental backup - private long logBytes; - - // the region server timestamp for tables: - // <table, <rs, timestamp>> - private Map<String, HashMap<String, String>> incrTimeRanges; - - // dependency of this backup, including all the dependent images to do PIT recovery - private Map<String, BackupImage> dependency; - - // the state of backup image - private String imageState; - - // the indicator of the image compaction - private boolean isCompacted = false; - - // the merge chain of the original backups, null if not a merged backup - private LinkedList<String> mergeChain; - - /** - * Construct manifest for a ongoing backup. - * @param backupCtx The ongoing backup context - */ - public BackupManifest(BackupContext backupCtx) { - this.token = backupCtx.getBackupId(); - this.type = backupCtx.getType(); - this.rootDir = backupCtx.getTargetRootDir(); - if (this.type.equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - this.logBackupDir = backupCtx.getHLogTargetDir(); - this.logBytes = backupCtx.getTotalBytesCopied(); - } - this.startTs = backupCtx.getStartTs(); - this.completeTs = backupCtx.getEndTs(); - this.loadTableSet(backupCtx.getTableListAsString()); - this.setImageOriginal(); - } - - /** - * Construct a table level manifest for a backup of the named table. - * @param backupCtx The ongoing backup context - */ - public BackupManifest(BackupContext backupCtx, String table) { - this.token = backupCtx.getBackupId(); - this.type = backupCtx.getType(); - this.rootDir = backupCtx.getTargetRootDir(); - this.tableBackupDir = backupCtx.getBackupStatus(table).getTargetDir(); - if (this.type.equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - this.logBackupDir = backupCtx.getHLogTargetDir(); - this.logBytes = backupCtx.getTotalBytesCopied(); - } - this.startTs = backupCtx.getStartTs(); - this.completeTs = backupCtx.getEndTs(); - this.loadTableSet(table); - this.setImageOriginal(); - } - - /** - * Construct manifest from a backup directory. - * @param conf configuration - * @param backupPath backup path - * @throws BackupException exception - */ - public BackupManifest(Configuration conf, Path backupPath) throws BackupException { - - LOG.debug("Loading manifest from: " + backupPath.toString()); - // The input backupDir may not exactly be the backup table dir. - // It could be the backup log dir where there is also a manifest file stored. - // This variable's purpose is to keep the correct and original location so - // that we can store/persist it. - this.tableBackupDir = backupPath.toString(); - this.config = conf; - try { - - FileSystem fs = backupPath.getFileSystem(conf); - FileStatus[] subFiles = FSUtils.listStatus(fs, backupPath); - if (subFiles == null) { - String errorMsg = backupPath.toString() + " does not exist"; - LOG.error(errorMsg); - throw new IOException(errorMsg); - } - for (FileStatus subFile : subFiles) { - if (subFile.getPath().getName().equals(FILE_NAME)) { - - // load and set manifest field from file content - FSDataInputStream in = fs.open(subFile.getPath()); - Properties props = new Properties(); - try { - props.load(in); - } catch (IOException e) { - LOG.error("Error when loading from manifest file!"); - throw e; - } finally { - in.close(); - } - - this.version = props.getProperty(TAG_VERSION); - this.token = props.getProperty(TAG_BACKUPID); - this.type = props.getProperty(TAG_BACKUPTYPE); - // Here the parameter backupDir is where the manifest file is. - // There should always be a manifest file under: - // backupRootDir/namespace/table/backupId/.backup.manifest - this.rootDir = backupPath.getParent().getParent().getParent().toString(); - - Path p = backupPath.getParent(); - if (p.getName().equals(HConstants.HREGION_LOGDIR_NAME)) { - this.rootDir = p.getParent().toString(); - } else { - this.rootDir = p.getParent().getParent().toString(); - } - - this.loadTableSet(props.getProperty(TAG_TABLESET)); - - this.startTs = Long.parseLong(props.getProperty(TAG_STARTTS)); - this.completeTs = Long.parseLong(props.getProperty(TAG_COMPLETETS)); - this.tableBytes = Long.parseLong(props.getProperty(TAG_TABLEBYTES)); - if (this.type.equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - this.logBytes = (Long.parseLong(props.getProperty(TAG_LOGBYTES))); - LOG.debug("convert will be implemented by future jira"); - } - this.loadIncrementalTimeRanges(props.getProperty(TAG_INCRTIMERANGE)); - this.loadDependency(props.getProperty(TAG_DEPENDENCY)); - this.imageState = props.getProperty(TAG_IMAGESTATE); - this.isCompacted = - props.getProperty(TAG_COMPACTION).equalsIgnoreCase("TRUE") ? true : false; - LOG.debug("merge and from existing snapshot will be implemented by future jira"); - LOG.debug("Loaded manifest instance from manifest file: " - + FSUtils.getPath(subFile.getPath())); - return; - } - } - String errorMsg = "No manifest file found in: " + backupPath.toString(); - LOG.error(errorMsg); - throw new IOException(errorMsg); - - } catch (IOException e) { - throw new BackupException(e.getMessage()); - } - } - - public String getType() { - return type; - } - - public void setType(String type) { - this.type = type; - } - - /** - * Load table set from a table set list string (t1;t2;t3;...). - * @param tableSetStr Table set list string - */ - private void loadTableSet(String tableSetStr) { - - LOG.debug("Loading table set: " + tableSetStr); - - String[] tableSet = tableSetStr.split(";"); - this.tableSet = this.getTableSet(); - if (this.tableSet.size() > 0) { - this.tableSet.clear(); - } - for (int i = 0; i < tableSet.length; i++) { - this.tableSet.add(tableSet[i]); - } - - LOG.debug(tableSet.length + " tables exist in table set."); - } - - public void setImageOriginal() { - this.imageState = IMAGE_STATE_ORIG; - } - - /** - * Get the table set of this image. - * @return The table set list - */ - public ArrayList<String> getTableSet() { - if (this.tableSet == null) { - this.tableSet = new ArrayList<String>(); - } - return this.tableSet; - } - - /** - * Persist the manifest file. - * @throws IOException IOException when storing the manifest file. - */ - public void store(Configuration conf) throws BackupException { - Properties props = new Properties(); - props.setProperty(TAG_VERSION, this.version); - props.setProperty(TAG_BACKUPID, this.token); - props.setProperty(TAG_BACKUPTYPE, this.type); - props.setProperty(TAG_TABLESET, this.getTableSetStr()); - LOG.debug("convert will be supported in future jira"); - // String convertedTables = this.getConvertedTableSetStr(); - // if (convertedTables != null ) - // props.setProperty(TAG_CONVERTEDTABLESET, convertedTables); - props.setProperty(TAG_STARTTS, Long.toString(this.startTs)); - props.setProperty(TAG_COMPLETETS, Long.toString(this.completeTs)); - props.setProperty(TAG_TABLEBYTES, Long.toString(this.tableBytes)); - if (this.type.equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - props.setProperty(TAG_LOGBYTES, Long.toString(this.logBytes)); - } - props.setProperty(TAG_INCRTIMERANGE, this.getIncrTimestampStr()); - props.setProperty(TAG_DEPENDENCY, this.getDependencyStr()); - props.setProperty(TAG_IMAGESTATE, this.getImageState()); - props.setProperty(TAG_COMPACTION, this.isCompacted ? "TRUE" : "FALSE"); - LOG.debug("merge will be supported in future jira"); - // props.setProperty(TAG_MERGECHAIN, this.getMergeChainStr()); - LOG.debug("backup from existing snapshot will be supported in future jira"); - // props.setProperty(TAG_FROMSNAPSHOT, this.isFromSnapshot() ? "TRUE" : "FALSE"); - - // write the file, overwrite if already exist - Path manifestFilePath = - new Path((this.tableBackupDir != null ? this.tableBackupDir : this.logBackupDir) - + File.separator + FILE_NAME); - try { - FSDataOutputStream out = manifestFilePath.getFileSystem(conf).create(manifestFilePath, true); - props.store(out, "HBase backup manifest."); - out.close(); - } catch (IOException e) { - throw new BackupException(e.getMessage()); - } - - LOG.debug("Manifest file stored to " + this.tableBackupDir != null ? this.tableBackupDir - : this.logBackupDir + File.separator + FILE_NAME); - } - - /** - * Get the table set string in the format of t1;t2;t3... - */ - private String getTableSetStr() { - return BackupUtil.concat(getTableSet(), ";"); - } - - public String getImageState() { - return imageState; - } - - public String getVersion() { - return version; - } - - /** - * Get this backup image. - * @return the backup image. - */ - public BackupImage getBackupImage() { - return this.getDependency().get(this.token); - } - - /** - * Add dependent backup image for this backup. - * @param image The direct dependent backup image - */ - public void addDependentImage(BackupImage image) { - this.getDependency().get(this.token).addAncestor(image); - this.setDependencyMap(this.getDependency(), image); - } - - /** - * Get the dependency' string in the json format. - */ - private String getDependencyStr() { - BackupImage thisImage = this.getDependency().get(this.token); - if (thisImage == null) { - LOG.warn("There is no dependency set yet."); - return null; - } - - ObjectMapper mapper = new ObjectMapper(); - try { - return mapper.writeValueAsString(thisImage); - } catch (JsonGenerationException e) { - LOG.error("Error when generating dependency string from backup image.", e); - return ERROR_DEPENDENCY; - } catch (JsonMappingException e) { - LOG.error("Error when generating dependency string from backup image.", e); - return ERROR_DEPENDENCY; - } catch (IOException e) { - LOG.error("Error when generating dependency string from backup image.", e); - return ERROR_DEPENDENCY; - } - } - - /** - * Get all dependent backup images. The image of this backup is also contained. - * @return The dependent backup images map - */ - public Map<String, BackupImage> getDependency() { - if (this.dependency == null) { - this.dependency = new HashMap<String, BackupImage>(); - LOG.debug(this.rootDir + " " + this.token + " " + this.type); - this.dependency.put(this.token, - new BackupImage(this.token, this.type, this.rootDir, this.getTableSetStr(), this.startTs, - this.completeTs)); - } - return this.dependency; - } - - /** - * Set the incremental timestamp map directly. - * @param incrTimestampMap timestamp map - */ - public void setIncrTimestampMap(HashMap<String, HashMap<String, String>> incrTimestampMap) { - this.incrTimeRanges = incrTimestampMap; - } - - /** - * Get the incremental time range string in the format of: - * t1,rs1:ts,rs2:ts,...;t2,rs1:ts,rs2:ts,...;t3,rs1:ts,rs2:ts,... - */ - private String getIncrTimestampStr() { - StringBuilder sb = new StringBuilder(); - for (Entry<String, HashMap<String, String>> tableEntry : this.getIncrTimestamps().entrySet()) { - sb.append(tableEntry.getKey() + ","); // table - for (Entry<String, String> rsEntry : tableEntry.getValue().entrySet()) { - sb.append(rsEntry.getKey() + ":"); // region server - sb.append(rsEntry.getValue() + ","); // timestamp - } - if (sb.length() > 1 && sb.charAt(sb.length() - 1) == ',') { - sb.deleteCharAt(sb.length() - 1); - } - sb.append(";"); - } - if (sb.length() > 1 && sb.charAt(sb.length() - 1) == ';') { - sb.deleteCharAt(sb.length() - 1); - } - return sb.toString(); - } - - public Map<String, HashMap<String, String>> getIncrTimestamps() { - if (this.incrTimeRanges == null) { - this.incrTimeRanges = new HashMap<String, HashMap<String, String>>(); - } - return this.incrTimeRanges; - } - - /** - * Load incremental timestamps from a given string, and store them in the collection. The - * timestamps in string is in the format of - * t1,rs1:ts,rs2:ts,...;t2,rs1:ts,rs2:ts,...;t3,rs1:ts,rs2:ts,... - * @param timeRangesInStr Incremental time ranges in string - */ - private void loadIncrementalTimeRanges(String timeRangesStr) throws IOException { - - LOG.debug("Loading table's incremental time ranges of region servers from string in manifest: " - + timeRangesStr); - - Map<String, HashMap<String, String>> timeRangeMap = this.getIncrTimestamps(); - - String[] entriesOfTables = timeRangesStr.split(";"); - for (int i = 0; i < entriesOfTables.length; i++) { - String[] itemsForTable = entriesOfTables[i].split(","); - - // validate the incremental timestamps string format for a table: - // t1,rs1:ts,rs2:ts,... - if (itemsForTable.length < 1) { - String errorMsg = "Wrong incremental time range format: " + timeRangesStr; - LOG.error(errorMsg); - throw new IOException(errorMsg); - } - - HashMap<String, String> rsTimestampMap = new HashMap<String, String>(); - for (int j = 1; j < itemsForTable.length; j++) { - String[] rsTsEntry = itemsForTable[j].split(":"); - - // validate the incremental timestamps string format for a region server: - // rs1:ts - if (rsTsEntry.length != 2) { - String errorMsg = "Wrong incremental timestamp format: " + itemsForTable[j]; - LOG.error(errorMsg); - throw new IOException(errorMsg); - } - - // an entry for timestamp of a region server - rsTimestampMap.put(rsTsEntry[0], rsTsEntry[1]); - } - - timeRangeMap.put(itemsForTable[0], rsTimestampMap); - } - - // all entries have been loaded - LOG.debug(entriesOfTables.length + " tables' incremental time ranges have been loaded."); - } - - /** - * Get the image list of this backup for restore in time order. - * @param reverse If true, then output in reverse order, otherwise in time order from old to new - * @return the backup image list for restore in time order - */ - public ArrayList<BackupImage> getRestoreDependentList(boolean reverse) { - TreeMap<Long, BackupImage> restoreImages = new TreeMap<Long, BackupImage>(); - for (BackupImage image : this.getDependency().values()) { - restoreImages.put(Long.valueOf(image.startTs), image); - } - return new ArrayList<BackupImage>(reverse ? (restoreImages.descendingMap().values()) - : (restoreImages.values())); - } - - /** - * Get the dependent image list for a specific table of this backup in time order from old to new - * if want to restore to this backup image level. - * @param table table - * @return the backup image list for a table in time order - */ - public ArrayList<BackupImage> getDependentListByTable(String table) { - ArrayList<BackupImage> tableImageList = new ArrayList<BackupImage>(); - ArrayList<BackupImage> imageList = getRestoreDependentList(true); - for (BackupImage image : imageList) { - if (image.hasTable(table)) { - tableImageList.add(image); - if (image.getType().equals(BackupRestoreConstants.BACKUP_TYPE_FULL)) { - break; - } - } - } - Collections.reverse(tableImageList); - return tableImageList; - } - - /** - * Get the full dependent image list in the whole dependency scope for a specific table of this - * backup in time order from old to new. - * @param table table - * @return the full backup image list for a table in time order in the whole scope of the - * dependency of this image - */ - public ArrayList<BackupImage> getAllDependentListByTable(String table) { - ArrayList<BackupImage> tableImageList = new ArrayList<BackupImage>(); - ArrayList<BackupImage> imageList = getRestoreDependentList(false); - for (BackupImage image : imageList) { - if (image.hasTable(table)) { - tableImageList.add(image); - } - } - return tableImageList; - } - - /** - * Load dependency from a dependency json string. - * @param dependencyStr The dependency string - * @throws IOException exception - */ - private void loadDependency(String dependencyStr) throws IOException { - - LOG.debug("Loading dependency: " + dependencyStr); - - String msg = "Dependency is broken in the manifest."; - if (dependencyStr.equals(ERROR_DEPENDENCY)) { - throw new IOException(msg); - } - - ObjectMapper mapper = new ObjectMapper(); - BackupImage image = null; - try { - image = mapper.readValue(dependencyStr, BackupImage.class); - } catch (JsonParseException e) { - LOG.error(msg); - throw new IOException(e.getMessage()); - } catch (JsonMappingException e) { - LOG.error(msg); - throw new IOException(e.getMessage()); - } catch (IOException e) { - LOG.error(msg); - throw new IOException(e.getMessage()); - } - LOG.debug("Manifest's current backup image information:"); - LOG.debug(" Token: " + image.getBackupId()); - LOG.debug(" Backup directory: " + image.getRootDir()); - this.setDependencyMap(this.getDependency(), image); - - LOG.debug("Dependent images map:"); - for (Entry<String, BackupImage> entry : this.getDependency().entrySet()) { - LOG.debug(" " + entry.getKey() + " : " + entry.getValue().getBackupId() + " -- " - + entry.getValue().getRootDir()); - } - - LOG.debug("Dependency has been loaded."); - } - - /** - * Recursively set the dependency map of the backup images. - * @param map The dependency map - * @param image The backup image - */ - private void setDependencyMap(Map<String, BackupImage> map, BackupImage image) { - if (image == null) { - return; - } else { - map.put(image.getBackupId(), image); - for (BackupImage img : image.getAncestors()) { - setDependencyMap(map, img); - } - } - } - - /** - * Check whether backup image1 could cover backup image2 or not. - * @param image1 backup image 1 - * @param image2 backup image 2 - * @return true if image1 can cover image2, otherwise false - */ - public static boolean canCoverImage(BackupImage image1, BackupImage image2) { - // image1 can cover image2 only when the following conditions are satisfied: - // - image1 must not be an incremental image; - // - image1 must be taken after image2 has been taken; - // - table set of image1 must cover the table set of image2. - if (image1.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - return false; - } - if (image1.getStartTs() < image2.getStartTs()) { - return false; - } - String[] image1TableSet = image1.getTableSet().split(";"); - String[] image2TableSet = image2.getTableSet().split(";"); - boolean found = false; - for (int i = 0; i < image2TableSet.length; i++) { - found = false; - for (int j = 0; j < image1TableSet.length; j++) { - if (image2TableSet[i].equals(image1TableSet[j])) { - found = true; - break; - } - } - if (!found) { - return false; - } - } - - LOG.debug("Backup image " + image1.getBackupId() + " can cover " + image2.getBackupId()); - return true; - } - - /** - * Check whether backup image set could cover a backup image or not. - * @param fullImages The backup image set - * @param image The target backup image - * @return true if fullImages can cover image, otherwise false - */ - public static boolean canCoverImage(ArrayList<BackupImage> fullImages, BackupImage image) { - // fullImages can cover image only when the following conditions are satisfied: - // - each image of fullImages must not be an incremental image; - // - each image of fullImages must be taken after image has been taken; - // - sum table set of fullImages must cover the table set of image. - for (BackupImage image1 : fullImages) { - if (image1.getType().equals(BackupRestoreConstants.BACKUP_TYPE_INCR)) { - return false; - } - if (image1.getStartTs() < image.getStartTs()) { - return false; - } - } - - ArrayList<String> image1TableSet = new ArrayList<String>(); - for (BackupImage image1 : fullImages) { - String[] tableSet = image1.getTableSet().split(";"); - for (String table : tableSet) { - image1TableSet.add(table); - } - } - ArrayList<String> image2TableSet = new ArrayList<String>(); - String[] tableSet = image.getTableSet().split(";"); - for (String table : tableSet) { - image2TableSet.add(table); - } - - for (int i = 0; i < image2TableSet.size(); i++) { - if (image1TableSet.contains(image2TableSet.get(i)) == false) { - return false; - } - } - - LOG.debug("Full image set can cover image " + image.getBackupId()); - return true; - } - -} http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreConstants.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreConstants.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreConstants.java deleted file mode 100644 index 9610b27..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreConstants.java +++ /dev/null @@ -1,66 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.backup; - -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; - -/** - * HConstants holds a bunch of HBase Backup and Restore constants - */ [email protected] [email protected] -public final class BackupRestoreConstants { - - // constants for znode data keys in backup znode - public static final String BACKUP_PROGRESS = "progress"; - public static final String BACKUP_START_TIME = "startTs"; - public static final String BACKUP_INPROGRESS_PHASE = "phase"; - public static final String BACKUP_COMPLETE_TIME = "completeTs"; - public static final String BACKUP_FAIL_TIME = "failedTs"; - public static final String BACKUP_FAIL_PHASE = "failedphase"; - public static final String BACKUP_FAIL_MSG = "failedmessage"; - public static final String BACKUP_ROOT_PATH = "targetRootDir"; - public static final String BACKUP_REQUEST_TABLE_LIST = "tablelist"; - public static final String BACKUP_REQUEST_TYPE = "type"; - public static final String BACKUP_BYTES_COPIED = "bytescopied"; - public static final String BACKUP_ANCESTORS = "ancestors"; - public static final String BACKUP_EXISTINGSNAPSHOT = "snapshot"; - - public static final String BACKUP_TYPE_FULL = "full"; - public static final String BACKUP_TYPE_INCR = "incremental"; - - // delimiter in tablename list in restore command - public static final String TABLENAME_DELIMITER_IN_COMMAND = ","; - - // delimiter in znode data - public static final String ZNODE_DATA_DELIMITER = ","; - - public static final String CONF_STAGING_ROOT = "snapshot.export.staging.root"; - - public static final String BACKUPID_PREFIX = "backup_"; - - public static enum BACKUP_COMMAND { - CREATE, CANCEL, DELETE, DESCRIBE, HISTORY, STATUS, CONVERT, MERGE, STOP, SHOW, HELP, - } - - private BackupRestoreConstants() { - // Can't be instantiated with this ctor. - } -} http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreServiceFactory.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreServiceFactory.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreServiceFactory.java deleted file mode 100644 index 1c38cf4..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupRestoreServiceFactory.java +++ /dev/null @@ -1,61 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -package org.apache.hadoop.hbase.backup; - -import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hbase.backup.mapreduce.MapReduceBackupCopyService; -import org.apache.hadoop.hbase.backup.mapreduce.MapReduceRestoreService; -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; -import org.apache.hadoop.util.ReflectionUtils; - [email protected] [email protected] -public final class BackupRestoreServiceFactory { - - public final static String HBASE_INCR_RESTORE_IMPL_CLASS = "hbase.incremental.restore.class"; - public final static String HBASE_BACKUP_COPY_IMPL_CLASS = "hbase.backup.copy.class"; - - private BackupRestoreServiceFactory(){ - throw new AssertionError("Instantiating utility class..."); - } - - /** - * Gets incremental restore service - * @param conf - configuration - * @return incremental backup service instance - */ - public static IncrementalRestoreService getIncrementalRestoreService(Configuration conf) { - Class<? extends IncrementalRestoreService> cls = - conf.getClass(HBASE_INCR_RESTORE_IMPL_CLASS, MapReduceRestoreService.class, - IncrementalRestoreService.class); - return ReflectionUtils.newInstance(cls, conf); - } - - /** - * Gets backup copy service - * @param conf - configuration - * @return backup copy service - */ - public static BackupCopyService getBackupCopyService(Configuration conf) { - Class<? extends BackupCopyService> cls = - conf.getClass(HBASE_BACKUP_COPY_IMPL_CLASS, MapReduceBackupCopyService.class, - BackupCopyService.class); - return ReflectionUtils.newInstance(cls, conf); - } -} http://git-wip-us.apache.org/repos/asf/hbase/blob/449fb812/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupStatus.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupStatus.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupStatus.java deleted file mode 100644 index c26e1d4..0000000 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/BackupStatus.java +++ /dev/null @@ -1,67 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.backup; - -import java.io.Serializable; - -import org.apache.hadoop.hbase.classification.InterfaceAudience; -import org.apache.hadoop.hbase.classification.InterfaceStability; - -/** - * Backup status and related information encapsulated for a table. - * At this moment only TargetDir and SnapshotName is encapsulated here. - * future Jira will be implemented for progress, bytesCopies, phase, etc. - */ - [email protected] [email protected] -public class BackupStatus implements Serializable { - - private static final long serialVersionUID = -5968397963548535982L; - - // table name for backup - private String table; - - // target directory of the backup image for this table - private String targetDir; - - // snapshot name for offline/online snapshot - private String snapshotName = null; - - public BackupStatus(String table, String targetRootDir, String backupId) { - this.table = table; - this.targetDir = HBackupFileSystem.getTableBackupDir(targetRootDir, backupId, table); - } - - public String getSnapshotName() { - return snapshotName; - } - - public void setSnapshotName(String snapshotName) { - this.snapshotName = snapshotName; - } - - public String getTargetDir() { - return targetDir; - } - - public String getTable() { - return table; - } -}
