http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/apidocs/overview-tree.html ---------------------------------------------------------------------- diff --git a/apidocs/overview-tree.html b/apidocs/overview-tree.html index 72a723d..d331c6f 100644 --- a/apidocs/overview-tree.html +++ b/apidocs/overview-tree.html @@ -891,32 +891,32 @@ <li type="circle">java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true" title="class or interface in java.lang"><span class="typeNameLink">Enum</span></a><E> (implements java.lang.<a href="https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true" title="class or interface in java.lang">Comparable</a><T>, java.io.<a href="https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true" title="class or interface in java.io">Serializable</a>) <ul> <li type="circle">org.apache.hadoop.hbase.util.<a href="org/apache/hadoop/hbase/util/Order.html" title="enum in org.apache.hadoop.hbase.util"><span class="typeNameLink">Order</span></a></li> -<li type="circle">org.apache.hadoop.hbase.<a href="org/apache/hadoop/hbase/KeepDeletedCells.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">KeepDeletedCells</span></a></li> <li type="circle">org.apache.hadoop.hbase.<a href="org/apache/hadoop/hbase/MemoryCompactionPolicy.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">MemoryCompactionPolicy</span></a></li> +<li type="circle">org.apache.hadoop.hbase.<a href="org/apache/hadoop/hbase/KeepDeletedCells.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">KeepDeletedCells</span></a></li> <li type="circle">org.apache.hadoop.hbase.<a href="org/apache/hadoop/hbase/CompareOperator.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CompareOperator</span></a></li> <li type="circle">org.apache.hadoop.hbase.<a href="org/apache/hadoop/hbase/CellBuilderType.html" title="enum in org.apache.hadoop.hbase"><span class="typeNameLink">CellBuilderType</span></a></li> -<li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">BitComparator.BitwiseOp</span></a></li> <li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/FilterList.Operator.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">FilterList.Operator</span></a></li> <li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/CompareFilter.CompareOp.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">CompareFilter.CompareOp</span></a></li> -<li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/Filter.ReturnCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">Filter.ReturnCode</span></a></li> +<li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/BitComparator.BitwiseOp.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">BitComparator.BitwiseOp</span></a></li> <li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/RegexStringComparator.EngineType.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">RegexStringComparator.EngineType</span></a></li> +<li type="circle">org.apache.hadoop.hbase.filter.<a href="org/apache/hadoop/hbase/filter/Filter.ReturnCode.html" title="enum in org.apache.hadoop.hbase.filter"><span class="typeNameLink">Filter.ReturnCode</span></a></li> <li type="circle">org.apache.hadoop.hbase.io.encoding.<a href="org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.html" title="enum in org.apache.hadoop.hbase.io.encoding"><span class="typeNameLink">DataBlockEncoding</span></a></li> <li type="circle">org.apache.hadoop.hbase.regionserver.<a href="org/apache/hadoop/hbase/regionserver/BloomType.html" title="enum in org.apache.hadoop.hbase.regionserver"><span class="typeNameLink">BloomType</span></a></li> +<li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/SpaceViolationPolicy.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">SpaceViolationPolicy</span></a></li> <li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/ThrottlingException.Type.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">ThrottlingException.Type</span></a></li> <li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/QuotaScope.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">QuotaScope</span></a></li> -<li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/ThrottleType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">ThrottleType</span></a></li> <li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/QuotaType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">QuotaType</span></a></li> -<li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/SpaceViolationPolicy.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">SpaceViolationPolicy</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/Durability.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Durability</span></a></li> +<li type="circle">org.apache.hadoop.hbase.quotas.<a href="org/apache/hadoop/hbase/quotas/ThrottleType.html" title="enum in org.apache.hadoop.hbase.quotas"><span class="typeNameLink">ThrottleType</span></a></li> <li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/SnapshotType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">SnapshotType</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/MasterSwitchType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MasterSwitchType</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/CompactType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactType</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/Durability.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Durability</span></a></li> <li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/MobCompactPartitionPolicy.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MobCompactPartitionPolicy</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/CompactionState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactionState</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/Scan.ReadType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Scan.ReadType</span></a></li> -<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/RequestController.ReturnCode.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">RequestController.ReturnCode</span></a></li> <li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/IsolationLevel.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">IsolationLevel</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/RequestController.ReturnCode.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">RequestController.ReturnCode</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/Scan.ReadType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Scan.ReadType</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/CompactionState.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactionState</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/MasterSwitchType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">MasterSwitchType</span></a></li> <li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/Consistency.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">Consistency</span></a></li> +<li type="circle">org.apache.hadoop.hbase.client.<a href="org/apache/hadoop/hbase/client/CompactType.html" title="enum in org.apache.hadoop.hbase.client"><span class="typeNameLink">CompactType</span></a></li> <li type="circle">org.apache.hadoop.hbase.client.security.<a href="org/apache/hadoop/hbase/client/security/SecurityCapability.html" title="enum in org.apache.hadoop.hbase.client.security"><span class="typeNameLink">SecurityCapability</span></a></li> </ul> </li>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/book.html ---------------------------------------------------------------------- diff --git a/book.html b/book.html index 79cba7d..80bdf26 100644 --- a/book.html +++ b/book.html @@ -6610,12 +6610,56 @@ Quitting...</code></pre> <h2 id="_upgrade_paths"><a class="anchor" href="#_upgrade_paths"></a>13. Upgrade Paths</h2> <div class="sectionbody"> <div class="sect2"> -<h3 id="upgrade1.0"><a class="anchor" href="#upgrade1.0"></a>13.1. Upgrading from 0.98.x to 1.x</h3> +<h3 id="upgrade1.4"><a class="anchor" href="#upgrade1.4"></a>13.1. Upgrading to 1.4+</h3> +<div class="sect3"> +<h4 id="_replication_peer_s_tablecfs_config"><a class="anchor" href="#_replication_peer_s_tablecfs_config"></a>13.1.1. Replication peer’s TableCFs config</h4> +<div class="paragraph"> +<p>Before 1.4, the table name can’t include namespace for replication peer’s TableCFs config. It was fixed by add TableCFs to ReplicationPeerConfig which was stored on Zookeeper. So when upgrade to 1.4, you have to update the original ReplicationPeerConfig data on Zookeeper firstly. There are four steps to upgrade when your cluster have a replication peer with TableCFs config.</p> +</div> +<div class="ulist"> +<ul> +<li> +<p>Disable the replication peer.</p> +</li> +<li> +<p>If master has permission to write replication peer znode, then rolling update master directly. If not, use TableCFsUpdater tool to update the replication peer’s config.</p> +</li> +</ul> +</div> +<div class="listingblock"> +<div class="content"> +<pre>$ bin/hbase org.apache.hadoop.hbase.replication.master.TableCFsUpdater update</pre> +</div> +</div> +<div class="ulist"> +<ul> +<li> +<p>Rolling update regionservers.</p> +</li> +<li> +<p>Enable the replication peer.</p> +</li> +</ul> +</div> +<div class="paragraph"> +<p>Notes:</p> +</div> +<div class="ulist"> +<ul> +<li> +<p>Can’t use the old client(before 1.4) to change the replication peer’s config. Because the client will write config to Zookeeper directly, the old client will miss TableCFs config. And the old client write TableCFs config to the old tablecfs znode, it will not work for new version regionserver.</p> +</li> +</ul> +</div> +</div> +</div> +<div class="sect2"> +<h3 id="upgrade1.0"><a class="anchor" href="#upgrade1.0"></a>13.2. Upgrading from 0.98.x to 1.x</h3> <div class="paragraph"> <p>In this section we first note the significant changes that come in with 1.0.0+ HBase and then we go over the upgrade process. Be sure to read the significant changes section with care so you avoid surprises.</p> </div> <div class="sect3"> -<h4 id="_changes_of_note"><a class="anchor" href="#_changes_of_note"></a>13.1.1. Changes of Note!</h4> +<h4 id="_changes_of_note"><a class="anchor" href="#_changes_of_note"></a>13.2.1. Changes of Note!</h4> <div class="paragraph"> <p>In here we list important changes that are in 1.0.0+ since 0.98.x., changes you should be aware that will go into effect once you upgrade.</p> </div> @@ -6657,7 +6701,7 @@ using 0.98.11 servers with any other client version.</p> </div> </div> <div class="sect3"> -<h4 id="upgrade1.0.rolling.upgrade"><a class="anchor" href="#upgrade1.0.rolling.upgrade"></a>13.1.2. Rolling upgrade from 0.98.x to HBase 1.0.0</h4> +<h4 id="upgrade1.0.rolling.upgrade"><a class="anchor" href="#upgrade1.0.rolling.upgrade"></a>13.2.2. Rolling upgrade from 0.98.x to HBase 1.0.0</h4> <div class="admonitionblock note"> <table> <tr> @@ -6676,7 +6720,7 @@ You cannot do a <a href="#hbase.rolling.upgrade">rolling upgrade</a> from 0.96.x </div> </div> <div class="sect3"> -<h4 id="upgrade1.0.scanner.caching"><a class="anchor" href="#upgrade1.0.scanner.caching"></a>13.1.3. Scanner Caching has Changed</h4> +<h4 id="upgrade1.0.scanner.caching"><a class="anchor" href="#upgrade1.0.scanner.caching"></a>13.2.3. Scanner Caching has Changed</h4> <div class="paragraph"> <div class="title">From 0.98.x to 1.x</div> <p>In hbase-1.x, the default Scan caching 'number of rows' changed. @@ -6689,14 +6733,14 @@ for further discussion.</p> </div> </div> <div class="sect3"> -<h4 id="upgrade1.0.from.0.94"><a class="anchor" href="#upgrade1.0.from.0.94"></a>13.1.4. Upgrading to 1.0 from 0.94</h4> +<h4 id="upgrade1.0.from.0.94"><a class="anchor" href="#upgrade1.0.from.0.94"></a>13.2.4. Upgrading to 1.0 from 0.94</h4> <div class="paragraph"> <p>You cannot rolling upgrade from 0.94.x to 1.x.x. You must stop your cluster, install the 1.x.x software, run the migration described at <a href="#executing.the.0.96.upgrade">Executing the 0.96 Upgrade</a> (substituting 1.x.x. wherever we make mention of 0.96.x in the section below), and then restart. Be sure to upgrade your ZooKeeper if it is a version less than the required 3.4.x.</p> </div> </div> </div> <div class="sect2"> -<h3 id="upgrade0.98"><a class="anchor" href="#upgrade0.98"></a>13.2. Upgrading from 0.96.x to 0.98.x</h3> +<h3 id="upgrade0.98"><a class="anchor" href="#upgrade0.98"></a>13.3. Upgrading from 0.96.x to 0.98.x</h3> <div class="paragraph"> <p>A rolling upgrade from 0.96.x to 0.98.x works. The two versions are not binary compatible.</p> </div> @@ -6708,15 +6752,15 @@ for further discussion.</p> </div> </div> <div class="sect2"> -<h3 id="_upgrading_from_0_94_x_to_0_98_x"><a class="anchor" href="#_upgrading_from_0_94_x_to_0_98_x"></a>13.3. Upgrading from 0.94.x to 0.98.x</h3> +<h3 id="_upgrading_from_0_94_x_to_0_98_x"><a class="anchor" href="#_upgrading_from_0_94_x_to_0_98_x"></a>13.4. Upgrading from 0.94.x to 0.98.x</h3> <div class="paragraph"> <p>A rolling upgrade from 0.94.x directly to 0.98.x does not work. The upgrade path follows the same procedures as <a href="#upgrade0.96">Upgrading from 0.94.x to 0.96.x</a>. Additional steps are required to use some of the new features of 0.98.x. See <a href="#upgrade0.98">Upgrading from 0.96.x to 0.98.x</a> for an abbreviated list of these features.</p> </div> </div> <div class="sect2"> -<h3 id="upgrade0.96"><a class="anchor" href="#upgrade0.96"></a>13.4. Upgrading from 0.94.x to 0.96.x</h3> +<h3 id="upgrade0.96"><a class="anchor" href="#upgrade0.96"></a>13.5. Upgrading from 0.94.x to 0.96.x</h3> <div class="sect3"> -<h4 id="_the_singularity"><a class="anchor" href="#_the_singularity"></a>13.4.1. The "Singularity"</h4> +<h4 id="_the_singularity"><a class="anchor" href="#_the_singularity"></a>13.5.1. The "Singularity"</h4> <div class="paragraph"> <p>You will have to stop your old 0.94.x cluster completely to upgrade. If you are replicating between clusters, both clusters will have to go down to upgrade. Make sure it is a clean shutdown. The less WAL files around, the faster the upgrade will run (the upgrade will split any log files it finds in the filesystem as part of the upgrade process). All clients must be upgraded to 0.96 too.</p> </div> @@ -6725,7 +6769,7 @@ for further discussion.</p> </div> </div> <div class="sect3"> -<h4 id="executing.the.0.96.upgrade"><a class="anchor" href="#executing.the.0.96.upgrade"></a>13.4.2. Executing the 0.96 Upgrade</h4> +<h4 id="executing.the.0.96.upgrade"><a class="anchor" href="#executing.the.0.96.upgrade"></a>13.5.2. Executing the 0.96 Upgrade</h4> <div class="admonitionblock note"> <table> <tr> @@ -6890,7 +6934,7 @@ Successfully completed Log splitting</pre> </div> </div> <div class="sect2"> -<h3 id="s096.migration.troubleshooting"><a class="anchor" href="#s096.migration.troubleshooting"></a>13.5. Troubleshooting</h3> +<h3 id="s096.migration.troubleshooting"><a class="anchor" href="#s096.migration.troubleshooting"></a>13.6. Troubleshooting</h3> <div id="s096.migration.troubleshooting.old.client" class="paragraph"> <div class="title">Old Client connecting to 0.96 cluster</div> <p>It will fail with an exception like the below. Upgrade.</p> @@ -6912,7 +6956,7 @@ Successfully completed Log splitting</pre> </div> </div> <div class="sect3"> -<h4 id="_upgrading_code_meta_code_to_use_protocol_buffers_protobuf"><a class="anchor" href="#_upgrading_code_meta_code_to_use_protocol_buffers_protobuf"></a>13.5.1. Upgrading <code>META</code> to use Protocol Buffers (Protobuf)</h4> +<h4 id="_upgrading_code_meta_code_to_use_protocol_buffers_protobuf"><a class="anchor" href="#_upgrading_code_meta_code_to_use_protocol_buffers_protobuf"></a>13.6.1. Upgrading <code>META</code> to use Protocol Buffers (Protobuf)</h4> <div class="paragraph"> <p>When you upgrade from versions prior to 0.96, <code>META</code> needs to be converted to use protocol buffers. This is controlled by the configuration option <code>hbase.MetaMigrationConvertingToPB</code>, which is set to <code>true</code> by default. Therefore, by default, no action is required on your part.</p> </div> @@ -6922,15 +6966,15 @@ Successfully completed Log splitting</pre> </div> </div> <div class="sect2"> -<h3 id="upgrade0.94"><a class="anchor" href="#upgrade0.94"></a>13.6. Upgrading from 0.92.x to 0.94.x</h3> +<h3 id="upgrade0.94"><a class="anchor" href="#upgrade0.94"></a>13.7. Upgrading from 0.92.x to 0.94.x</h3> <div class="paragraph"> <p>We used to think that 0.92 and 0.94 were interface compatible and that you can do a rolling upgrade between these versions but then we figured that <a href="https://issues.apache.org/jira/browse/HBASE-5357">HBASE-5357 Use builder pattern in HColumnDescriptor</a> changed method signatures so rather than return <code>void</code> they instead return <code>HColumnDescriptor</code>. This will throw <code>java.lang.NoSuchMethodError: org.apache.hadoop.hbase.HColumnDescriptor.setMaxVersions(I)V</code> so 0.92 and 0.94 are NOT compatible. You cannot do a rolling upgrade between them.</p> </div> </div> <div class="sect2"> -<h3 id="upgrade0.92"><a class="anchor" href="#upgrade0.92"></a>13.7. Upgrading from 0.90.x to 0.92.x</h3> +<h3 id="upgrade0.92"><a class="anchor" href="#upgrade0.92"></a>13.8. Upgrading from 0.90.x to 0.92.x</h3> <div class="sect3"> -<h4 id="_upgrade_guide"><a class="anchor" href="#_upgrade_guide"></a>13.7.1. Upgrade Guide</h4> +<h4 id="_upgrade_guide"><a class="anchor" href="#_upgrade_guide"></a>13.8.1. Upgrade Guide</h4> <div class="paragraph"> <p>You will find that 0.92.0 runs a little differently to 0.90.x releases. Here are a few things to watch out for upgrading from 0.90.x to 0.92.0.</p> </div> @@ -7020,7 +7064,7 @@ Successfully completed Log splitting</pre> </div> </div> <div class="sect2"> -<h3 id="upgrade0.90"><a class="anchor" href="#upgrade0.90"></a>13.8. Upgrading to HBase 0.90.x from 0.20.x or 0.89.x</h3> +<h3 id="upgrade0.90"><a class="anchor" href="#upgrade0.90"></a>13.9. Upgrading to HBase 0.90.x from 0.20.x or 0.89.x</h3> <div class="paragraph"> <p>This version of 0.90.x HBase can be started on data written by HBase 0.20.x or HBase 0.89.x. There is no need of a migration step. HBase 0.89.x and 0.90.x does write out the name of region directories differently — it names them with a md5 hash of the region name rather than a jenkins hash — so this means that once started, there is no going back to HBase 0.20.x.</p> </div> @@ -18754,60 +18798,45 @@ two tables need to be backed up and be recoverable as a group.</p> <p>The following is an outline of the steps and examples of commands that are used to backup the data for the <em>green</em> application and to recover the data later. All commands are run when logged in as HBase superuser.</p> </div> -<div class="olist arabic"> -<ol class="arabic"> +<div class="ulist"> +<ul> <li> <p>A backup set called <em>green_set</em> is created as an alias for both the transactions table and the customer table. The backup set can be used for all operations to avoid typing each table name. The backup set name is case-sensitive and should be formed with only printable characters and without spaces.</p> -</li> -</ol> -</div> -<div class="listingblock"> +<div class="literalblock"> <div class="content"> -<pre class="CodeRay highlight"><code data-lang="java"><span class="error">$</span> hbase backup set add green_set transactions -<span class="error">$</span> hbase backup set add green_set customer</code></pre> +<pre>$ hbase backup set add green_set transactions +$ hbase backup set add green_set customer</pre> </div> </div> -<div class="olist arabic"> -<ol class="arabic"> +</li> <li> <p>The first backup of green_set data must be a full backup. The following command example shows how credentials are passed to Amazon S3 and specifies the file system with the s3a: prefix.</p> -</li> -</ol> -</div> -<div class="listingblock"> +<div class="literalblock"> <div class="content"> -<pre class="CodeRay highlight"><code data-lang="java"><span class="error">$</span> ACCESS_KEY=ABCDEFGHIJKLMNOPQRST -<span class="error">$</span> SECRET_KEY=<span class="integer">123456789</span>abcdefghijklmnopqrstuvwxyzABCD -<span class="error">$</span> sudo -u hbase hbase backup create full\ - s3a:<span class="comment">//$ACCESS_KEY:SECRET_KEY@prodhbasebackups/backups -s green_set</span></code></pre> +<pre>$ ACCESS_KEY=ABCDEFGHIJKLMNOPQRST +$ SECRET_KEY=123456789abcdefghijklmnopqrstuvwxyzABCD +$ sudo -u hbase hbase backup create full\ + s3a://$ACCESS_KEY:SECRET_KEY@prodhbasebackups/backups -s green_set</pre> </div> </div> -<div class="olist arabic"> -<ol class="arabic"> +</li> <li> <p>Incremental backups should be run according to a schedule that ensures essential data recovery in the event of a catastrophe. At this retail company, the HBase admin team decides that automated daily backups secures the data sufficiently. The team decides that they can implement this by modifying an existing Cron job that is defined in <code>/etc/crontab</code>. Consequently, IT modifies the Cron job by adding the following line:</p> -</li> -</ol> -</div> -<div class="listingblock"> +<div class="literalblock"> <div class="content"> -<pre class="CodeRay highlight"><code data-lang="java"><span class="annotation">@daily</span> hbase hbase backup create incremental s3a:<span class="comment">//$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups -s green_set</span></code></pre> +<pre>@daily hbase hbase backup create incremental s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups -s green_set</pre> </div> </div> -<div class="olist arabic"> -<ol class="arabic"> +</li> <li> <p>A catastrophic IT incident disables the production cluster that the green application uses. An HBase system administrator of the backup cluster must restore the <em>green_set</em> dataset to the point in time closest to the recovery objective.</p> -</li> -</ol> -</div> <div class="admonitionblock note"> <table> <tr> @@ -18826,39 +18855,38 @@ and protecting a detailed log of backup IDs outside the production cluster in yo <p>The HBase administrator runs the following command on the directory where backups are stored to print the list of successful backup IDs on the console:</p> </div> -<div class="paragraph"> -<p><code>hdfs dfs -ls -t /prodhbasebackups/backups</code></p> +<div class="literalblock"> +<div class="content"> +<pre>`hdfs dfs -ls -t /prodhbasebackups/backups`</pre> </div> -<div class="olist arabic"> -<ol class="arabic"> +</div> +</li> <li> <p>The admin scans the list to see which backup was created at a date and time closest to the recovery objective. To do this, the admin converts the calendar timestamp of the recovery point in time to Unix time because backup IDs are uniquely identified with Unix time. The backup IDs are listed in reverse chronological order, meaning the most recent successful backup appears first.</p> -</li> -</ol> -</div> <div class="paragraph"> <p>The admin notices that the following line in the command output corresponds with the <em>green_set</em> backup that needs to be restored:</p> </div> -<div class="paragraph"> -<p><code>/prodhbasebackups/backups/backup_1467823988425</code></p> +<div class="literalblock"> +<div class="content"> +<pre>/prodhbasebackups/backups/backup_1467823988425`</pre> </div> -<div class="olist arabic"> -<ol class="arabic"> +</div> +</li> <li> <p>The admin restores green_set invoking the backup ID and the -overwrite option. The -overwrite option truncates all existing data in the destination and populates the tables with data from the backup dataset. Without this flag, the backup data is appended to the existing data in the destination. In this case, the admin decides to overwrite the data because it is corrupted.</p> -</li> -</ol> -</div> -<div class="listingblock"> +<div class="literalblock"> <div class="content"> -<pre class="CodeRay highlight"><code data-lang="java"><span class="error">$</span> sudo -u hbase hbase restore -s green_set \ - s3a:<span class="comment">//$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite</span></code></pre> +<pre>$ sudo -u hbase hbase restore -s green_set \ + s3a://$ACCESS_KEY:$SECRET_KEY@prodhbasebackups/backups backup_1467823988425 \ -overwrite</pre> </div> </div> +</li> +</ul> +</div> </div> </div> <div class="sect1"> @@ -30718,14 +30746,20 @@ Make sure you are pulling from the repository when tests run and that you are no <p>See <a href="https://www.apache.org/dev/publishing-maven-artifacts.html">Publishing Maven Artifacts</a> for some pointers on this maven staging process.</p> </div> <div class="paragraph"> -<p>If the HBase version ends in <code>-SNAPSHOT</code>, the artifacts go elsewhere. +<p>+ +If the HBase version ends in <code>-SNAPSHOT</code>, the artifacts go elsewhere. They are put into the Apache snapshots repository directly and are immediately available. Making a SNAPSHOT release, this is what you want to happen.</p> </div> <div class="paragraph"> -<p>At this stage, you have two tarballs in your 'build output directory' and a set of artifacts in a staging area of the maven repository, in the 'closed' state. +<p>+ +At this stage, you have two tarballs in your 'build output directory' and a set of artifacts +in a staging area of the maven repository, in the 'closed' state. Next sign, fingerprint and then 'stage' your release candiate build output directory via svnpubsub by committing -your directory to <a href="https://dist.apache.org/repos/dist/dev/hbase/">The 'dev' distribution directory</a> (See comments on <a href="https://issues.apache.org/jira/browse/HBASE-10554">HBASE-10554 Please delete old releases from mirroring system</a> but in essence it is an svn checkout of <a href="https://dist.apache.org/repos/dist/dev/hbase — releases" class="bare">https://dist.apache.org/repos/dist/dev/hbase — releases</a> are at <a href="https://dist.apache.org/repos/dist/release/hbase" class="bare">https://dist.apache.org/repos/dist/release/hbase</a>). In the <em>version directory</em> run the following commands:</p> +your directory to <a href="https://dist.apache.org/repos/dist/dev/hbase/">The dev distribution directory</a> +(See comments on <a href="https://issues.apache.org/jira/browse/HBASE-10554">HBASE-10554 Please delete old releases from mirroring system</a> +but in essence it is an svn checkout of <a href="https://dist.apache.org/repos/dist/dev/hbase">dev/hbase</a> — releases are at +<a href="https://dist.apache.org/repos/dist/release/hbase">release/hbase</a>). In the <em>version directory</em> run the following commands:</p> </div> <div class="listingblock"> <div class="content"> @@ -37300,7 +37334,7 @@ The server will return cellblocks compressed using this same compressor as long <div id="footer"> <div id="footer-text"> Version 3.0.0-SNAPSHOT<br> -Last updated 2018-03-22 14:29:40 UTC +Last updated 2018-03-23 14:29:57 UTC </div> </div> </body> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/bulk-loads.html ---------------------------------------------------------------------- diff --git a/bulk-loads.html b/bulk-loads.html index 2a63526..f6656ad 100644 --- a/bulk-loads.html +++ b/bulk-loads.html @@ -7,7 +7,7 @@ <head> <meta charset="UTF-8" /> <meta name="viewport" content="width=device-width, initial-scale=1.0" /> - <meta name="Date-Revision-yyyymmdd" content="20180322" /> + <meta name="Date-Revision-yyyymmdd" content="20180323" /> <meta http-equiv="Content-Language" content="en" /> <title>Apache HBase – Bulk Loads in Apache HBase (TM) @@ -299,7 +299,7 @@ under the License. --> <a href="https://www.apache.org/">The Apache Software Foundation</a>. All rights reserved. - <li id="publishDate" class="pull-right">Last Published: 2018-03-22</li> + <li id="publishDate" class="pull-right">Last Published: 2018-03-23</li> </p> </div>
