http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/MultiTableHFileOutputFormat.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/MultiTableHFileOutputFormat.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/MultiTableHFileOutputFormat.html index 65a78e7..664869e 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/MultiTableHFileOutputFormat.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/MultiTableHFileOutputFormat.html @@ -29,106 +29,105 @@ <span class="sourceLineNo">021</span>import com.google.common.annotations.VisibleForTesting;<a name="line.21"></a> <span class="sourceLineNo">022</span>import org.apache.commons.logging.Log;<a name="line.22"></a> <span class="sourceLineNo">023</span>import org.apache.commons.logging.LogFactory;<a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.hadoop.hbase.HTableDescriptor;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.hadoop.hbase.classification.InterfaceAudience;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.hadoop.hbase.client.RegionLocator;<a name="line.26"></a> -<span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.27"></a> -<span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.28"></a> -<span class="sourceLineNo">029</span>import org.apache.hadoop.mapreduce.Job;<a name="line.29"></a> -<span class="sourceLineNo">030</span><a name="line.30"></a> -<span class="sourceLineNo">031</span>import java.io.IOException;<a name="line.31"></a> -<span class="sourceLineNo">032</span>import java.nio.charset.Charset;<a name="line.32"></a> -<span class="sourceLineNo">033</span>import java.util.List;<a name="line.33"></a> -<span class="sourceLineNo">034</span><a name="line.34"></a> -<span class="sourceLineNo">035</span>/**<a name="line.35"></a> -<span class="sourceLineNo">036</span> * Create 3 level tree directory, first level is using table name as parent<a name="line.36"></a> -<span class="sourceLineNo">037</span> * directory and then use family name as child directory, and all related HFiles<a name="line.37"></a> -<span class="sourceLineNo">038</span> * for one family are under child directory<a name="line.38"></a> -<span class="sourceLineNo">039</span> * -tableName1<a name="line.39"></a> -<span class="sourceLineNo">040</span> * -columnFamilyName1<a name="line.40"></a> -<span class="sourceLineNo">041</span> * -columnFamilyName2<a name="line.41"></a> -<span class="sourceLineNo">042</span> * -HFiles<a name="line.42"></a> -<span class="sourceLineNo">043</span> * -tableName2<a name="line.43"></a> -<span class="sourceLineNo">044</span> * -columnFamilyName1<a name="line.44"></a> -<span class="sourceLineNo">045</span> * -HFiles<a name="line.45"></a> -<span class="sourceLineNo">046</span> * -columnFamilyName2<a name="line.46"></a> -<span class="sourceLineNo">047</span> */<a name="line.47"></a> -<span class="sourceLineNo">048</span>@InterfaceAudience.Public<a name="line.48"></a> -<span class="sourceLineNo">049</span>@VisibleForTesting<a name="line.49"></a> -<span class="sourceLineNo">050</span>public class MultiTableHFileOutputFormat extends HFileOutputFormat2 {<a name="line.50"></a> -<span class="sourceLineNo">051</span> private static final Log LOG = LogFactory.getLog(MultiTableHFileOutputFormat.class);<a name="line.51"></a> -<span class="sourceLineNo">052</span><a name="line.52"></a> -<span class="sourceLineNo">053</span> /**<a name="line.53"></a> -<span class="sourceLineNo">054</span> * Creates a composite key to use as a mapper output key when using<a name="line.54"></a> -<span class="sourceLineNo">055</span> * MultiTableHFileOutputFormat.configureIncrementaLoad to set up bulk ingest job<a name="line.55"></a> -<span class="sourceLineNo">056</span> *<a name="line.56"></a> -<span class="sourceLineNo">057</span> * @param tableName Name of the Table - Eg: TableName.getNameAsString()<a name="line.57"></a> -<span class="sourceLineNo">058</span> * @param suffix Usually represents a rowkey when creating a mapper key or column family<a name="line.58"></a> -<span class="sourceLineNo">059</span> * @return byte[] representation of composite key<a name="line.59"></a> -<span class="sourceLineNo">060</span> */<a name="line.60"></a> -<span class="sourceLineNo">061</span> public static byte[] createCompositeKey(byte[] tableName,<a name="line.61"></a> -<span class="sourceLineNo">062</span> byte[] suffix) {<a name="line.62"></a> -<span class="sourceLineNo">063</span> return combineTableNameSuffix(tableName, suffix);<a name="line.63"></a> -<span class="sourceLineNo">064</span> }<a name="line.64"></a> -<span class="sourceLineNo">065</span><a name="line.65"></a> -<span class="sourceLineNo">066</span> /**<a name="line.66"></a> -<span class="sourceLineNo">067</span> * Alternate api which accepts an ImmutableBytesWritable for the suffix<a name="line.67"></a> -<span class="sourceLineNo">068</span> * @see MultiTableHFileOutputFormat#createCompositeKey(byte[], byte[])<a name="line.68"></a> -<span class="sourceLineNo">069</span> */<a name="line.69"></a> -<span class="sourceLineNo">070</span> public static byte[] createCompositeKey(byte[] tableName,<a name="line.70"></a> -<span class="sourceLineNo">071</span> ImmutableBytesWritable suffix) {<a name="line.71"></a> -<span class="sourceLineNo">072</span> return combineTableNameSuffix(tableName, suffix.get());<a name="line.72"></a> -<span class="sourceLineNo">073</span> }<a name="line.73"></a> -<span class="sourceLineNo">074</span><a name="line.74"></a> -<span class="sourceLineNo">075</span> /**<a name="line.75"></a> -<span class="sourceLineNo">076</span> * Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the<a name="line.76"></a> -<span class="sourceLineNo">077</span> * suffix<a name="line.77"></a> -<span class="sourceLineNo">078</span> * @see MultiTableHFileOutputFormat#createCompositeKey(byte[], byte[])<a name="line.78"></a> -<span class="sourceLineNo">079</span> */<a name="line.79"></a> -<span class="sourceLineNo">080</span> public static byte[] createCompositeKey(String tableName,<a name="line.80"></a> -<span class="sourceLineNo">081</span> ImmutableBytesWritable suffix) {<a name="line.81"></a> -<span class="sourceLineNo">082</span> return combineTableNameSuffix(tableName.getBytes(Charset.forName("UTF-8")), suffix.get());<a name="line.82"></a> -<span class="sourceLineNo">083</span> }<a name="line.83"></a> -<span class="sourceLineNo">084</span><a name="line.84"></a> -<span class="sourceLineNo">085</span> /**<a name="line.85"></a> -<span class="sourceLineNo">086</span> * Analogous to<a name="line.86"></a> -<span class="sourceLineNo">087</span> * {@link HFileOutputFormat2#configureIncrementalLoad(Job, HTableDescriptor, RegionLocator)},<a name="line.87"></a> -<span class="sourceLineNo">088</span> * this function will configure the requisite number of reducers to write HFiles for multple<a name="line.88"></a> -<span class="sourceLineNo">089</span> * tables simultaneously<a name="line.89"></a> -<span class="sourceLineNo">090</span> *<a name="line.90"></a> -<span class="sourceLineNo">091</span> * @param job See {@link org.apache.hadoop.mapreduce.Job}<a name="line.91"></a> -<span class="sourceLineNo">092</span> * @param multiTableDescriptors Table descriptor and region locator pairs<a name="line.92"></a> -<span class="sourceLineNo">093</span> * @throws IOException<a name="line.93"></a> -<span class="sourceLineNo">094</span> */<a name="line.94"></a> -<span class="sourceLineNo">095</span> public static void configureIncrementalLoad(Job job, List<TableInfo><a name="line.95"></a> -<span class="sourceLineNo">096</span> multiTableDescriptors)<a name="line.96"></a> -<span class="sourceLineNo">097</span> throws IOException {<a name="line.97"></a> -<span class="sourceLineNo">098</span> MultiTableHFileOutputFormat.configureIncrementalLoad(job, multiTableDescriptors,<a name="line.98"></a> -<span class="sourceLineNo">099</span> MultiTableHFileOutputFormat.class);<a name="line.99"></a> -<span class="sourceLineNo">100</span> }<a name="line.100"></a> -<span class="sourceLineNo">101</span><a name="line.101"></a> -<span class="sourceLineNo">102</span> final private static int validateCompositeKey(byte[] keyBytes) {<a name="line.102"></a> -<span class="sourceLineNo">103</span><a name="line.103"></a> -<span class="sourceLineNo">104</span> int separatorIdx = Bytes.indexOf(keyBytes, HFileOutputFormat2.tableSeparator);<a name="line.104"></a> -<span class="sourceLineNo">105</span><a name="line.105"></a> -<span class="sourceLineNo">106</span> // Either the separator was not found or a tablename wasn't present or a key wasn't present<a name="line.106"></a> -<span class="sourceLineNo">107</span> if (separatorIdx == -1) {<a name="line.107"></a> -<span class="sourceLineNo">108</span> throw new IllegalArgumentException("Invalid format for composite key [" + Bytes<a name="line.108"></a> -<span class="sourceLineNo">109</span> .toStringBinary(keyBytes) + "]. Cannot extract tablename and suffix from key");<a name="line.109"></a> -<span class="sourceLineNo">110</span> }<a name="line.110"></a> -<span class="sourceLineNo">111</span> return separatorIdx;<a name="line.111"></a> -<span class="sourceLineNo">112</span> }<a name="line.112"></a> -<span class="sourceLineNo">113</span><a name="line.113"></a> -<span class="sourceLineNo">114</span> protected static byte[] getTableName(byte[] keyBytes) {<a name="line.114"></a> -<span class="sourceLineNo">115</span> int separatorIdx = validateCompositeKey(keyBytes);<a name="line.115"></a> -<span class="sourceLineNo">116</span> return Bytes.copy(keyBytes, 0, separatorIdx);<a name="line.116"></a> -<span class="sourceLineNo">117</span> }<a name="line.117"></a> -<span class="sourceLineNo">118</span><a name="line.118"></a> -<span class="sourceLineNo">119</span> protected static byte[] getSuffix(byte[] keyBytes) {<a name="line.119"></a> -<span class="sourceLineNo">120</span> int separatorIdx = validateCompositeKey(keyBytes);<a name="line.120"></a> -<span class="sourceLineNo">121</span> return Bytes.copy(keyBytes, separatorIdx+1, keyBytes.length - separatorIdx - 1);<a name="line.121"></a> -<span class="sourceLineNo">122</span> }<a name="line.122"></a> -<span class="sourceLineNo">123</span>}<a name="line.123"></a> +<span class="sourceLineNo">024</span>import org.apache.hadoop.hbase.classification.InterfaceAudience;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.apache.hadoop.hbase.client.TableDescriptor;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.26"></a> +<span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.27"></a> +<span class="sourceLineNo">028</span>import org.apache.hadoop.mapreduce.Job;<a name="line.28"></a> +<span class="sourceLineNo">029</span><a name="line.29"></a> +<span class="sourceLineNo">030</span>import java.io.IOException;<a name="line.30"></a> +<span class="sourceLineNo">031</span>import java.nio.charset.Charset;<a name="line.31"></a> +<span class="sourceLineNo">032</span>import java.util.List;<a name="line.32"></a> +<span class="sourceLineNo">033</span><a name="line.33"></a> +<span class="sourceLineNo">034</span>/**<a name="line.34"></a> +<span class="sourceLineNo">035</span> * Create 3 level tree directory, first level is using table name as parent<a name="line.35"></a> +<span class="sourceLineNo">036</span> * directory and then use family name as child directory, and all related HFiles<a name="line.36"></a> +<span class="sourceLineNo">037</span> * for one family are under child directory<a name="line.37"></a> +<span class="sourceLineNo">038</span> * -tableName1<a name="line.38"></a> +<span class="sourceLineNo">039</span> * -columnFamilyName1<a name="line.39"></a> +<span class="sourceLineNo">040</span> * -columnFamilyName2<a name="line.40"></a> +<span class="sourceLineNo">041</span> * -HFiles<a name="line.41"></a> +<span class="sourceLineNo">042</span> * -tableName2<a name="line.42"></a> +<span class="sourceLineNo">043</span> * -columnFamilyName1<a name="line.43"></a> +<span class="sourceLineNo">044</span> * -HFiles<a name="line.44"></a> +<span class="sourceLineNo">045</span> * -columnFamilyName2<a name="line.45"></a> +<span class="sourceLineNo">046</span> */<a name="line.46"></a> +<span class="sourceLineNo">047</span>@InterfaceAudience.Public<a name="line.47"></a> +<span class="sourceLineNo">048</span>@VisibleForTesting<a name="line.48"></a> +<span class="sourceLineNo">049</span>public class MultiTableHFileOutputFormat extends HFileOutputFormat2 {<a name="line.49"></a> +<span class="sourceLineNo">050</span> private static final Log LOG = LogFactory.getLog(MultiTableHFileOutputFormat.class);<a name="line.50"></a> +<span class="sourceLineNo">051</span><a name="line.51"></a> +<span class="sourceLineNo">052</span> /**<a name="line.52"></a> +<span class="sourceLineNo">053</span> * Creates a composite key to use as a mapper output key when using<a name="line.53"></a> +<span class="sourceLineNo">054</span> * MultiTableHFileOutputFormat.configureIncrementaLoad to set up bulk ingest job<a name="line.54"></a> +<span class="sourceLineNo">055</span> *<a name="line.55"></a> +<span class="sourceLineNo">056</span> * @param tableName Name of the Table - Eg: TableName.getNameAsString()<a name="line.56"></a> +<span class="sourceLineNo">057</span> * @param suffix Usually represents a rowkey when creating a mapper key or column family<a name="line.57"></a> +<span class="sourceLineNo">058</span> * @return byte[] representation of composite key<a name="line.58"></a> +<span class="sourceLineNo">059</span> */<a name="line.59"></a> +<span class="sourceLineNo">060</span> public static byte[] createCompositeKey(byte[] tableName,<a name="line.60"></a> +<span class="sourceLineNo">061</span> byte[] suffix) {<a name="line.61"></a> +<span class="sourceLineNo">062</span> return combineTableNameSuffix(tableName, suffix);<a name="line.62"></a> +<span class="sourceLineNo">063</span> }<a name="line.63"></a> +<span class="sourceLineNo">064</span><a name="line.64"></a> +<span class="sourceLineNo">065</span> /**<a name="line.65"></a> +<span class="sourceLineNo">066</span> * Alternate api which accepts an ImmutableBytesWritable for the suffix<a name="line.66"></a> +<span class="sourceLineNo">067</span> * @see MultiTableHFileOutputFormat#createCompositeKey(byte[], byte[])<a name="line.67"></a> +<span class="sourceLineNo">068</span> */<a name="line.68"></a> +<span class="sourceLineNo">069</span> public static byte[] createCompositeKey(byte[] tableName,<a name="line.69"></a> +<span class="sourceLineNo">070</span> ImmutableBytesWritable suffix) {<a name="line.70"></a> +<span class="sourceLineNo">071</span> return combineTableNameSuffix(tableName, suffix.get());<a name="line.71"></a> +<span class="sourceLineNo">072</span> }<a name="line.72"></a> +<span class="sourceLineNo">073</span><a name="line.73"></a> +<span class="sourceLineNo">074</span> /**<a name="line.74"></a> +<span class="sourceLineNo">075</span> * Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the<a name="line.75"></a> +<span class="sourceLineNo">076</span> * suffix<a name="line.76"></a> +<span class="sourceLineNo">077</span> * @see MultiTableHFileOutputFormat#createCompositeKey(byte[], byte[])<a name="line.77"></a> +<span class="sourceLineNo">078</span> */<a name="line.78"></a> +<span class="sourceLineNo">079</span> public static byte[] createCompositeKey(String tableName,<a name="line.79"></a> +<span class="sourceLineNo">080</span> ImmutableBytesWritable suffix) {<a name="line.80"></a> +<span class="sourceLineNo">081</span> return combineTableNameSuffix(tableName.getBytes(Charset.forName("UTF-8")), suffix.get());<a name="line.81"></a> +<span class="sourceLineNo">082</span> }<a name="line.82"></a> +<span class="sourceLineNo">083</span><a name="line.83"></a> +<span class="sourceLineNo">084</span> /**<a name="line.84"></a> +<span class="sourceLineNo">085</span> * Analogous to<a name="line.85"></a> +<span class="sourceLineNo">086</span> * {@link HFileOutputFormat2#configureIncrementalLoad(Job, TableDescriptor, RegionLocator)},<a name="line.86"></a> +<span class="sourceLineNo">087</span> * this function will configure the requisite number of reducers to write HFiles for multple<a name="line.87"></a> +<span class="sourceLineNo">088</span> * tables simultaneously<a name="line.88"></a> +<span class="sourceLineNo">089</span> *<a name="line.89"></a> +<span class="sourceLineNo">090</span> * @param job See {@link org.apache.hadoop.mapreduce.Job}<a name="line.90"></a> +<span class="sourceLineNo">091</span> * @param multiTableDescriptors Table descriptor and region locator pairs<a name="line.91"></a> +<span class="sourceLineNo">092</span> * @throws IOException<a name="line.92"></a> +<span class="sourceLineNo">093</span> */<a name="line.93"></a> +<span class="sourceLineNo">094</span> public static void configureIncrementalLoad(Job job, List<TableInfo><a name="line.94"></a> +<span class="sourceLineNo">095</span> multiTableDescriptors)<a name="line.95"></a> +<span class="sourceLineNo">096</span> throws IOException {<a name="line.96"></a> +<span class="sourceLineNo">097</span> MultiTableHFileOutputFormat.configureIncrementalLoad(job, multiTableDescriptors,<a name="line.97"></a> +<span class="sourceLineNo">098</span> MultiTableHFileOutputFormat.class);<a name="line.98"></a> +<span class="sourceLineNo">099</span> }<a name="line.99"></a> +<span class="sourceLineNo">100</span><a name="line.100"></a> +<span class="sourceLineNo">101</span> final private static int validateCompositeKey(byte[] keyBytes) {<a name="line.101"></a> +<span class="sourceLineNo">102</span><a name="line.102"></a> +<span class="sourceLineNo">103</span> int separatorIdx = Bytes.indexOf(keyBytes, HFileOutputFormat2.tableSeparator);<a name="line.103"></a> +<span class="sourceLineNo">104</span><a name="line.104"></a> +<span class="sourceLineNo">105</span> // Either the separator was not found or a tablename wasn't present or a key wasn't present<a name="line.105"></a> +<span class="sourceLineNo">106</span> if (separatorIdx == -1) {<a name="line.106"></a> +<span class="sourceLineNo">107</span> throw new IllegalArgumentException("Invalid format for composite key [" + Bytes<a name="line.107"></a> +<span class="sourceLineNo">108</span> .toStringBinary(keyBytes) + "]. Cannot extract tablename and suffix from key");<a name="line.108"></a> +<span class="sourceLineNo">109</span> }<a name="line.109"></a> +<span class="sourceLineNo">110</span> return separatorIdx;<a name="line.110"></a> +<span class="sourceLineNo">111</span> }<a name="line.111"></a> +<span class="sourceLineNo">112</span><a name="line.112"></a> +<span class="sourceLineNo">113</span> protected static byte[] getTableName(byte[] keyBytes) {<a name="line.113"></a> +<span class="sourceLineNo">114</span> int separatorIdx = validateCompositeKey(keyBytes);<a name="line.114"></a> +<span class="sourceLineNo">115</span> return Bytes.copy(keyBytes, 0, separatorIdx);<a name="line.115"></a> +<span class="sourceLineNo">116</span> }<a name="line.116"></a> +<span class="sourceLineNo">117</span><a name="line.117"></a> +<span class="sourceLineNo">118</span> protected static byte[] getSuffix(byte[] keyBytes) {<a name="line.118"></a> +<span class="sourceLineNo">119</span> int separatorIdx = validateCompositeKey(keyBytes);<a name="line.119"></a> +<span class="sourceLineNo">120</span> return Bytes.copy(keyBytes, separatorIdx+1, keyBytes.length - separatorIdx - 1);<a name="line.120"></a> +<span class="sourceLineNo">121</span> }<a name="line.121"></a> +<span class="sourceLineNo">122</span>}<a name="line.122"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html index f625c0b..1eae269 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html @@ -315,7 +315,7 @@ <span class="sourceLineNo">307</span> try (Connection conn = ConnectionFactory.createConnection(conf);<a name="line.307"></a> <span class="sourceLineNo">308</span> Table table = conn.getTable(tableName);<a name="line.308"></a> <span class="sourceLineNo">309</span> RegionLocator regionLocator = conn.getRegionLocator(tableName)) {<a name="line.309"></a> -<span class="sourceLineNo">310</span> HFileOutputFormat2.configureIncrementalLoad(job, table.getTableDescriptor(), regionLocator);<a name="line.310"></a> +<span class="sourceLineNo">310</span> HFileOutputFormat2.configureIncrementalLoad(job, table.getDescriptor(), regionLocator);<a name="line.310"></a> <span class="sourceLineNo">311</span> }<a name="line.311"></a> <span class="sourceLineNo">312</span> TableMapReduceUtil.addDependencyJarsForClasses(job.getConfiguration(),<a name="line.312"></a> <span class="sourceLineNo">313</span> com.google.common.base.Preconditions.class);<a name="line.313"></a>