http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapred/HRegionPartitioner.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapred/HRegionPartitioner.html b/apidocs/src-html/org/apache/hadoop/hbase/mapred/HRegionPartitioner.html index 08be3b2..77665c7 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapred/HRegionPartitioner.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapred/HRegionPartitioner.html @@ -27,80 +27,81 @@ <span class="sourceLineNo">019</span>package org.apache.hadoop.hbase.mapred;<a name="line.19"></a> <span class="sourceLineNo">020</span><a name="line.20"></a> <span class="sourceLineNo">021</span>import java.io.IOException;<a name="line.21"></a> -<span class="sourceLineNo">022</span><a name="line.22"></a> -<span class="sourceLineNo">023</span>import org.apache.commons.logging.Log;<a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.LogFactory;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.hadoop.hbase.HBaseConfiguration;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.hadoop.hbase.TableName;<a name="line.26"></a> -<span class="sourceLineNo">027</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.27"></a> -<span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.28"></a> -<span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.client.ConnectionFactory;<a name="line.29"></a> -<span class="sourceLineNo">030</span>import org.apache.hadoop.hbase.client.RegionLocator;<a name="line.30"></a> -<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.31"></a> -<span class="sourceLineNo">032</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.32"></a> -<span class="sourceLineNo">033</span>import org.apache.hadoop.mapred.JobConf;<a name="line.33"></a> -<span class="sourceLineNo">034</span>import org.apache.hadoop.mapred.Partitioner;<a name="line.34"></a> -<span class="sourceLineNo">035</span><a name="line.35"></a> -<span class="sourceLineNo">036</span>/**<a name="line.36"></a> -<span class="sourceLineNo">037</span> * This is used to partition the output keys into groups of keys.<a name="line.37"></a> -<span class="sourceLineNo">038</span> * Keys are grouped according to the regions that currently exist<a name="line.38"></a> -<span class="sourceLineNo">039</span> * so that each reducer fills a single region so load is distributed.<a name="line.39"></a> -<span class="sourceLineNo">040</span> *<a name="line.40"></a> -<span class="sourceLineNo">041</span> * @param <K2><a name="line.41"></a> -<span class="sourceLineNo">042</span> * @param <V2><a name="line.42"></a> -<span class="sourceLineNo">043</span> */<a name="line.43"></a> -<span class="sourceLineNo">044</span>@InterfaceAudience.Public<a name="line.44"></a> -<span class="sourceLineNo">045</span>public class HRegionPartitioner<K2,V2><a name="line.45"></a> -<span class="sourceLineNo">046</span>implements Partitioner<ImmutableBytesWritable, V2> {<a name="line.46"></a> -<span class="sourceLineNo">047</span> private static final Log LOG = LogFactory.getLog(HRegionPartitioner.class);<a name="line.47"></a> -<span class="sourceLineNo">048</span> // Connection and locator are not cleaned up; they just die when partitioner is done.<a name="line.48"></a> -<span class="sourceLineNo">049</span> private Connection connection;<a name="line.49"></a> -<span class="sourceLineNo">050</span> private RegionLocator locator;<a name="line.50"></a> -<span class="sourceLineNo">051</span> private byte[][] startKeys;<a name="line.51"></a> -<span class="sourceLineNo">052</span><a name="line.52"></a> +<span class="sourceLineNo">022</span>import org.apache.hadoop.hbase.HBaseConfiguration;<a name="line.22"></a> +<span class="sourceLineNo">023</span>import org.apache.hadoop.hbase.TableName;<a name="line.23"></a> +<span class="sourceLineNo">024</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.slf4j.Logger;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.slf4j.LoggerFactory;<a name="line.26"></a> +<span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.27"></a> +<span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.client.ConnectionFactory;<a name="line.28"></a> +<span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.client.RegionLocator;<a name="line.29"></a> +<span class="sourceLineNo">030</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.30"></a> +<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.31"></a> +<span class="sourceLineNo">032</span>import org.apache.hadoop.mapred.JobConf;<a name="line.32"></a> +<span class="sourceLineNo">033</span>import org.apache.hadoop.mapred.Partitioner;<a name="line.33"></a> +<span class="sourceLineNo">034</span><a name="line.34"></a> +<span class="sourceLineNo">035</span>/**<a name="line.35"></a> +<span class="sourceLineNo">036</span> * This is used to partition the output keys into groups of keys.<a name="line.36"></a> +<span class="sourceLineNo">037</span> * Keys are grouped according to the regions that currently exist<a name="line.37"></a> +<span class="sourceLineNo">038</span> * so that each reducer fills a single region so load is distributed.<a name="line.38"></a> +<span class="sourceLineNo">039</span> *<a name="line.39"></a> +<span class="sourceLineNo">040</span> * @param <K2><a name="line.40"></a> +<span class="sourceLineNo">041</span> * @param <V2><a name="line.41"></a> +<span class="sourceLineNo">042</span> */<a name="line.42"></a> +<span class="sourceLineNo">043</span>@InterfaceAudience.Public<a name="line.43"></a> +<span class="sourceLineNo">044</span>public class HRegionPartitioner<K2,V2><a name="line.44"></a> +<span class="sourceLineNo">045</span>implements Partitioner<ImmutableBytesWritable, V2> {<a name="line.45"></a> +<span class="sourceLineNo">046</span> private static final Logger LOG = LoggerFactory.getLogger(HRegionPartitioner.class);<a name="line.46"></a> +<span class="sourceLineNo">047</span> // Connection and locator are not cleaned up; they just die when partitioner is done.<a name="line.47"></a> +<span class="sourceLineNo">048</span> private Connection connection;<a name="line.48"></a> +<span class="sourceLineNo">049</span> private RegionLocator locator;<a name="line.49"></a> +<span class="sourceLineNo">050</span> private byte[][] startKeys;<a name="line.50"></a> +<span class="sourceLineNo">051</span><a name="line.51"></a> +<span class="sourceLineNo">052</span> @Override<a name="line.52"></a> <span class="sourceLineNo">053</span> public void configure(JobConf job) {<a name="line.53"></a> <span class="sourceLineNo">054</span> try {<a name="line.54"></a> <span class="sourceLineNo">055</span> this.connection = ConnectionFactory.createConnection(HBaseConfiguration.create(job));<a name="line.55"></a> <span class="sourceLineNo">056</span> TableName tableName = TableName.valueOf(job.get(TableOutputFormat.OUTPUT_TABLE));<a name="line.56"></a> <span class="sourceLineNo">057</span> this.locator = this.connection.getRegionLocator(tableName);<a name="line.57"></a> <span class="sourceLineNo">058</span> } catch (IOException e) {<a name="line.58"></a> -<span class="sourceLineNo">059</span> LOG.error(e);<a name="line.59"></a> +<span class="sourceLineNo">059</span> LOG.error(e.toString(), e);<a name="line.59"></a> <span class="sourceLineNo">060</span> }<a name="line.60"></a> <span class="sourceLineNo">061</span><a name="line.61"></a> <span class="sourceLineNo">062</span> try {<a name="line.62"></a> <span class="sourceLineNo">063</span> this.startKeys = this.locator.getStartKeys();<a name="line.63"></a> <span class="sourceLineNo">064</span> } catch (IOException e) {<a name="line.64"></a> -<span class="sourceLineNo">065</span> LOG.error(e);<a name="line.65"></a> +<span class="sourceLineNo">065</span> LOG.error(e.toString(), e);<a name="line.65"></a> <span class="sourceLineNo">066</span> }<a name="line.66"></a> <span class="sourceLineNo">067</span> }<a name="line.67"></a> <span class="sourceLineNo">068</span><a name="line.68"></a> -<span class="sourceLineNo">069</span> public int getPartition(ImmutableBytesWritable key, V2 value, int numPartitions) {<a name="line.69"></a> -<span class="sourceLineNo">070</span> byte[] region = null;<a name="line.70"></a> -<span class="sourceLineNo">071</span> // Only one region return 0<a name="line.71"></a> -<span class="sourceLineNo">072</span> if (this.startKeys.length == 1){<a name="line.72"></a> -<span class="sourceLineNo">073</span> return 0;<a name="line.73"></a> -<span class="sourceLineNo">074</span> }<a name="line.74"></a> -<span class="sourceLineNo">075</span> try {<a name="line.75"></a> -<span class="sourceLineNo">076</span> // Not sure if this is cached after a split so we could have problems<a name="line.76"></a> -<span class="sourceLineNo">077</span> // here if a region splits while mapping<a name="line.77"></a> -<span class="sourceLineNo">078</span> region = locator.getRegionLocation(key.get()).getRegionInfo().getStartKey();<a name="line.78"></a> -<span class="sourceLineNo">079</span> } catch (IOException e) {<a name="line.79"></a> -<span class="sourceLineNo">080</span> LOG.error(e);<a name="line.80"></a> -<span class="sourceLineNo">081</span> }<a name="line.81"></a> -<span class="sourceLineNo">082</span> for (int i = 0; i < this.startKeys.length; i++){<a name="line.82"></a> -<span class="sourceLineNo">083</span> if (Bytes.compareTo(region, this.startKeys[i]) == 0 ){<a name="line.83"></a> -<span class="sourceLineNo">084</span> if (i >= numPartitions-1){<a name="line.84"></a> -<span class="sourceLineNo">085</span> // cover if we have less reduces then regions.<a name="line.85"></a> -<span class="sourceLineNo">086</span> return (Integer.toString(i).hashCode()<a name="line.86"></a> -<span class="sourceLineNo">087</span> & Integer.MAX_VALUE) % numPartitions;<a name="line.87"></a> -<span class="sourceLineNo">088</span> }<a name="line.88"></a> -<span class="sourceLineNo">089</span> return i;<a name="line.89"></a> -<span class="sourceLineNo">090</span> }<a name="line.90"></a> -<span class="sourceLineNo">091</span> }<a name="line.91"></a> -<span class="sourceLineNo">092</span> // if above fails to find start key that match we need to return something<a name="line.92"></a> -<span class="sourceLineNo">093</span> return 0;<a name="line.93"></a> -<span class="sourceLineNo">094</span> }<a name="line.94"></a> -<span class="sourceLineNo">095</span>}<a name="line.95"></a> +<span class="sourceLineNo">069</span> @Override<a name="line.69"></a> +<span class="sourceLineNo">070</span> public int getPartition(ImmutableBytesWritable key, V2 value, int numPartitions) {<a name="line.70"></a> +<span class="sourceLineNo">071</span> byte[] region = null;<a name="line.71"></a> +<span class="sourceLineNo">072</span> // Only one region return 0<a name="line.72"></a> +<span class="sourceLineNo">073</span> if (this.startKeys.length == 1){<a name="line.73"></a> +<span class="sourceLineNo">074</span> return 0;<a name="line.74"></a> +<span class="sourceLineNo">075</span> }<a name="line.75"></a> +<span class="sourceLineNo">076</span> try {<a name="line.76"></a> +<span class="sourceLineNo">077</span> // Not sure if this is cached after a split so we could have problems<a name="line.77"></a> +<span class="sourceLineNo">078</span> // here if a region splits while mapping<a name="line.78"></a> +<span class="sourceLineNo">079</span> region = locator.getRegionLocation(key.get()).getRegionInfo().getStartKey();<a name="line.79"></a> +<span class="sourceLineNo">080</span> } catch (IOException e) {<a name="line.80"></a> +<span class="sourceLineNo">081</span> LOG.error(e.toString(), e);<a name="line.81"></a> +<span class="sourceLineNo">082</span> }<a name="line.82"></a> +<span class="sourceLineNo">083</span> for (int i = 0; i < this.startKeys.length; i++){<a name="line.83"></a> +<span class="sourceLineNo">084</span> if (Bytes.compareTo(region, this.startKeys[i]) == 0 ){<a name="line.84"></a> +<span class="sourceLineNo">085</span> if (i >= numPartitions-1){<a name="line.85"></a> +<span class="sourceLineNo">086</span> // cover if we have less reduces then regions.<a name="line.86"></a> +<span class="sourceLineNo">087</span> return (Integer.toString(i).hashCode()<a name="line.87"></a> +<span class="sourceLineNo">088</span> & Integer.MAX_VALUE) % numPartitions;<a name="line.88"></a> +<span class="sourceLineNo">089</span> }<a name="line.89"></a> +<span class="sourceLineNo">090</span> return i;<a name="line.90"></a> +<span class="sourceLineNo">091</span> }<a name="line.91"></a> +<span class="sourceLineNo">092</span> }<a name="line.92"></a> +<span class="sourceLineNo">093</span> // if above fails to find start key that match we need to return something<a name="line.93"></a> +<span class="sourceLineNo">094</span> return 0;<a name="line.94"></a> +<span class="sourceLineNo">095</span> }<a name="line.95"></a> +<span class="sourceLineNo">096</span>}<a name="line.96"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapred/IdentityTableReduce.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapred/IdentityTableReduce.html b/apidocs/src-html/org/apache/hadoop/hbase/mapred/IdentityTableReduce.html index 045de2b..88b2605 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapred/IdentityTableReduce.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapred/IdentityTableReduce.html @@ -29,9 +29,9 @@ <span class="sourceLineNo">021</span>import java.io.IOException;<a name="line.21"></a> <span class="sourceLineNo">022</span>import java.util.Iterator;<a name="line.22"></a> <span class="sourceLineNo">023</span><a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.Log;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.commons.logging.LogFactory;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.26"></a> +<span class="sourceLineNo">024</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.slf4j.Logger;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.slf4j.LoggerFactory;<a name="line.26"></a> <span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.client.Put;<a name="line.27"></a> <span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.28"></a> <span class="sourceLineNo">029</span>import org.apache.hadoop.mapred.MapReduceBase;<a name="line.29"></a> @@ -46,8 +46,8 @@ <span class="sourceLineNo">038</span>extends MapReduceBase<a name="line.38"></a> <span class="sourceLineNo">039</span>implements TableReduce<ImmutableBytesWritable, Put> {<a name="line.39"></a> <span class="sourceLineNo">040</span> @SuppressWarnings("unused")<a name="line.40"></a> -<span class="sourceLineNo">041</span> private static final Log LOG =<a name="line.41"></a> -<span class="sourceLineNo">042</span> LogFactory.getLog(IdentityTableReduce.class.getName());<a name="line.42"></a> +<span class="sourceLineNo">041</span> private static final Logger LOG =<a name="line.41"></a> +<span class="sourceLineNo">042</span> LoggerFactory.getLogger(IdentityTableReduce.class.getName());<a name="line.42"></a> <span class="sourceLineNo">043</span><a name="line.43"></a> <span class="sourceLineNo">044</span> /**<a name="line.44"></a> <span class="sourceLineNo">045</span> * No aggregation, output pairs of (key, record)<a name="line.45"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormat.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormat.html b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormat.html index 8486e0b..4e5c800 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormat.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormat.html @@ -28,9 +28,9 @@ <span class="sourceLineNo">020</span><a name="line.20"></a> <span class="sourceLineNo">021</span>import java.io.IOException;<a name="line.21"></a> <span class="sourceLineNo">022</span><a name="line.22"></a> -<span class="sourceLineNo">023</span>import org.apache.commons.logging.Log;<a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.LogFactory;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.25"></a> +<span class="sourceLineNo">023</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.23"></a> +<span class="sourceLineNo">024</span>import org.slf4j.Logger;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.slf4j.LoggerFactory;<a name="line.25"></a> <span class="sourceLineNo">026</span>import org.apache.hadoop.fs.Path;<a name="line.26"></a> <span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.TableName;<a name="line.27"></a> <span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.28"></a> @@ -47,7 +47,7 @@ <span class="sourceLineNo">039</span>@InterfaceAudience.Public<a name="line.39"></a> <span class="sourceLineNo">040</span>public class TableInputFormat extends TableInputFormatBase implements<a name="line.40"></a> <span class="sourceLineNo">041</span> JobConfigurable {<a name="line.41"></a> -<span class="sourceLineNo">042</span> private static final Log LOG = LogFactory.getLog(TableInputFormat.class);<a name="line.42"></a> +<span class="sourceLineNo">042</span> private static final Logger LOG = LoggerFactory.getLogger(TableInputFormat.class);<a name="line.42"></a> <span class="sourceLineNo">043</span><a name="line.43"></a> <span class="sourceLineNo">044</span> /**<a name="line.44"></a> <span class="sourceLineNo">045</span> * space delimited list of columns<a name="line.45"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormatBase.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormatBase.html b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormatBase.html index 4927b45..0b6e2d0 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormatBase.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableInputFormatBase.html @@ -29,9 +29,9 @@ <span class="sourceLineNo">021</span>import java.io.Closeable;<a name="line.21"></a> <span class="sourceLineNo">022</span>import java.io.IOException;<a name="line.22"></a> <span class="sourceLineNo">023</span><a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.Log;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.commons.logging.LogFactory;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.26"></a> +<span class="sourceLineNo">024</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.slf4j.Logger;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.slf4j.LoggerFactory;<a name="line.26"></a> <span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.HConstants;<a name="line.27"></a> <span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.TableName;<a name="line.28"></a> <span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.29"></a> @@ -87,7 +87,7 @@ <span class="sourceLineNo">079</span>@InterfaceAudience.Public<a name="line.79"></a> <span class="sourceLineNo">080</span>public abstract class TableInputFormatBase<a name="line.80"></a> <span class="sourceLineNo">081</span>implements InputFormat<ImmutableBytesWritable, Result> {<a name="line.81"></a> -<span class="sourceLineNo">082</span> private static final Log LOG = LogFactory.getLog(TableInputFormatBase.class);<a name="line.82"></a> +<span class="sourceLineNo">082</span> private static final Logger LOG = LoggerFactory.getLogger(TableInputFormatBase.class);<a name="line.82"></a> <span class="sourceLineNo">083</span> private byte [][] inputColumns;<a name="line.83"></a> <span class="sourceLineNo">084</span> private Table table;<a name="line.84"></a> <span class="sourceLineNo">085</span> private RegionLocator regionLocator;<a name="line.85"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.html b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.html index e85d130..71ef26b 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapred/TableRecordReaderImpl.html @@ -27,244 +27,243 @@ <span class="sourceLineNo">019</span>package org.apache.hadoop.hbase.mapred;<a name="line.19"></a> <span class="sourceLineNo">020</span><a name="line.20"></a> <span class="sourceLineNo">021</span>import java.io.IOException;<a name="line.21"></a> -<span class="sourceLineNo">022</span><a name="line.22"></a> -<span class="sourceLineNo">023</span>import org.apache.commons.logging.Log;<a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.LogFactory;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.hadoop.conf.Configuration;<a name="line.26"></a> -<span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.client.Result;<a name="line.27"></a> -<span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.client.ResultScanner;<a name="line.28"></a> -<span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.client.Scan;<a name="line.29"></a> -<span class="sourceLineNo">030</span>import org.apache.hadoop.hbase.client.ScannerCallable;<a name="line.30"></a> -<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.client.Table;<a name="line.31"></a> -<span class="sourceLineNo">032</span>import org.apache.hadoop.hbase.DoNotRetryIOException;<a name="line.32"></a> -<span class="sourceLineNo">033</span>import org.apache.hadoop.hbase.filter.Filter;<a name="line.33"></a> -<span class="sourceLineNo">034</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.34"></a> -<span class="sourceLineNo">035</span>import org.apache.hadoop.hbase.mapreduce.TableInputFormat;<a name="line.35"></a> -<span class="sourceLineNo">036</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.36"></a> -<span class="sourceLineNo">037</span>import org.apache.hadoop.util.StringUtils;<a name="line.37"></a> -<span class="sourceLineNo">038</span><a name="line.38"></a> -<span class="sourceLineNo">039</span>import static org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl.LOG_PER_ROW_COUNT;<a name="line.39"></a> -<span class="sourceLineNo">040</span><a name="line.40"></a> -<span class="sourceLineNo">041</span>/**<a name="line.41"></a> -<span class="sourceLineNo">042</span> * Iterate over an HBase table data, return (Text, RowResult) pairs<a name="line.42"></a> -<span class="sourceLineNo">043</span> */<a name="line.43"></a> -<span class="sourceLineNo">044</span>@InterfaceAudience.Public<a name="line.44"></a> -<span class="sourceLineNo">045</span>public class TableRecordReaderImpl {<a name="line.45"></a> -<span class="sourceLineNo">046</span> private static final Log LOG = LogFactory.getLog(TableRecordReaderImpl.class);<a name="line.46"></a> -<span class="sourceLineNo">047</span><a name="line.47"></a> -<span class="sourceLineNo">048</span> private byte [] startRow;<a name="line.48"></a> -<span class="sourceLineNo">049</span> private byte [] endRow;<a name="line.49"></a> -<span class="sourceLineNo">050</span> private byte [] lastSuccessfulRow;<a name="line.50"></a> -<span class="sourceLineNo">051</span> private Filter trrRowFilter;<a name="line.51"></a> -<span class="sourceLineNo">052</span> private ResultScanner scanner;<a name="line.52"></a> -<span class="sourceLineNo">053</span> private Table htable;<a name="line.53"></a> -<span class="sourceLineNo">054</span> private byte [][] trrInputColumns;<a name="line.54"></a> -<span class="sourceLineNo">055</span> private long timestamp;<a name="line.55"></a> -<span class="sourceLineNo">056</span> private int rowcount;<a name="line.56"></a> -<span class="sourceLineNo">057</span> private boolean logScannerActivity = false;<a name="line.57"></a> -<span class="sourceLineNo">058</span> private int logPerRowCount = 100;<a name="line.58"></a> -<span class="sourceLineNo">059</span><a name="line.59"></a> -<span class="sourceLineNo">060</span> /**<a name="line.60"></a> -<span class="sourceLineNo">061</span> * Restart from survivable exceptions by creating a new scanner.<a name="line.61"></a> -<span class="sourceLineNo">062</span> *<a name="line.62"></a> -<span class="sourceLineNo">063</span> * @param firstRow<a name="line.63"></a> -<span class="sourceLineNo">064</span> * @throws IOException<a name="line.64"></a> -<span class="sourceLineNo">065</span> */<a name="line.65"></a> -<span class="sourceLineNo">066</span> public void restart(byte[] firstRow) throws IOException {<a name="line.66"></a> -<span class="sourceLineNo">067</span> Scan currentScan;<a name="line.67"></a> -<span class="sourceLineNo">068</span> if ((endRow != null) && (endRow.length > 0)) {<a name="line.68"></a> -<span class="sourceLineNo">069</span> if (trrRowFilter != null) {<a name="line.69"></a> -<span class="sourceLineNo">070</span> Scan scan = new Scan(firstRow, endRow);<a name="line.70"></a> -<span class="sourceLineNo">071</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.71"></a> -<span class="sourceLineNo">072</span> scan.setFilter(trrRowFilter);<a name="line.72"></a> -<span class="sourceLineNo">073</span> scan.setCacheBlocks(false);<a name="line.73"></a> -<span class="sourceLineNo">074</span> this.scanner = this.htable.getScanner(scan);<a name="line.74"></a> -<span class="sourceLineNo">075</span> currentScan = scan;<a name="line.75"></a> -<span class="sourceLineNo">076</span> } else {<a name="line.76"></a> -<span class="sourceLineNo">077</span> LOG.debug("TIFB.restart, firstRow: " +<a name="line.77"></a> -<span class="sourceLineNo">078</span> Bytes.toStringBinary(firstRow) + ", endRow: " +<a name="line.78"></a> -<span class="sourceLineNo">079</span> Bytes.toStringBinary(endRow));<a name="line.79"></a> -<span class="sourceLineNo">080</span> Scan scan = new Scan(firstRow, endRow);<a name="line.80"></a> -<span class="sourceLineNo">081</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.81"></a> -<span class="sourceLineNo">082</span> this.scanner = this.htable.getScanner(scan);<a name="line.82"></a> -<span class="sourceLineNo">083</span> currentScan = scan;<a name="line.83"></a> -<span class="sourceLineNo">084</span> }<a name="line.84"></a> -<span class="sourceLineNo">085</span> } else {<a name="line.85"></a> -<span class="sourceLineNo">086</span> LOG.debug("TIFB.restart, firstRow: " +<a name="line.86"></a> -<span class="sourceLineNo">087</span> Bytes.toStringBinary(firstRow) + ", no endRow");<a name="line.87"></a> -<span class="sourceLineNo">088</span><a name="line.88"></a> -<span class="sourceLineNo">089</span> Scan scan = new Scan(firstRow);<a name="line.89"></a> -<span class="sourceLineNo">090</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.90"></a> -<span class="sourceLineNo">091</span> scan.setFilter(trrRowFilter);<a name="line.91"></a> -<span class="sourceLineNo">092</span> this.scanner = this.htable.getScanner(scan);<a name="line.92"></a> -<span class="sourceLineNo">093</span> currentScan = scan;<a name="line.93"></a> -<span class="sourceLineNo">094</span> }<a name="line.94"></a> -<span class="sourceLineNo">095</span> if (logScannerActivity) {<a name="line.95"></a> -<span class="sourceLineNo">096</span> LOG.info("Current scan=" + currentScan.toString());<a name="line.96"></a> -<span class="sourceLineNo">097</span> timestamp = System.currentTimeMillis();<a name="line.97"></a> -<span class="sourceLineNo">098</span> rowcount = 0;<a name="line.98"></a> -<span class="sourceLineNo">099</span> }<a name="line.99"></a> -<span class="sourceLineNo">100</span> }<a name="line.100"></a> -<span class="sourceLineNo">101</span><a name="line.101"></a> -<span class="sourceLineNo">102</span> /**<a name="line.102"></a> -<span class="sourceLineNo">103</span> * Build the scanner. Not done in constructor to allow for extension.<a name="line.103"></a> -<span class="sourceLineNo">104</span> *<a name="line.104"></a> -<span class="sourceLineNo">105</span> * @throws IOException<a name="line.105"></a> -<span class="sourceLineNo">106</span> */<a name="line.106"></a> -<span class="sourceLineNo">107</span> public void init() throws IOException {<a name="line.107"></a> -<span class="sourceLineNo">108</span> restart(startRow);<a name="line.108"></a> -<span class="sourceLineNo">109</span> }<a name="line.109"></a> -<span class="sourceLineNo">110</span><a name="line.110"></a> -<span class="sourceLineNo">111</span> byte[] getStartRow() {<a name="line.111"></a> -<span class="sourceLineNo">112</span> return this.startRow;<a name="line.112"></a> -<span class="sourceLineNo">113</span> }<a name="line.113"></a> -<span class="sourceLineNo">114</span> /**<a name="line.114"></a> -<span class="sourceLineNo">115</span> * @param htable the {@link org.apache.hadoop.hbase.HTableDescriptor} to scan.<a name="line.115"></a> -<span class="sourceLineNo">116</span> */<a name="line.116"></a> -<span class="sourceLineNo">117</span> public void setHTable(Table htable) {<a name="line.117"></a> -<span class="sourceLineNo">118</span> Configuration conf = htable.getConfiguration();<a name="line.118"></a> -<span class="sourceLineNo">119</span> logScannerActivity = conf.getBoolean(<a name="line.119"></a> -<span class="sourceLineNo">120</span> ScannerCallable.LOG_SCANNER_ACTIVITY, false);<a name="line.120"></a> -<span class="sourceLineNo">121</span> logPerRowCount = conf.getInt(LOG_PER_ROW_COUNT, 100);<a name="line.121"></a> -<span class="sourceLineNo">122</span> this.htable = htable;<a name="line.122"></a> -<span class="sourceLineNo">123</span> }<a name="line.123"></a> -<span class="sourceLineNo">124</span><a name="line.124"></a> -<span class="sourceLineNo">125</span> /**<a name="line.125"></a> -<span class="sourceLineNo">126</span> * @param inputColumns the columns to be placed in {@link Result}.<a name="line.126"></a> -<span class="sourceLineNo">127</span> */<a name="line.127"></a> -<span class="sourceLineNo">128</span> public void setInputColumns(final byte [][] inputColumns) {<a name="line.128"></a> -<span class="sourceLineNo">129</span> this.trrInputColumns = inputColumns;<a name="line.129"></a> -<span class="sourceLineNo">130</span> }<a name="line.130"></a> -<span class="sourceLineNo">131</span><a name="line.131"></a> -<span class="sourceLineNo">132</span> /**<a name="line.132"></a> -<span class="sourceLineNo">133</span> * @param startRow the first row in the split<a name="line.133"></a> -<span class="sourceLineNo">134</span> */<a name="line.134"></a> -<span class="sourceLineNo">135</span> public void setStartRow(final byte [] startRow) {<a name="line.135"></a> -<span class="sourceLineNo">136</span> this.startRow = startRow;<a name="line.136"></a> -<span class="sourceLineNo">137</span> }<a name="line.137"></a> -<span class="sourceLineNo">138</span><a name="line.138"></a> -<span class="sourceLineNo">139</span> /**<a name="line.139"></a> -<span class="sourceLineNo">140</span> *<a name="line.140"></a> -<span class="sourceLineNo">141</span> * @param endRow the last row in the split<a name="line.141"></a> -<span class="sourceLineNo">142</span> */<a name="line.142"></a> -<span class="sourceLineNo">143</span> public void setEndRow(final byte [] endRow) {<a name="line.143"></a> -<span class="sourceLineNo">144</span> this.endRow = endRow;<a name="line.144"></a> -<span class="sourceLineNo">145</span> }<a name="line.145"></a> -<span class="sourceLineNo">146</span><a name="line.146"></a> -<span class="sourceLineNo">147</span> /**<a name="line.147"></a> -<span class="sourceLineNo">148</span> * @param rowFilter the {@link Filter} to be used.<a name="line.148"></a> -<span class="sourceLineNo">149</span> */<a name="line.149"></a> -<span class="sourceLineNo">150</span> public void setRowFilter(Filter rowFilter) {<a name="line.150"></a> -<span class="sourceLineNo">151</span> this.trrRowFilter = rowFilter;<a name="line.151"></a> -<span class="sourceLineNo">152</span> }<a name="line.152"></a> -<span class="sourceLineNo">153</span><a name="line.153"></a> -<span class="sourceLineNo">154</span> public void close() {<a name="line.154"></a> -<span class="sourceLineNo">155</span> if (this.scanner != null) {<a name="line.155"></a> -<span class="sourceLineNo">156</span> this.scanner.close();<a name="line.156"></a> -<span class="sourceLineNo">157</span> }<a name="line.157"></a> -<span class="sourceLineNo">158</span> try {<a name="line.158"></a> -<span class="sourceLineNo">159</span> this.htable.close();<a name="line.159"></a> -<span class="sourceLineNo">160</span> } catch (IOException ioe) {<a name="line.160"></a> -<span class="sourceLineNo">161</span> LOG.warn("Error closing table", ioe);<a name="line.161"></a> -<span class="sourceLineNo">162</span> }<a name="line.162"></a> -<span class="sourceLineNo">163</span> }<a name="line.163"></a> -<span class="sourceLineNo">164</span><a name="line.164"></a> -<span class="sourceLineNo">165</span> /**<a name="line.165"></a> -<span class="sourceLineNo">166</span> * @return ImmutableBytesWritable<a name="line.166"></a> -<span class="sourceLineNo">167</span> *<a name="line.167"></a> -<span class="sourceLineNo">168</span> * @see org.apache.hadoop.mapred.RecordReader#createKey()<a name="line.168"></a> -<span class="sourceLineNo">169</span> */<a name="line.169"></a> -<span class="sourceLineNo">170</span> public ImmutableBytesWritable createKey() {<a name="line.170"></a> -<span class="sourceLineNo">171</span> return new ImmutableBytesWritable();<a name="line.171"></a> -<span class="sourceLineNo">172</span> }<a name="line.172"></a> -<span class="sourceLineNo">173</span><a name="line.173"></a> -<span class="sourceLineNo">174</span> /**<a name="line.174"></a> -<span class="sourceLineNo">175</span> * @return RowResult<a name="line.175"></a> -<span class="sourceLineNo">176</span> *<a name="line.176"></a> -<span class="sourceLineNo">177</span> * @see org.apache.hadoop.mapred.RecordReader#createValue()<a name="line.177"></a> -<span class="sourceLineNo">178</span> */<a name="line.178"></a> -<span class="sourceLineNo">179</span> public Result createValue() {<a name="line.179"></a> -<span class="sourceLineNo">180</span> return new Result();<a name="line.180"></a> -<span class="sourceLineNo">181</span> }<a name="line.181"></a> -<span class="sourceLineNo">182</span><a name="line.182"></a> -<span class="sourceLineNo">183</span> public long getPos() {<a name="line.183"></a> -<span class="sourceLineNo">184</span> // This should be the ordinal tuple in the range;<a name="line.184"></a> -<span class="sourceLineNo">185</span> // not clear how to calculate...<a name="line.185"></a> -<span class="sourceLineNo">186</span> return 0;<a name="line.186"></a> -<span class="sourceLineNo">187</span> }<a name="line.187"></a> -<span class="sourceLineNo">188</span><a name="line.188"></a> -<span class="sourceLineNo">189</span> public float getProgress() {<a name="line.189"></a> -<span class="sourceLineNo">190</span> // Depends on the total number of tuples and getPos<a name="line.190"></a> -<span class="sourceLineNo">191</span> return 0;<a name="line.191"></a> -<span class="sourceLineNo">192</span> }<a name="line.192"></a> -<span class="sourceLineNo">193</span><a name="line.193"></a> -<span class="sourceLineNo">194</span> /**<a name="line.194"></a> -<span class="sourceLineNo">195</span> * @param key HStoreKey as input key.<a name="line.195"></a> -<span class="sourceLineNo">196</span> * @param value MapWritable as input value<a name="line.196"></a> -<span class="sourceLineNo">197</span> * @return true if there was more data<a name="line.197"></a> -<span class="sourceLineNo">198</span> * @throws IOException<a name="line.198"></a> -<span class="sourceLineNo">199</span> */<a name="line.199"></a> -<span class="sourceLineNo">200</span> public boolean next(ImmutableBytesWritable key, Result value)<a name="line.200"></a> -<span class="sourceLineNo">201</span> throws IOException {<a name="line.201"></a> -<span class="sourceLineNo">202</span> Result result;<a name="line.202"></a> -<span class="sourceLineNo">203</span> try {<a name="line.203"></a> -<span class="sourceLineNo">204</span> try {<a name="line.204"></a> -<span class="sourceLineNo">205</span> result = this.scanner.next();<a name="line.205"></a> -<span class="sourceLineNo">206</span> if (logScannerActivity) {<a name="line.206"></a> -<span class="sourceLineNo">207</span> rowcount ++;<a name="line.207"></a> -<span class="sourceLineNo">208</span> if (rowcount >= logPerRowCount) {<a name="line.208"></a> -<span class="sourceLineNo">209</span> long now = System.currentTimeMillis();<a name="line.209"></a> -<span class="sourceLineNo">210</span> LOG.info("Mapper took " + (now-timestamp)<a name="line.210"></a> -<span class="sourceLineNo">211</span> + "ms to process " + rowcount + " rows");<a name="line.211"></a> -<span class="sourceLineNo">212</span> timestamp = now;<a name="line.212"></a> -<span class="sourceLineNo">213</span> rowcount = 0;<a name="line.213"></a> -<span class="sourceLineNo">214</span> }<a name="line.214"></a> -<span class="sourceLineNo">215</span> }<a name="line.215"></a> -<span class="sourceLineNo">216</span> } catch (IOException e) {<a name="line.216"></a> -<span class="sourceLineNo">217</span> // do not retry if the exception tells us not to do so<a name="line.217"></a> -<span class="sourceLineNo">218</span> if (e instanceof DoNotRetryIOException) {<a name="line.218"></a> -<span class="sourceLineNo">219</span> throw e;<a name="line.219"></a> -<span class="sourceLineNo">220</span> }<a name="line.220"></a> -<span class="sourceLineNo">221</span> // try to handle all other IOExceptions by restarting<a name="line.221"></a> -<span class="sourceLineNo">222</span> // the scanner, if the second call fails, it will be rethrown<a name="line.222"></a> -<span class="sourceLineNo">223</span> LOG.debug("recovered from " + StringUtils.stringifyException(e));<a name="line.223"></a> -<span class="sourceLineNo">224</span> if (lastSuccessfulRow == null) {<a name="line.224"></a> -<span class="sourceLineNo">225</span> LOG.warn("We are restarting the first next() invocation," +<a name="line.225"></a> -<span class="sourceLineNo">226</span> " if your mapper has restarted a few other times like this" +<a name="line.226"></a> -<span class="sourceLineNo">227</span> " then you should consider killing this job and investigate" +<a name="line.227"></a> -<span class="sourceLineNo">228</span> " why it's taking so long.");<a name="line.228"></a> -<span class="sourceLineNo">229</span> }<a name="line.229"></a> -<span class="sourceLineNo">230</span> if (lastSuccessfulRow == null) {<a name="line.230"></a> -<span class="sourceLineNo">231</span> restart(startRow);<a name="line.231"></a> -<span class="sourceLineNo">232</span> } else {<a name="line.232"></a> -<span class="sourceLineNo">233</span> restart(lastSuccessfulRow);<a name="line.233"></a> -<span class="sourceLineNo">234</span> this.scanner.next(); // skip presumed already mapped row<a name="line.234"></a> -<span class="sourceLineNo">235</span> }<a name="line.235"></a> -<span class="sourceLineNo">236</span> result = this.scanner.next();<a name="line.236"></a> -<span class="sourceLineNo">237</span> }<a name="line.237"></a> -<span class="sourceLineNo">238</span><a name="line.238"></a> -<span class="sourceLineNo">239</span> if (result != null && result.size() > 0) {<a name="line.239"></a> -<span class="sourceLineNo">240</span> key.set(result.getRow());<a name="line.240"></a> -<span class="sourceLineNo">241</span> lastSuccessfulRow = key.get();<a name="line.241"></a> -<span class="sourceLineNo">242</span> value.copyFrom(result);<a name="line.242"></a> -<span class="sourceLineNo">243</span> return true;<a name="line.243"></a> -<span class="sourceLineNo">244</span> }<a name="line.244"></a> -<span class="sourceLineNo">245</span> return false;<a name="line.245"></a> -<span class="sourceLineNo">246</span> } catch (IOException ioe) {<a name="line.246"></a> -<span class="sourceLineNo">247</span> if (logScannerActivity) {<a name="line.247"></a> -<span class="sourceLineNo">248</span> long now = System.currentTimeMillis();<a name="line.248"></a> -<span class="sourceLineNo">249</span> LOG.info("Mapper took " + (now-timestamp)<a name="line.249"></a> -<span class="sourceLineNo">250</span> + "ms to process " + rowcount + " rows");<a name="line.250"></a> -<span class="sourceLineNo">251</span> LOG.info(ioe);<a name="line.251"></a> -<span class="sourceLineNo">252</span> String lastRow = lastSuccessfulRow == null ?<a name="line.252"></a> -<span class="sourceLineNo">253</span> "null" : Bytes.toStringBinary(lastSuccessfulRow);<a name="line.253"></a> -<span class="sourceLineNo">254</span> LOG.info("lastSuccessfulRow=" + lastRow);<a name="line.254"></a> -<span class="sourceLineNo">255</span> }<a name="line.255"></a> -<span class="sourceLineNo">256</span> throw ioe;<a name="line.256"></a> -<span class="sourceLineNo">257</span> }<a name="line.257"></a> -<span class="sourceLineNo">258</span> }<a name="line.258"></a> -<span class="sourceLineNo">259</span>}<a name="line.259"></a> +<span class="sourceLineNo">022</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.22"></a> +<span class="sourceLineNo">023</span>import org.slf4j.Logger;<a name="line.23"></a> +<span class="sourceLineNo">024</span>import org.slf4j.LoggerFactory;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.apache.hadoop.conf.Configuration;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.apache.hadoop.hbase.client.Result;<a name="line.26"></a> +<span class="sourceLineNo">027</span>import org.apache.hadoop.hbase.client.ResultScanner;<a name="line.27"></a> +<span class="sourceLineNo">028</span>import org.apache.hadoop.hbase.client.Scan;<a name="line.28"></a> +<span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.client.ScannerCallable;<a name="line.29"></a> +<span class="sourceLineNo">030</span>import org.apache.hadoop.hbase.client.Table;<a name="line.30"></a> +<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.DoNotRetryIOException;<a name="line.31"></a> +<span class="sourceLineNo">032</span>import org.apache.hadoop.hbase.filter.Filter;<a name="line.32"></a> +<span class="sourceLineNo">033</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.33"></a> +<span class="sourceLineNo">034</span>import org.apache.hadoop.hbase.mapreduce.TableInputFormat;<a name="line.34"></a> +<span class="sourceLineNo">035</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.35"></a> +<span class="sourceLineNo">036</span>import org.apache.hadoop.util.StringUtils;<a name="line.36"></a> +<span class="sourceLineNo">037</span><a name="line.37"></a> +<span class="sourceLineNo">038</span>import static org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl.LOG_PER_ROW_COUNT;<a name="line.38"></a> +<span class="sourceLineNo">039</span><a name="line.39"></a> +<span class="sourceLineNo">040</span>/**<a name="line.40"></a> +<span class="sourceLineNo">041</span> * Iterate over an HBase table data, return (Text, RowResult) pairs<a name="line.41"></a> +<span class="sourceLineNo">042</span> */<a name="line.42"></a> +<span class="sourceLineNo">043</span>@InterfaceAudience.Public<a name="line.43"></a> +<span class="sourceLineNo">044</span>public class TableRecordReaderImpl {<a name="line.44"></a> +<span class="sourceLineNo">045</span> private static final Logger LOG = LoggerFactory.getLogger(TableRecordReaderImpl.class);<a name="line.45"></a> +<span class="sourceLineNo">046</span><a name="line.46"></a> +<span class="sourceLineNo">047</span> private byte [] startRow;<a name="line.47"></a> +<span class="sourceLineNo">048</span> private byte [] endRow;<a name="line.48"></a> +<span class="sourceLineNo">049</span> private byte [] lastSuccessfulRow;<a name="line.49"></a> +<span class="sourceLineNo">050</span> private Filter trrRowFilter;<a name="line.50"></a> +<span class="sourceLineNo">051</span> private ResultScanner scanner;<a name="line.51"></a> +<span class="sourceLineNo">052</span> private Table htable;<a name="line.52"></a> +<span class="sourceLineNo">053</span> private byte [][] trrInputColumns;<a name="line.53"></a> +<span class="sourceLineNo">054</span> private long timestamp;<a name="line.54"></a> +<span class="sourceLineNo">055</span> private int rowcount;<a name="line.55"></a> +<span class="sourceLineNo">056</span> private boolean logScannerActivity = false;<a name="line.56"></a> +<span class="sourceLineNo">057</span> private int logPerRowCount = 100;<a name="line.57"></a> +<span class="sourceLineNo">058</span><a name="line.58"></a> +<span class="sourceLineNo">059</span> /**<a name="line.59"></a> +<span class="sourceLineNo">060</span> * Restart from survivable exceptions by creating a new scanner.<a name="line.60"></a> +<span class="sourceLineNo">061</span> *<a name="line.61"></a> +<span class="sourceLineNo">062</span> * @param firstRow<a name="line.62"></a> +<span class="sourceLineNo">063</span> * @throws IOException<a name="line.63"></a> +<span class="sourceLineNo">064</span> */<a name="line.64"></a> +<span class="sourceLineNo">065</span> public void restart(byte[] firstRow) throws IOException {<a name="line.65"></a> +<span class="sourceLineNo">066</span> Scan currentScan;<a name="line.66"></a> +<span class="sourceLineNo">067</span> if ((endRow != null) && (endRow.length > 0)) {<a name="line.67"></a> +<span class="sourceLineNo">068</span> if (trrRowFilter != null) {<a name="line.68"></a> +<span class="sourceLineNo">069</span> Scan scan = new Scan(firstRow, endRow);<a name="line.69"></a> +<span class="sourceLineNo">070</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.70"></a> +<span class="sourceLineNo">071</span> scan.setFilter(trrRowFilter);<a name="line.71"></a> +<span class="sourceLineNo">072</span> scan.setCacheBlocks(false);<a name="line.72"></a> +<span class="sourceLineNo">073</span> this.scanner = this.htable.getScanner(scan);<a name="line.73"></a> +<span class="sourceLineNo">074</span> currentScan = scan;<a name="line.74"></a> +<span class="sourceLineNo">075</span> } else {<a name="line.75"></a> +<span class="sourceLineNo">076</span> LOG.debug("TIFB.restart, firstRow: " +<a name="line.76"></a> +<span class="sourceLineNo">077</span> Bytes.toStringBinary(firstRow) + ", endRow: " +<a name="line.77"></a> +<span class="sourceLineNo">078</span> Bytes.toStringBinary(endRow));<a name="line.78"></a> +<span class="sourceLineNo">079</span> Scan scan = new Scan(firstRow, endRow);<a name="line.79"></a> +<span class="sourceLineNo">080</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.80"></a> +<span class="sourceLineNo">081</span> this.scanner = this.htable.getScanner(scan);<a name="line.81"></a> +<span class="sourceLineNo">082</span> currentScan = scan;<a name="line.82"></a> +<span class="sourceLineNo">083</span> }<a name="line.83"></a> +<span class="sourceLineNo">084</span> } else {<a name="line.84"></a> +<span class="sourceLineNo">085</span> LOG.debug("TIFB.restart, firstRow: " +<a name="line.85"></a> +<span class="sourceLineNo">086</span> Bytes.toStringBinary(firstRow) + ", no endRow");<a name="line.86"></a> +<span class="sourceLineNo">087</span><a name="line.87"></a> +<span class="sourceLineNo">088</span> Scan scan = new Scan(firstRow);<a name="line.88"></a> +<span class="sourceLineNo">089</span> TableInputFormat.addColumns(scan, trrInputColumns);<a name="line.89"></a> +<span class="sourceLineNo">090</span> scan.setFilter(trrRowFilter);<a name="line.90"></a> +<span class="sourceLineNo">091</span> this.scanner = this.htable.getScanner(scan);<a name="line.91"></a> +<span class="sourceLineNo">092</span> currentScan = scan;<a name="line.92"></a> +<span class="sourceLineNo">093</span> }<a name="line.93"></a> +<span class="sourceLineNo">094</span> if (logScannerActivity) {<a name="line.94"></a> +<span class="sourceLineNo">095</span> LOG.info("Current scan=" + currentScan.toString());<a name="line.95"></a> +<span class="sourceLineNo">096</span> timestamp = System.currentTimeMillis();<a name="line.96"></a> +<span class="sourceLineNo">097</span> rowcount = 0;<a name="line.97"></a> +<span class="sourceLineNo">098</span> }<a name="line.98"></a> +<span class="sourceLineNo">099</span> }<a name="line.99"></a> +<span class="sourceLineNo">100</span><a name="line.100"></a> +<span class="sourceLineNo">101</span> /**<a name="line.101"></a> +<span class="sourceLineNo">102</span> * Build the scanner. Not done in constructor to allow for extension.<a name="line.102"></a> +<span class="sourceLineNo">103</span> *<a name="line.103"></a> +<span class="sourceLineNo">104</span> * @throws IOException<a name="line.104"></a> +<span class="sourceLineNo">105</span> */<a name="line.105"></a> +<span class="sourceLineNo">106</span> public void init() throws IOException {<a name="line.106"></a> +<span class="sourceLineNo">107</span> restart(startRow);<a name="line.107"></a> +<span class="sourceLineNo">108</span> }<a name="line.108"></a> +<span class="sourceLineNo">109</span><a name="line.109"></a> +<span class="sourceLineNo">110</span> byte[] getStartRow() {<a name="line.110"></a> +<span class="sourceLineNo">111</span> return this.startRow;<a name="line.111"></a> +<span class="sourceLineNo">112</span> }<a name="line.112"></a> +<span class="sourceLineNo">113</span> /**<a name="line.113"></a> +<span class="sourceLineNo">114</span> * @param htable the {@link org.apache.hadoop.hbase.HTableDescriptor} to scan.<a name="line.114"></a> +<span class="sourceLineNo">115</span> */<a name="line.115"></a> +<span class="sourceLineNo">116</span> public void setHTable(Table htable) {<a name="line.116"></a> +<span class="sourceLineNo">117</span> Configuration conf = htable.getConfiguration();<a name="line.117"></a> +<span class="sourceLineNo">118</span> logScannerActivity = conf.getBoolean(<a name="line.118"></a> +<span class="sourceLineNo">119</span> ScannerCallable.LOG_SCANNER_ACTIVITY, false);<a name="line.119"></a> +<span class="sourceLineNo">120</span> logPerRowCount = conf.getInt(LOG_PER_ROW_COUNT, 100);<a name="line.120"></a> +<span class="sourceLineNo">121</span> this.htable = htable;<a name="line.121"></a> +<span class="sourceLineNo">122</span> }<a name="line.122"></a> +<span class="sourceLineNo">123</span><a name="line.123"></a> +<span class="sourceLineNo">124</span> /**<a name="line.124"></a> +<span class="sourceLineNo">125</span> * @param inputColumns the columns to be placed in {@link Result}.<a name="line.125"></a> +<span class="sourceLineNo">126</span> */<a name="line.126"></a> +<span class="sourceLineNo">127</span> public void setInputColumns(final byte [][] inputColumns) {<a name="line.127"></a> +<span class="sourceLineNo">128</span> this.trrInputColumns = inputColumns;<a name="line.128"></a> +<span class="sourceLineNo">129</span> }<a name="line.129"></a> +<span class="sourceLineNo">130</span><a name="line.130"></a> +<span class="sourceLineNo">131</span> /**<a name="line.131"></a> +<span class="sourceLineNo">132</span> * @param startRow the first row in the split<a name="line.132"></a> +<span class="sourceLineNo">133</span> */<a name="line.133"></a> +<span class="sourceLineNo">134</span> public void setStartRow(final byte [] startRow) {<a name="line.134"></a> +<span class="sourceLineNo">135</span> this.startRow = startRow;<a name="line.135"></a> +<span class="sourceLineNo">136</span> }<a name="line.136"></a> +<span class="sourceLineNo">137</span><a name="line.137"></a> +<span class="sourceLineNo">138</span> /**<a name="line.138"></a> +<span class="sourceLineNo">139</span> *<a name="line.139"></a> +<span class="sourceLineNo">140</span> * @param endRow the last row in the split<a name="line.140"></a> +<span class="sourceLineNo">141</span> */<a name="line.141"></a> +<span class="sourceLineNo">142</span> public void setEndRow(final byte [] endRow) {<a name="line.142"></a> +<span class="sourceLineNo">143</span> this.endRow = endRow;<a name="line.143"></a> +<span class="sourceLineNo">144</span> }<a name="line.144"></a> +<span class="sourceLineNo">145</span><a name="line.145"></a> +<span class="sourceLineNo">146</span> /**<a name="line.146"></a> +<span class="sourceLineNo">147</span> * @param rowFilter the {@link Filter} to be used.<a name="line.147"></a> +<span class="sourceLineNo">148</span> */<a name="line.148"></a> +<span class="sourceLineNo">149</span> public void setRowFilter(Filter rowFilter) {<a name="line.149"></a> +<span class="sourceLineNo">150</span> this.trrRowFilter = rowFilter;<a name="line.150"></a> +<span class="sourceLineNo">151</span> }<a name="line.151"></a> +<span class="sourceLineNo">152</span><a name="line.152"></a> +<span class="sourceLineNo">153</span> public void close() {<a name="line.153"></a> +<span class="sourceLineNo">154</span> if (this.scanner != null) {<a name="line.154"></a> +<span class="sourceLineNo">155</span> this.scanner.close();<a name="line.155"></a> +<span class="sourceLineNo">156</span> }<a name="line.156"></a> +<span class="sourceLineNo">157</span> try {<a name="line.157"></a> +<span class="sourceLineNo">158</span> this.htable.close();<a name="line.158"></a> +<span class="sourceLineNo">159</span> } catch (IOException ioe) {<a name="line.159"></a> +<span class="sourceLineNo">160</span> LOG.warn("Error closing table", ioe);<a name="line.160"></a> +<span class="sourceLineNo">161</span> }<a name="line.161"></a> +<span class="sourceLineNo">162</span> }<a name="line.162"></a> +<span class="sourceLineNo">163</span><a name="line.163"></a> +<span class="sourceLineNo">164</span> /**<a name="line.164"></a> +<span class="sourceLineNo">165</span> * @return ImmutableBytesWritable<a name="line.165"></a> +<span class="sourceLineNo">166</span> *<a name="line.166"></a> +<span class="sourceLineNo">167</span> * @see org.apache.hadoop.mapred.RecordReader#createKey()<a name="line.167"></a> +<span class="sourceLineNo">168</span> */<a name="line.168"></a> +<span class="sourceLineNo">169</span> public ImmutableBytesWritable createKey() {<a name="line.169"></a> +<span class="sourceLineNo">170</span> return new ImmutableBytesWritable();<a name="line.170"></a> +<span class="sourceLineNo">171</span> }<a name="line.171"></a> +<span class="sourceLineNo">172</span><a name="line.172"></a> +<span class="sourceLineNo">173</span> /**<a name="line.173"></a> +<span class="sourceLineNo">174</span> * @return RowResult<a name="line.174"></a> +<span class="sourceLineNo">175</span> *<a name="line.175"></a> +<span class="sourceLineNo">176</span> * @see org.apache.hadoop.mapred.RecordReader#createValue()<a name="line.176"></a> +<span class="sourceLineNo">177</span> */<a name="line.177"></a> +<span class="sourceLineNo">178</span> public Result createValue() {<a name="line.178"></a> +<span class="sourceLineNo">179</span> return new Result();<a name="line.179"></a> +<span class="sourceLineNo">180</span> }<a name="line.180"></a> +<span class="sourceLineNo">181</span><a name="line.181"></a> +<span class="sourceLineNo">182</span> public long getPos() {<a name="line.182"></a> +<span class="sourceLineNo">183</span> // This should be the ordinal tuple in the range;<a name="line.183"></a> +<span class="sourceLineNo">184</span> // not clear how to calculate...<a name="line.184"></a> +<span class="sourceLineNo">185</span> return 0;<a name="line.185"></a> +<span class="sourceLineNo">186</span> }<a name="line.186"></a> +<span class="sourceLineNo">187</span><a name="line.187"></a> +<span class="sourceLineNo">188</span> public float getProgress() {<a name="line.188"></a> +<span class="sourceLineNo">189</span> // Depends on the total number of tuples and getPos<a name="line.189"></a> +<span class="sourceLineNo">190</span> return 0;<a name="line.190"></a> +<span class="sourceLineNo">191</span> }<a name="line.191"></a> +<span class="sourceLineNo">192</span><a name="line.192"></a> +<span class="sourceLineNo">193</span> /**<a name="line.193"></a> +<span class="sourceLineNo">194</span> * @param key HStoreKey as input key.<a name="line.194"></a> +<span class="sourceLineNo">195</span> * @param value MapWritable as input value<a name="line.195"></a> +<span class="sourceLineNo">196</span> * @return true if there was more data<a name="line.196"></a> +<span class="sourceLineNo">197</span> * @throws IOException<a name="line.197"></a> +<span class="sourceLineNo">198</span> */<a name="line.198"></a> +<span class="sourceLineNo">199</span> public boolean next(ImmutableBytesWritable key, Result value)<a name="line.199"></a> +<span class="sourceLineNo">200</span> throws IOException {<a name="line.200"></a> +<span class="sourceLineNo">201</span> Result result;<a name="line.201"></a> +<span class="sourceLineNo">202</span> try {<a name="line.202"></a> +<span class="sourceLineNo">203</span> try {<a name="line.203"></a> +<span class="sourceLineNo">204</span> result = this.scanner.next();<a name="line.204"></a> +<span class="sourceLineNo">205</span> if (logScannerActivity) {<a name="line.205"></a> +<span class="sourceLineNo">206</span> rowcount ++;<a name="line.206"></a> +<span class="sourceLineNo">207</span> if (rowcount >= logPerRowCount) {<a name="line.207"></a> +<span class="sourceLineNo">208</span> long now = System.currentTimeMillis();<a name="line.208"></a> +<span class="sourceLineNo">209</span> LOG.info("Mapper took " + (now-timestamp)<a name="line.209"></a> +<span class="sourceLineNo">210</span> + "ms to process " + rowcount + " rows");<a name="line.210"></a> +<span class="sourceLineNo">211</span> timestamp = now;<a name="line.211"></a> +<span class="sourceLineNo">212</span> rowcount = 0;<a name="line.212"></a> +<span class="sourceLineNo">213</span> }<a name="line.213"></a> +<span class="sourceLineNo">214</span> }<a name="line.214"></a> +<span class="sourceLineNo">215</span> } catch (IOException e) {<a name="line.215"></a> +<span class="sourceLineNo">216</span> // do not retry if the exception tells us not to do so<a name="line.216"></a> +<span class="sourceLineNo">217</span> if (e instanceof DoNotRetryIOException) {<a name="line.217"></a> +<span class="sourceLineNo">218</span> throw e;<a name="line.218"></a> +<span class="sourceLineNo">219</span> }<a name="line.219"></a> +<span class="sourceLineNo">220</span> // try to handle all other IOExceptions by restarting<a name="line.220"></a> +<span class="sourceLineNo">221</span> // the scanner, if the second call fails, it will be rethrown<a name="line.221"></a> +<span class="sourceLineNo">222</span> LOG.debug("recovered from " + StringUtils.stringifyException(e));<a name="line.222"></a> +<span class="sourceLineNo">223</span> if (lastSuccessfulRow == null) {<a name="line.223"></a> +<span class="sourceLineNo">224</span> LOG.warn("We are restarting the first next() invocation," +<a name="line.224"></a> +<span class="sourceLineNo">225</span> " if your mapper has restarted a few other times like this" +<a name="line.225"></a> +<span class="sourceLineNo">226</span> " then you should consider killing this job and investigate" +<a name="line.226"></a> +<span class="sourceLineNo">227</span> " why it's taking so long.");<a name="line.227"></a> +<span class="sourceLineNo">228</span> }<a name="line.228"></a> +<span class="sourceLineNo">229</span> if (lastSuccessfulRow == null) {<a name="line.229"></a> +<span class="sourceLineNo">230</span> restart(startRow);<a name="line.230"></a> +<span class="sourceLineNo">231</span> } else {<a name="line.231"></a> +<span class="sourceLineNo">232</span> restart(lastSuccessfulRow);<a name="line.232"></a> +<span class="sourceLineNo">233</span> this.scanner.next(); // skip presumed already mapped row<a name="line.233"></a> +<span class="sourceLineNo">234</span> }<a name="line.234"></a> +<span class="sourceLineNo">235</span> result = this.scanner.next();<a name="line.235"></a> +<span class="sourceLineNo">236</span> }<a name="line.236"></a> +<span class="sourceLineNo">237</span><a name="line.237"></a> +<span class="sourceLineNo">238</span> if (result != null && result.size() > 0) {<a name="line.238"></a> +<span class="sourceLineNo">239</span> key.set(result.getRow());<a name="line.239"></a> +<span class="sourceLineNo">240</span> lastSuccessfulRow = key.get();<a name="line.240"></a> +<span class="sourceLineNo">241</span> value.copyFrom(result);<a name="line.241"></a> +<span class="sourceLineNo">242</span> return true;<a name="line.242"></a> +<span class="sourceLineNo">243</span> }<a name="line.243"></a> +<span class="sourceLineNo">244</span> return false;<a name="line.244"></a> +<span class="sourceLineNo">245</span> } catch (IOException ioe) {<a name="line.245"></a> +<span class="sourceLineNo">246</span> if (logScannerActivity) {<a name="line.246"></a> +<span class="sourceLineNo">247</span> long now = System.currentTimeMillis();<a name="line.247"></a> +<span class="sourceLineNo">248</span> LOG.info("Mapper took " + (now-timestamp)<a name="line.248"></a> +<span class="sourceLineNo">249</span> + "ms to process " + rowcount + " rows");<a name="line.249"></a> +<span class="sourceLineNo">250</span> LOG.info(ioe.toString(), ioe);<a name="line.250"></a> +<span class="sourceLineNo">251</span> String lastRow = lastSuccessfulRow == null ?<a name="line.251"></a> +<span class="sourceLineNo">252</span> "null" : Bytes.toStringBinary(lastSuccessfulRow);<a name="line.252"></a> +<span class="sourceLineNo">253</span> LOG.info("lastSuccessfulRow=" + lastRow);<a name="line.253"></a> +<span class="sourceLineNo">254</span> }<a name="line.254"></a> +<span class="sourceLineNo">255</span> throw ioe;<a name="line.255"></a> +<span class="sourceLineNo">256</span> }<a name="line.256"></a> +<span class="sourceLineNo">257</span> }<a name="line.257"></a> +<span class="sourceLineNo">258</span>}<a name="line.258"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html index 6f48392..ab79d90 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html @@ -28,11 +28,11 @@ <span class="sourceLineNo">020</span><a name="line.20"></a> <span class="sourceLineNo">021</span>import java.io.IOException;<a name="line.21"></a> <span class="sourceLineNo">022</span><a name="line.22"></a> -<span class="sourceLineNo">023</span>import org.apache.commons.logging.Log;<a name="line.23"></a> -<span class="sourceLineNo">024</span>import org.apache.commons.logging.LogFactory;<a name="line.24"></a> -<span class="sourceLineNo">025</span>import org.apache.hadoop.hbase.CompareOperator;<a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.hadoop.hbase.HConstants;<a name="line.26"></a> -<span class="sourceLineNo">027</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.27"></a> +<span class="sourceLineNo">023</span>import org.apache.hadoop.hbase.CompareOperator;<a name="line.23"></a> +<span class="sourceLineNo">024</span>import org.apache.hadoop.hbase.HConstants;<a name="line.24"></a> +<span class="sourceLineNo">025</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.25"></a> +<span class="sourceLineNo">026</span>import org.slf4j.Logger;<a name="line.26"></a> +<span class="sourceLineNo">027</span>import org.slf4j.LoggerFactory;<a name="line.27"></a> <span class="sourceLineNo">028</span>import org.apache.hadoop.conf.Configuration;<a name="line.28"></a> <span class="sourceLineNo">029</span>import org.apache.hadoop.conf.Configured;<a name="line.29"></a> <span class="sourceLineNo">030</span>import org.apache.hadoop.fs.Path;<a name="line.30"></a> @@ -80,8 +80,8 @@ <span class="sourceLineNo">072</span> */<a name="line.72"></a> <span class="sourceLineNo">073</span>@InterfaceAudience.Public<a name="line.73"></a> <span class="sourceLineNo">074</span>public class CellCounter extends Configured implements Tool {<a name="line.74"></a> -<span class="sourceLineNo">075</span> private static final Log LOG =<a name="line.75"></a> -<span class="sourceLineNo">076</span> LogFactory.getLog(CellCounter.class.getName());<a name="line.76"></a> +<span class="sourceLineNo">075</span> private static final Logger LOG =<a name="line.75"></a> +<span class="sourceLineNo">076</span> LoggerFactory.getLogger(CellCounter.class.getName());<a name="line.76"></a> <span class="sourceLineNo">077</span><a name="line.77"></a> <span class="sourceLineNo">078</span><a name="line.78"></a> <span class="sourceLineNo">079</span> /**<a name="line.79"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CopyTable.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CopyTable.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CopyTable.html index 46fb016..f7c1cae 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CopyTable.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CopyTable.html @@ -31,15 +31,15 @@ <span class="sourceLineNo">023</span>import java.util.Map;<a name="line.23"></a> <span class="sourceLineNo">024</span>import java.util.Random;<a name="line.24"></a> <span class="sourceLineNo">025</span><a name="line.25"></a> -<span class="sourceLineNo">026</span>import org.apache.commons.logging.Log;<a name="line.26"></a> -<span class="sourceLineNo">027</span>import org.apache.commons.logging.LogFactory;<a name="line.27"></a> -<span class="sourceLineNo">028</span>import org.apache.hadoop.conf.Configured;<a name="line.28"></a> -<span class="sourceLineNo">029</span>import org.apache.hadoop.fs.FileSystem;<a name="line.29"></a> -<span class="sourceLineNo">030</span>import org.apache.hadoop.fs.Path;<a name="line.30"></a> -<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.HBaseConfiguration;<a name="line.31"></a> -<span class="sourceLineNo">032</span>import org.apache.hadoop.hbase.HConstants;<a name="line.32"></a> -<span class="sourceLineNo">033</span>import org.apache.hadoop.hbase.TableName;<a name="line.33"></a> -<span class="sourceLineNo">034</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.34"></a> +<span class="sourceLineNo">026</span>import org.apache.hadoop.conf.Configured;<a name="line.26"></a> +<span class="sourceLineNo">027</span>import org.apache.hadoop.fs.FileSystem;<a name="line.27"></a> +<span class="sourceLineNo">028</span>import org.apache.hadoop.fs.Path;<a name="line.28"></a> +<span class="sourceLineNo">029</span>import org.apache.hadoop.hbase.HBaseConfiguration;<a name="line.29"></a> +<span class="sourceLineNo">030</span>import org.apache.hadoop.hbase.HConstants;<a name="line.30"></a> +<span class="sourceLineNo">031</span>import org.apache.hadoop.hbase.TableName;<a name="line.31"></a> +<span class="sourceLineNo">032</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.32"></a> +<span class="sourceLineNo">033</span>import org.slf4j.Logger;<a name="line.33"></a> +<span class="sourceLineNo">034</span>import org.slf4j.LoggerFactory;<a name="line.34"></a> <span class="sourceLineNo">035</span>import org.apache.hadoop.hbase.client.Admin;<a name="line.35"></a> <span class="sourceLineNo">036</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.36"></a> <span class="sourceLineNo">037</span>import org.apache.hadoop.hbase.client.ConnectionFactory;<a name="line.37"></a> @@ -57,7 +57,7 @@ <span class="sourceLineNo">049</span> */<a name="line.49"></a> <span class="sourceLineNo">050</span>@InterfaceAudience.Public<a name="line.50"></a> <span class="sourceLineNo">051</span>public class CopyTable extends Configured implements Tool {<a name="line.51"></a> -<span class="sourceLineNo">052</span> private static final Log LOG = LogFactory.getLog(CopyTable.class);<a name="line.52"></a> +<span class="sourceLineNo">052</span> private static final Logger LOG = LoggerFactory.getLogger(CopyTable.class);<a name="line.52"></a> <span class="sourceLineNo">053</span><a name="line.53"></a> <span class="sourceLineNo">054</span> final static String NAME = "copytable";<a name="line.54"></a> <span class="sourceLineNo">055</span> long startTime = 0;<a name="line.55"></a> http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html ---------------------------------------------------------------------- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html index 63ead06..d0a626e 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html @@ -48,57 +48,57 @@ <span class="sourceLineNo">040</span>import java.util.stream.Collectors;<a name="line.40"></a> <span class="sourceLineNo">041</span><a name="line.41"></a> <span class="sourceLineNo">042</span>import org.apache.commons.lang3.StringUtils;<a name="line.42"></a> -<span class="sourceLineNo">043</span>import org.apache.commons.logging.Log;<a name="line.43"></a> -<span class="sourceLineNo">044</span>import org.apache.commons.logging.LogFactory;<a name="line.44"></a> -<span class="sourceLineNo">045</span>import org.apache.hadoop.conf.Configuration;<a name="line.45"></a> -<span class="sourceLineNo">046</span>import org.apache.hadoop.fs.FileSystem;<a name="line.46"></a> -<span class="sourceLineNo">047</span>import org.apache.hadoop.fs.Path;<a name="line.47"></a> -<span class="sourceLineNo">048</span>import org.apache.hadoop.hbase.Cell;<a name="line.48"></a> -<span class="sourceLineNo">049</span>import org.apache.hadoop.hbase.CellComparator;<a name="line.49"></a> -<span class="sourceLineNo">050</span>import org.apache.hadoop.hbase.CellUtil;<a name="line.50"></a> -<span class="sourceLineNo">051</span>import org.apache.hadoop.hbase.HConstants;<a name="line.51"></a> -<span class="sourceLineNo">052</span>import org.apache.hadoop.hbase.HRegionLocation;<a name="line.52"></a> -<span class="sourceLineNo">053</span>import org.apache.hadoop.hbase.HTableDescriptor;<a name="line.53"></a> -<span class="sourceLineNo">054</span>import org.apache.hadoop.hbase.PrivateCellUtil;<a name="line.54"></a> -<span class="sourceLineNo">055</span>import org.apache.hadoop.hbase.KeyValue;<a name="line.55"></a> -<span class="sourceLineNo">056</span>import org.apache.hadoop.hbase.TableName;<a name="line.56"></a> -<span class="sourceLineNo">057</span>import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;<a name="line.57"></a> -<span class="sourceLineNo">058</span>import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;<a name="line.58"></a> -<span class="sourceLineNo">059</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.59"></a> -<span class="sourceLineNo">060</span>import org.apache.hadoop.hbase.client.ConnectionFactory;<a name="line.60"></a> -<span class="sourceLineNo">061</span>import org.apache.hadoop.hbase.client.Put;<a name="line.61"></a> -<span class="sourceLineNo">062</span>import org.apache.hadoop.hbase.client.RegionLocator;<a name="line.62"></a> -<span class="sourceLineNo">063</span>import org.apache.hadoop.hbase.client.Table;<a name="line.63"></a> -<span class="sourceLineNo">064</span>import org.apache.hadoop.hbase.client.TableDescriptor;<a name="line.64"></a> -<span class="sourceLineNo">065</span>import org.apache.hadoop.hbase.fs.HFileSystem;<a name="line.65"></a> -<span class="sourceLineNo">066</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.66"></a> -<span class="sourceLineNo">067</span>import org.apache.hadoop.hbase.io.compress.Compression;<a name="line.67"></a> -<span class="sourceLineNo">068</span>import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;<a name="line.68"></a> -<span class="sourceLineNo">069</span>import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;<a name="line.69"></a> -<span class="sourceLineNo">070</span>import org.apache.hadoop.hbase.io.hfile.CacheConfig;<a name="line.70"></a> -<span class="sourceLineNo">071</span>import org.apache.hadoop.hbase.io.hfile.HFile;<a name="line.71"></a> -<span class="sourceLineNo">072</span>import org.apache.hadoop.hbase.io.hfile.HFileContext;<a name="line.72"></a> -<span class="sourceLineNo">073</span>import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder;<a name="line.73"></a> -<span class="sourceLineNo">074</span>import org.apache.hadoop.hbase.io.hfile.HFileWriterImpl;<a name="line.74"></a> -<span class="sourceLineNo">075</span>import org.apache.hadoop.hbase.regionserver.BloomType;<a name="line.75"></a> -<span class="sourceLineNo">076</span>import org.apache.hadoop.hbase.regionserver.HStore;<a name="line.76"></a> -<span class="sourceLineNo">077</span>import org.apache.hadoop.hbase.regionserver.StoreFileWriter;<a name="line.77"></a> -<span class="sourceLineNo">078</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.78"></a> -<span class="sourceLineNo">079</span>import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;<a name="line.79"></a> -<span class="sourceLineNo">080</span>import org.apache.hadoop.hbase.util.FSUtils;<a name="line.80"></a> -<span class="sourceLineNo">081</span>import org.apache.hadoop.hbase.util.MapReduceCell;<a name="line.81"></a> -<span class="sourceLineNo">082</span>import org.apache.hadoop.io.NullWritable;<a name="line.82"></a> -<span class="sourceLineNo">083</span>import org.apache.hadoop.io.SequenceFile;<a name="line.83"></a> -<span class="sourceLineNo">084</span>import org.apache.hadoop.io.Text;<a name="line.84"></a> -<span class="sourceLineNo">085</span>import org.apache.hadoop.mapreduce.Job;<a name="line.85"></a> -<span class="sourceLineNo">086</span>import org.apache.hadoop.mapreduce.OutputCommitter;<a name="line.86"></a> -<span class="sourceLineNo">087</span>import org.apache.hadoop.mapreduce.OutputFormat;<a name="line.87"></a> -<span class="sourceLineNo">088</span>import org.apache.hadoop.mapreduce.RecordWriter;<a name="line.88"></a> -<span class="sourceLineNo">089</span>import org.apache.hadoop.mapreduce.TaskAttemptContext;<a name="line.89"></a> -<span class="sourceLineNo">090</span>import org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter;<a name="line.90"></a> -<span class="sourceLineNo">091</span>import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;<a name="line.91"></a> -<span class="sourceLineNo">092</span>import org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner;<a name="line.92"></a> -<span class="sourceLineNo">093</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.93"></a> +<span class="sourceLineNo">043</span>import org.apache.hadoop.conf.Configuration;<a name="line.43"></a> +<span class="sourceLineNo">044</span>import org.apache.hadoop.fs.FileSystem;<a name="line.44"></a> +<span class="sourceLineNo">045</span>import org.apache.hadoop.fs.Path;<a name="line.45"></a> +<span class="sourceLineNo">046</span>import org.apache.hadoop.hbase.Cell;<a name="line.46"></a> +<span class="sourceLineNo">047</span>import org.apache.hadoop.hbase.CellComparator;<a name="line.47"></a> +<span class="sourceLineNo">048</span>import org.apache.hadoop.hbase.CellUtil;<a name="line.48"></a> +<span class="sourceLineNo">049</span>import org.apache.hadoop.hbase.HConstants;<a name="line.49"></a> +<span class="sourceLineNo">050</span>import org.apache.hadoop.hbase.HRegionLocation;<a name="line.50"></a> +<span class="sourceLineNo">051</span>import org.apache.hadoop.hbase.HTableDescriptor;<a name="line.51"></a> +<span class="sourceLineNo">052</span>import org.apache.hadoop.hbase.PrivateCellUtil;<a name="line.52"></a> +<span class="sourceLineNo">053</span>import org.apache.hadoop.hbase.KeyValue;<a name="line.53"></a> +<span class="sourceLineNo">054</span>import org.apache.hadoop.hbase.TableName;<a name="line.54"></a> +<span class="sourceLineNo">055</span>import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;<a name="line.55"></a> +<span class="sourceLineNo">056</span>import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;<a name="line.56"></a> +<span class="sourceLineNo">057</span>import org.apache.hadoop.hbase.client.Connection;<a name="line.57"></a> +<span class="sourceLineNo">058</span>import org.apache.hadoop.hbase.client.ConnectionFactory;<a name="line.58"></a> +<span class="sourceLineNo">059</span>import org.apache.hadoop.hbase.client.Put;<a name="line.59"></a> +<span class="sourceLineNo">060</span>import org.apache.hadoop.hbase.client.RegionLocator;<a name="line.60"></a> +<span class="sourceLineNo">061</span>import org.apache.hadoop.hbase.client.Table;<a name="line.61"></a> +<span class="sourceLineNo">062</span>import org.apache.hadoop.hbase.client.TableDescriptor;<a name="line.62"></a> +<span class="sourceLineNo">063</span>import org.apache.hadoop.hbase.fs.HFileSystem;<a name="line.63"></a> +<span class="sourceLineNo">064</span>import org.apache.hadoop.hbase.io.ImmutableBytesWritable;<a name="line.64"></a> +<span class="sourceLineNo">065</span>import org.apache.hadoop.hbase.io.compress.Compression;<a name="line.65"></a> +<span class="sourceLineNo">066</span>import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;<a name="line.66"></a> +<span class="sourceLineNo">067</span>import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;<a name="line.67"></a> +<span class="sourceLineNo">068</span>import org.apache.hadoop.hbase.io.hfile.CacheConfig;<a name="line.68"></a> +<span class="sourceLineNo">069</span>import org.apache.hadoop.hbase.io.hfile.HFile;<a name="line.69"></a> +<span class="sourceLineNo">070</span>import org.apache.hadoop.hbase.io.hfile.HFileContext;<a name="line.70"></a> +<span class="sourceLineNo">071</span>import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder;<a name="line.71"></a> +<span class="sourceLineNo">072</span>import org.apache.hadoop.hbase.io.hfile.HFileWriterImpl;<a name="line.72"></a> +<span class="sourceLineNo">073</span>import org.apache.hadoop.hbase.regionserver.BloomType;<a name="line.73"></a> +<span class="sourceLineNo">074</span>import org.apache.hadoop.hbase.regionserver.HStore;<a name="line.74"></a> +<span class="sourceLineNo">075</span>import org.apache.hadoop.hbase.regionserver.StoreFileWriter;<a name="line.75"></a> +<span class="sourceLineNo">076</span>import org.apache.hadoop.hbase.util.Bytes;<a name="line.76"></a> +<span class="sourceLineNo">077</span>import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;<a name="line.77"></a> +<span class="sourceLineNo">078</span>import org.apache.hadoop.hbase.util.FSUtils;<a name="line.78"></a> +<span class="sourceLineNo">079</span>import org.apache.hadoop.hbase.util.MapReduceCell;<a name="line.79"></a> +<span class="sourceLineNo">080</span>import org.apache.hadoop.io.NullWritable;<a name="line.80"></a> +<span class="sourceLineNo">081</span>import org.apache.hadoop.io.SequenceFile;<a name="line.81"></a> +<span class="sourceLineNo">082</span>import org.apache.hadoop.io.Text;<a name="line.82"></a> +<span class="sourceLineNo">083</span>import org.apache.hadoop.mapreduce.Job;<a name="line.83"></a> +<span class="sourceLineNo">084</span>import org.apache.hadoop.mapreduce.OutputCommitter;<a name="line.84"></a> +<span class="sourceLineNo">085</span>import org.apache.hadoop.mapreduce.OutputFormat;<a name="line.85"></a> +<span class="sourceLineNo">086</span>import org.apache.hadoop.mapreduce.RecordWriter;<a name="line.86"></a> +<span class="sourceLineNo">087</span>import org.apache.hadoop.mapreduce.TaskAttemptContext;<a name="line.87"></a> +<span class="sourceLineNo">088</span>import org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter;<a name="line.88"></a> +<span class="sourceLineNo">089</span>import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;<a name="line.89"></a> +<span class="sourceLineNo">090</span>import org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner;<a name="line.90"></a> +<span class="sourceLineNo">091</span>import org.apache.yetus.audience.InterfaceAudience;<a name="line.91"></a> +<span class="sourceLineNo">092</span>import org.slf4j.Logger;<a name="line.92"></a> +<span class="sourceLineNo">093</span>import org.slf4j.LoggerFactory;<a name="line.93"></a> <span class="sourceLineNo">094</span>import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;<a name="line.94"></a> <span class="sourceLineNo">095</span><a name="line.95"></a> <span class="sourceLineNo">096</span>/**<a name="line.96"></a> @@ -113,7 +113,7 @@ <span class="sourceLineNo">105</span>@InterfaceAudience.Public<a name="line.105"></a> <span class="sourceLineNo">106</span>public class HFileOutputFormat2<a name="line.106"></a> <span class="sourceLineNo">107</span> extends FileOutputFormat<ImmutableBytesWritable, Cell> {<a name="line.107"></a> -<span class="sourceLineNo">108</span> private static final Log LOG = LogFactory.getLog(HFileOutputFormat2.class);<a name="line.108"></a> +<span class="sourceLineNo">108</span> private static final Logger LOG = LoggerFactory.getLogger(HFileOutputFormat2.class);<a name="line.108"></a> <span class="sourceLineNo">109</span> static class TableInfo {<a name="line.109"></a> <span class="sourceLineNo">110</span> private TableDescriptor tableDesctiptor;<a name="line.110"></a> <span class="sourceLineNo">111</span> private RegionLocator regionLocator;<a name="line.111"></a>
