Re: hbase.master parameter?
That property hasn't been in effect since 0.90 (far as I can remember). Ever since we switched master discovery to ZK, the property has been obsolete. On Fri, Oct 4, 2013 at 5:13 AM, Jay Vyas jayunit...@gmail.com wrote: What happened to the hbase.master parameter? I dont see it in the docs... was it deprecated? It appears to still have an effect in 94.7 -- Jay Vyas http://jayunit100.blogspot.com -- Harsh J
Re: Spatial data posting in HBase
If you mean insert and query spatial data, look at algorithms that are distributed databases compliant : geohash, z-index, voronoi diagram... Well, that makes me want to write a blog article about these topics :) On Tue, Sep 24, 2013 at 3:43 PM, Ted Yu yuzhih...@gmail.com wrote: There're plenty of examples in unit tests. e.g. : Put put = new Put(Bytes.toBytes(row + String.format(%1$04d, i))); put.add(family, null, value); table.put(put); value can be obtained through Bytes.toBytes(). table is an HTable. Cheers On Tue, Sep 24, 2013 at 4:15 AM, cto ankur...@tcs.com wrote: Hi , I am very new in HBase. Could you please let me know , how to insert spatial data (Latitude / Longitude) in HBase using Java . -- View this message in context: http://apache-hbase.679495.n3.nabble.com/Spatial-data-posting-in-HBase-tp4051123.html Sent from the HBase User mailing list archive at Nabble.com. -- Adrien Mogenet http://www.borntosegfault.com
Newly added regionserver is not severing requests
I'm setting up a Hbase cluster on a cloud infrastructure. HBase version: 0.94.11 Hadoop version: 1.0.4 Currently I have 4 nodes in my cluster (1 master, 3 regionservers) and I'm using YCSB (yahoo benchmarks) to create a table (500.000 rows) and send requests (Asynchronous requests). Everything works fine with this setup (as I'm monitoring the hole process with ganglia and I'm getting lamda, throughput, latency combined with the YCSB's output), but the problem occurs when I add a new regionserver on-the-fly as it doesn't getting any requests. What on-the-fly means: While the YCSB is sending request to the cluster, I'm adding new regionservers using python scripts. Addition Process (while the cluster is serving requests): 1. I'm creating a new VM which will act as the new regionserver and configure every needed aspect (hbase, hadoop, /etc/host, connect to private network, etc) 2. Stoping **hbase** balancer 3. Configuring every node in the cluster with the new node's information * adding hostname to regioservers files * adding hostname to hadoop's slave file * adding hostname and IP to /etc/host file of every node * etc 4. Executing on the master node: * `hadoop/bin/start-dfs.sh` * `hadoop/bin/start-mapred.sh` * `hbase/bin/start-hbase.sh` (I've also tried to run `hbase start regionserver` on the newly added node and does exactly the same with the last command - starts the regionserver) 5. Once the newly added node is up and running I'm executing **hadoop** load balancer 6. When the hadoop load balancer stops I'm starting again the **hbase** load balancer I'm connecting over ssh to the master node and check that the load balancers (hbase/hadoop) did their job as both the blocks and regions are uniformly spread across all the regionservers/slaves including the new one. But when I run status 'simple' in the hbase shell I see that the new regionservers are not getting any requests. (below is the output of the command after adding 2 new regionserver okeanos-nodes-4/5) |hbase(main):008:0 status 'simple' 5 live servers okeanos-nodes-1:60020 1380865800330 requestsPerSecond=5379, numberOfOnlineRegions=4, usedHeapMB=175, maxHeapMB=3067 okeanos-nodes-2:60020 1380865800738 requestsPerSecond=5674, numberOfOnlineRegions=4, usedHeapMB=161, maxHeapMB=3067 okeanos-nodes-5:60020 1380867725605 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=27, maxHeapMB=3067 okeanos-nodes-3:60020 1380865800162 requestsPerSecond=3871, numberOfOnlineRegions=5, usedHeapMB=162, maxHeapMB=3067 okeanos-nodes-4:60020 1380866702216 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=29, maxHeapMB=3067 0 dead servers Aggregate load: 14924, regions: 19| The fact that they don't serve any requests is also evidenced by the CPU usage, in a serving regionserver is about 70% while in these 2 regioservers is about 2%. Below is the output of|hadoop dfsadmin -report|, as you can see the block are evenly distributed (according to|hadoop balancer -threshold 2|). |root@okeanos-nodes-master:~# /opt/hadoop-1.0.4/bin/hadoop dfsadmin -report Configured Capacity: 105701683200 (98.44 GB) Present Capacity: 86440648704 (80.5 GB) DFS Remaining: 84188446720 (78.41 GB) DFS Used: 2252201984 (2.1 GB) DFS Used%: 2.61% Under replicated blocks: 0 Blocks with corrupt replicas: 0 Missing blocks: 0 - Datanodes available: 5 (5 total, 0 dead) Name: 10.0.0.11:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 309166080 (294.84 MB) Non DFS Used: 3851579392 (3.59 GB) DFS Remaining: 16979591168(15.81 GB) DFS Used%: 1.46% DFS Remaining%: 80.32% Last contact: Fri Oct 04 11:30:31 EEST 2013 Name: 10.0.0.3:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 531652608 (507.02 MB) Non DFS Used: 3852300288 (3.59 GB) DFS Remaining: 16756383744(15.61 GB) DFS Used%: 2.51% DFS Remaining%: 79.26% Last contact: Fri Oct 04 11:30:32 EEST 2013 Name: 10.0.0.5:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 502910976 (479.61 MB) Non DFS Used: 3853029376 (3.59 GB) DFS Remaining: 16784396288(15.63 GB) DFS Used%: 2.38% DFS Remaining%: 79.4% Last contact: Fri Oct 04 11:30:32 EEST 2013 Name: 10.0.0.4:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 421974016 (402.43 MB) Non DFS Used: 3852365824 (3.59 GB) DFS Remaining: 16865996800(15.71 GB) DFS Used%: 2% DFS Remaining%: 79.78% Last contact: Fri Oct 04 11:30:29 EEST 2013 Name: 10.0.0.10:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 486498304 (463.96 MB) Non DFS Used: 3851759616 (3.59 GB) DFS Remaining: 16802078720(15.65 GB) DFS Used%: 2.3% DFS Remaining%: 79.48% Last contact: Fri Oct 04 11:30:29 EEST 2013| I've tried stopping YCSB,
Re: Newly added regionserver is not severing requests
One possibility could be that the regions got balanced after the write load is complete. That means, when the regions were being written they were with one RS and once that is done, the region got assigned to the idle RS. Are you sure you are that YCSB writes to the regions after balancing too? Also you can run your benchmark now (after regions are balanced) and write some data to the regions on idle RS and see if it increases the request count. On Fri, Oct 4, 2013 at 2:37 PM, Thanasis Naskos anas...@csd.auth.gr wrote: I'm setting up a Hbase cluster on a cloud infrastructure. HBase version: 0.94.11 Hadoop version: 1.0.4 Currently I have 4 nodes in my cluster (1 master, 3 regionservers) and I'm using YCSB (yahoo benchmarks) to create a table (500.000 rows) and send requests (Asynchronous requests). Everything works fine with this setup (as I'm monitoring the hole process with ganglia and I'm getting lamda, throughput, latency combined with the YCSB's output), but the problem occurs when I add a new regionserver on-the-fly as it doesn't getting any requests. What on-the-fly means: While the YCSB is sending request to the cluster, I'm adding new regionservers using python scripts. Addition Process (while the cluster is serving requests): 1. I'm creating a new VM which will act as the new regionserver and configure every needed aspect (hbase, hadoop, /etc/host, connect to private network, etc) 2. Stoping **hbase** balancer 3. Configuring every node in the cluster with the new node's information * adding hostname to regioservers files * adding hostname to hadoop's slave file * adding hostname and IP to /etc/host file of every node * etc 4. Executing on the master node: * `hadoop/bin/start-dfs.sh` * `hadoop/bin/start-mapred.sh` * `hbase/bin/start-hbase.sh` (I've also tried to run `hbase start regionserver` on the newly added node and does exactly the same with the last command - starts the regionserver) 5. Once the newly added node is up and running I'm executing **hadoop** load balancer 6. When the hadoop load balancer stops I'm starting again the **hbase** load balancer I'm connecting over ssh to the master node and check that the load balancers (hbase/hadoop) did their job as both the blocks and regions are uniformly spread across all the regionservers/slaves including the new one. But when I run status 'simple' in the hbase shell I see that the new regionservers are not getting any requests. (below is the output of the command after adding 2 new regionserver okeanos-nodes-4/5) |hbase(main):008:0 status 'simple' 5 live servers okeanos-nodes-1:60020 1380865800330 requestsPerSecond=5379, numberOfOnlineRegions=4, usedHeapMB=175, maxHeapMB=3067 okeanos-nodes-2:60020 1380865800738 requestsPerSecond=5674, numberOfOnlineRegions=4, usedHeapMB=161, maxHeapMB=3067 okeanos-nodes-5:60020 1380867725605 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=27, maxHeapMB=3067 okeanos-nodes-3:60020 1380865800162 requestsPerSecond=3871, numberOfOnlineRegions=5, usedHeapMB=162, maxHeapMB=3067 okeanos-nodes-4:60020 1380866702216 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=29, maxHeapMB=3067 0 dead servers Aggregate load: 14924, regions: 19| The fact that they don't serve any requests is also evidenced by the CPU usage, in a serving regionserver is about 70% while in these 2 regioservers is about 2%. Below is the output of|hadoop dfsadmin -report|, as you can see the block are evenly distributed (according to|hadoop balancer -threshold 2|). |root@okeanos-nodes-master:~# /opt/hadoop-1.0.4/bin/hadoop dfsadmin -report Configured Capacity: 105701683200 (98.44 GB) Present Capacity: 86440648704 (80.5 GB) DFS Remaining: 84188446720 (78.41 GB) DFS Used: 2252201984 (2.1 GB) DFS Used%: 2.61% Under replicated blocks: 0 Blocks with corrupt replicas: 0 Missing blocks: 0 --**--- Datanodes available: 5 (5 total, 0 dead) Name: 10.0.0.11:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 309166080 (294.84 MB) Non DFS Used: 3851579392 (3.59 GB) DFS Remaining: 16979591168(15.81 GB) DFS Used%: 1.46% DFS Remaining%: 80.32% Last contact: Fri Oct 04 11:30:31 EEST 2013 Name: 10.0.0.3:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 531652608 (507.02 MB) Non DFS Used: 3852300288 (3.59 GB) DFS Remaining: 16756383744(15.61 GB) DFS Used%: 2.51% DFS Remaining%: 79.26% Last contact: Fri Oct 04 11:30:32 EEST 2013 Name: 10.0.0.5:50010 Decommission Status : Normal Configured Capacity: 21140336640 (19.69 GB) DFS Used: 502910976 (479.61 MB) Non DFS Used: 3853029376 (3.59 GB) DFS Remaining: 16784396288(15.63 GB) DFS Used%: 2.38% DFS Remaining%: 79.4% Last contact: Fri Oct 04
Re: Newly added regionserver is not severing requests
One possibility could be that the regions got balanced after the write load is complete. That means, when the regions were being written they were with one RS and once that is done, the region got assigned to the idle RS. I think that this is the case, but why is this wrong? I write the data to the database with 3 RS's and when the write load is finished I add one more RS and run hadoop and hbase load balancers to assign some data and regions (respectively) to this new node (without adding new data) Shouldn't this work? Are you sure you are that YCSB writes to the regions after balancing too? I should have mentioned that once the data is written to the RS's (3 RS's), YCSB sends only READ requests and doesn't write/insert/update anything else to the database even after new nodes (RS's) are added. Also you can run your benchmark now (after regions are balanced) and write some data to the regions on idle RS and see if it increases the request count. I've tried to add (put) a new row to the database from inside the idle RS (shell) and the row was inserted properly (I've checked it with get )... but as expected nothing changed still I have 2 RS's idle Thank you for your interest!! On 10/04/2013 12:45 PM, Bharath Vissapragada wrote: One possibility could be that the regions got balanced after the write load is complete. That means, when the regions were being written they were with one RS and once that is done, the region got assigned to the idle RS. Are you sure you are that YCSB writes to the regions after balancing too? Also you can run your benchmark now (after regions are balanced) and write some data to the regions on idle RS and see if it increases the request count. On Fri, Oct 4, 2013 at 2:37 PM, Thanasis Naskos anas...@csd.auth.gr wrote: I'm setting up a Hbase cluster on a cloud infrastructure. HBase version: 0.94.11 Hadoop version: 1.0.4 Currently I have 4 nodes in my cluster (1 master, 3 regionservers) and I'm using YCSB (yahoo benchmarks) to create a table (500.000 rows) and send requests (Asynchronous requests). Everything works fine with this setup (as I'm monitoring the hole process with ganglia and I'm getting lamda, throughput, latency combined with the YCSB's output), but the problem occurs when I add a new regionserver on-the-fly as it doesn't getting any requests. What on-the-fly means: While the YCSB is sending request to the cluster, I'm adding new regionservers using python scripts. Addition Process (while the cluster is serving requests): 1. I'm creating a new VM which will act as the new regionserver and configure every needed aspect (hbase, hadoop, /etc/host, connect to private network, etc) 2. Stoping **hbase** balancer 3. Configuring every node in the cluster with the new node's information * adding hostname to regioservers files * adding hostname to hadoop's slave file * adding hostname and IP to /etc/host file of every node * etc 4. Executing on the master node: * `hadoop/bin/start-dfs.sh` * `hadoop/bin/start-mapred.sh` * `hbase/bin/start-hbase.sh` (I've also tried to run `hbase start regionserver` on the newly added node and does exactly the same with the last command - starts the regionserver) 5. Once the newly added node is up and running I'm executing **hadoop** load balancer 6. When the hadoop load balancer stops I'm starting again the **hbase** load balancer I'm connecting over ssh to the master node and check that the load balancers (hbase/hadoop) did their job as both the blocks and regions are uniformly spread across all the regionservers/slaves including the new one. But when I run status 'simple' in the hbase shell I see that the new regionservers are not getting any requests. (below is the output of the command after adding 2 new regionserver okeanos-nodes-4/5) |hbase(main):008:0 status 'simple' 5 live servers okeanos-nodes-1:60020 1380865800330 requestsPerSecond=5379, numberOfOnlineRegions=4, usedHeapMB=175, maxHeapMB=3067 okeanos-nodes-2:60020 1380865800738 requestsPerSecond=5674, numberOfOnlineRegions=4, usedHeapMB=161, maxHeapMB=3067 okeanos-nodes-5:60020 1380867725605 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=27, maxHeapMB=3067 okeanos-nodes-3:60020 1380865800162 requestsPerSecond=3871, numberOfOnlineRegions=5, usedHeapMB=162, maxHeapMB=3067 okeanos-nodes-4:60020 1380866702216 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=29, maxHeapMB=3067 0 dead servers Aggregate load: 14924, regions: 19| The fact that they don't serve any requests is also evidenced by the CPU usage, in a serving regionserver is about 70% while in these 2 regioservers is about 2%. Below is the output of|hadoop dfsadmin -report|, as you can see the block are evenly distributed (according to|hadoop balancer -threshold 2|). |root@okeanos-nodes-master:~# /opt/hadoop-1.0.4/bin/hadoop
Re: Newly added regionserver is not severing requests
[SOLVED] I find what was going on and it had nothing to do with Hbase... I have forgotten to add the hostname and IP of the new RS to the YCSB server VM :-( Thanks again Bharath for your interest On 10/04/2013 02:18 PM, Thanasis Naskos wrote: One possibility could be that the regions got balanced after the write load is complete. That means, when the regions were being written they were with one RS and once that is done, the region got assigned to the idle RS. I think that this is the case, but why is this wrong? I write the data to the database with 3 RS's and when the write load is finished I add one more RS and run hadoop and hbase load balancers to assign some data and regions (respectively) to this new node (without adding new data) Shouldn't this work? Are you sure you are that YCSB writes to the regions after balancing too? I should have mentioned that once the data is written to the RS's (3 RS's), YCSB sends only READ requests and doesn't write/insert/update anything else to the database even after new nodes (RS's) are added. Also you can run your benchmark now (after regions are balanced) and write some data to the regions on idle RS and see if it increases the request count. I've tried to add (put) a new row to the database from inside the idle RS (shell) and the row was inserted properly (I've checked it with get )... but as expected nothing changed still I have 2 RS's idle Thank you for your interest!! On 10/04/2013 12:45 PM, Bharath Vissapragada wrote: One possibility could be that the regions got balanced after the write load is complete. That means, when the regions were being written they were with one RS and once that is done, the region got assigned to the idle RS. Are you sure you are that YCSB writes to the regions after balancing too? Also you can run your benchmark now (after regions are balanced) and write some data to the regions on idle RS and see if it increases the request count. On Fri, Oct 4, 2013 at 2:37 PM, Thanasis Naskos anas...@csd.auth.gr wrote: I'm setting up a Hbase cluster on a cloud infrastructure. HBase version: 0.94.11 Hadoop version: 1.0.4 Currently I have 4 nodes in my cluster (1 master, 3 regionservers) and I'm using YCSB (yahoo benchmarks) to create a table (500.000 rows) and send requests (Asynchronous requests). Everything works fine with this setup (as I'm monitoring the hole process with ganglia and I'm getting lamda, throughput, latency combined with the YCSB's output), but the problem occurs when I add a new regionserver on-the-fly as it doesn't getting any requests. What on-the-fly means: While the YCSB is sending request to the cluster, I'm adding new regionservers using python scripts. Addition Process (while the cluster is serving requests): 1. I'm creating a new VM which will act as the new regionserver and configure every needed aspect (hbase, hadoop, /etc/host, connect to private network, etc) 2. Stoping **hbase** balancer 3. Configuring every node in the cluster with the new node's information * adding hostname to regioservers files * adding hostname to hadoop's slave file * adding hostname and IP to /etc/host file of every node * etc 4. Executing on the master node: * `hadoop/bin/start-dfs.sh` * `hadoop/bin/start-mapred.sh` * `hbase/bin/start-hbase.sh` (I've also tried to run `hbase start regionserver` on the newly added node and does exactly the same with the last command - starts the regionserver) 5. Once the newly added node is up and running I'm executing **hadoop** load balancer 6. When the hadoop load balancer stops I'm starting again the **hbase** load balancer I'm connecting over ssh to the master node and check that the load balancers (hbase/hadoop) did their job as both the blocks and regions are uniformly spread across all the regionservers/slaves including the new one. But when I run status 'simple' in the hbase shell I see that the new regionservers are not getting any requests. (below is the output of the command after adding 2 new regionserver okeanos-nodes-4/5) |hbase(main):008:0 status 'simple' 5 live servers okeanos-nodes-1:60020 1380865800330 requestsPerSecond=5379, numberOfOnlineRegions=4, usedHeapMB=175, maxHeapMB=3067 okeanos-nodes-2:60020 1380865800738 requestsPerSecond=5674, numberOfOnlineRegions=4, usedHeapMB=161, maxHeapMB=3067 okeanos-nodes-5:60020 1380867725605 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=27, maxHeapMB=3067 okeanos-nodes-3:60020 1380865800162 requestsPerSecond=3871, numberOfOnlineRegions=5, usedHeapMB=162, maxHeapMB=3067 okeanos-nodes-4:60020 1380866702216 requestsPerSecond=0, numberOfOnlineRegions=3, usedHeapMB=29, maxHeapMB=3067 0 dead servers Aggregate load: 14924, regions: 19| The fact that they don't serve any requests is also evidenced by the CPU usage, in a
Re: hbase.master parameter?
Thanks for the feedback ! So - are you sure it has no effect? By obsolete - do we mean deprecated? In this case - which parameters have replaced it and how [specifically] ? Any help on this issue would be appreciated , because im seeing an effect when i have the parameter , will check my hbase version and confirm. On Fri, Oct 4, 2013 at 2:05 AM, Harsh J ha...@cloudera.com wrote: That property hasn't been in effect since 0.90 (far as I can remember). Ever since we switched master discovery to ZK, the property has been obsolete. On Fri, Oct 4, 2013 at 5:13 AM, Jay Vyas jayunit...@gmail.com wrote: What happened to the hbase.master parameter? I dont see it in the docs... was it deprecated? It appears to still have an effect in 94.7 -- Jay Vyas http://jayunit100.blogspot.com -- Harsh J -- Jay Vyas http://jayunit100.blogspot.com
Re: hbase.master parameter?
Oh wow. looking in the source this really is an old parameter: It appears that the answer to my question is : that these are the master parameters: src/main/resources/hbase-default.xml:namehbase.master.port/name src/main/resources/hbase-default.xml:namehbase.master.info.port/name src/main/resources/hbase-default.xml: namehbase.master.info.bindAddress/name src/main/resources/hbase-default.xml: namehbase.master.dns.interface/name src/main/resources/hbase-default.xml: namehbase.master.dns.nameserver/name src/main/resources/hbase-default.xml: namehbase.master.logcleaner.ttl/name src/main/resources/hbase-default.xml: namehbase.master.logcleaner.plugins/name src/main/resources/hbase-default.xml: valueorg.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner/value src/main/resources/hbase-default.xml: namehbase.master.keytab.file/name src/main/resources/hbase-default.xml: namehbase.master.kerberos.principal/name src/main/resources/hbase-default.xml: namehbase.master.hfilecleaner.plugins/name src/main/resources/hbase-default.xml: valueorg.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner/value src/test/resources/hbase-site.xml: namehbase.master.event.waiting.time/name src/test/resources/hbase-site.xml:namehbase.master.info.port/name src/test/resources/hbase-site.xml:descriptionThe port for the hbase master web UI src/test/resources/hbase-site.xml: namehbase.master.lease.thread.wakefrequency/name On Fri, Oct 4, 2013 at 11:06 AM, Jay Vyas jayunit...@gmail.com wrote: Thanks for the feedback ! So - are you sure it has no effect? By obsolete - do we mean deprecated? In this case - which parameters have replaced it and how [specifically] ? Any help on this issue would be appreciated , because im seeing an effect when i have the parameter , will check my hbase version and confirm. On Fri, Oct 4, 2013 at 2:05 AM, Harsh J ha...@cloudera.com wrote: That property hasn't been in effect since 0.90 (far as I can remember). Ever since we switched master discovery to ZK, the property has been obsolete. On Fri, Oct 4, 2013 at 5:13 AM, Jay Vyas jayunit...@gmail.com wrote: What happened to the hbase.master parameter? I dont see it in the docs... was it deprecated? It appears to still have an effect in 94.7 -- Jay Vyas http://jayunit100.blogspot.com -- Harsh J -- Jay Vyas http://jayunit100.blogspot.com -- Jay Vyas http://jayunit100.blogspot.com
Re: hbase.master parameter?
hbase.master was removed when we added zookeeper, so now a client will do a lookup in ZK instead of talking to a pre-determined master. So in a way, hbase.zookeeper.quorum is what replaces hbase.master FWIW that was done in 0.20.0 which was released in September of 2009, so hbase.master has been removed 4 years ago. J-D On Fri, Oct 4, 2013 at 8:11 AM, Jay Vyas jayunit...@gmail.com wrote: Oh wow. looking in the source this really is an old parameter: It appears that the answer to my question is : that these are the master parameters: src/main/resources/hbase-default.xml:namehbase.master.port/name src/main/resources/hbase-default.xml: namehbase.master.info.port/name src/main/resources/hbase-default.xml: namehbase.master.info.bindAddress/name src/main/resources/hbase-default.xml: namehbase.master.dns.interface/name src/main/resources/hbase-default.xml: namehbase.master.dns.nameserver/name src/main/resources/hbase-default.xml: namehbase.master.logcleaner.ttl/name src/main/resources/hbase-default.xml: namehbase.master.logcleaner.plugins/name src/main/resources/hbase-default.xml: valueorg.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner/value src/main/resources/hbase-default.xml: namehbase.master.keytab.file/name src/main/resources/hbase-default.xml: namehbase.master.kerberos.principal/name src/main/resources/hbase-default.xml: namehbase.master.hfilecleaner.plugins/name src/main/resources/hbase-default.xml: valueorg.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner/value src/test/resources/hbase-site.xml: namehbase.master.event.waiting.time/name src/test/resources/hbase-site.xml:namehbase.master.info.port/name src/test/resources/hbase-site.xml:descriptionThe port for the hbase master web UI src/test/resources/hbase-site.xml: namehbase.master.lease.thread.wakefrequency/name On Fri, Oct 4, 2013 at 11:06 AM, Jay Vyas jayunit...@gmail.com wrote: Thanks for the feedback ! So - are you sure it has no effect? By obsolete - do we mean deprecated? In this case - which parameters have replaced it and how [specifically] ? Any help on this issue would be appreciated , because im seeing an effect when i have the parameter , will check my hbase version and confirm. On Fri, Oct 4, 2013 at 2:05 AM, Harsh J ha...@cloudera.com wrote: That property hasn't been in effect since 0.90 (far as I can remember). Ever since we switched master discovery to ZK, the property has been obsolete. On Fri, Oct 4, 2013 at 5:13 AM, Jay Vyas jayunit...@gmail.com wrote: What happened to the hbase.master parameter? I dont see it in the docs... was it deprecated? It appears to still have an effect in 94.7 -- Jay Vyas http://jayunit100.blogspot.com -- Harsh J -- Jay Vyas http://jayunit100.blogspot.com -- Jay Vyas http://jayunit100.blogspot.com
Error starting Hbase 95 in Hadoop 2.0
Hi, The hdfs and zookeeper are both up running. When I try to start the HBase, it failed with this error below. I even tried to change the security.group.mpapping to ShellBasedUnixGroupsMappingWithFallback, still not working. Anyone know what could went wrong? Caused by: java.lang.UnsatisfiedLinkError: org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V at org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native Method) at org.apache.hadoop.security.JniBasedUnixGroupsMapping.clinit(JniBasedUnixGroupsMapping.java:49) at org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback.init(JniBasedUnixGroupsMappingWithFallback.java:38) ... 33 more Thanks Tian-Ying
Re: Error starting Hbase 95 in Hadoop 2.0
See http://search-hadoop.com/m/3NMzFay1Ha1 On Fri, Oct 4, 2013 at 2:44 PM, Tianying Chang tich...@ebaysf.com wrote: Hi, The hdfs and zookeeper are both up running. When I try to start the HBase, it failed with this error below. I even tried to change the security.group.mpapping to ShellBasedUnixGroupsMappingWithFallback, still not working. Anyone know what could went wrong? Caused by: java.lang.UnsatisfiedLinkError: org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V at org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native Method) at org.apache.hadoop.security.JniBasedUnixGroupsMapping.clinit(JniBasedUnixGroupsMapping.java:49) at org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback.init(JniBasedUnixGroupsMappingWithFallback.java:38) ... 33 more Thanks Tian-Ying