Thanks. This change did make the hbase cluster start. I ran into one issue that I will document in a separate email.
On Thu, Dec 22, 2011 at 3:14 PM, Andrei Savu <[email protected]> wrote: > Oops! This instance type needs a 32bit OS. This should work: > > whirr.hardware-id=m1.small > whirr.image-id=us-east-1/ami-6936fb00 > > -- Andrei > > > On Thu, Dec 22, 2011 at 10:11 PM, Andrei Savu <[email protected]> wrote: >> >> BTW I select this AMI from http://cloud.ubuntu.com/ami/ - Ubuntu 10.04.3 >> LTS >> >> >> On Thu, Dec 22, 2011 at 10:09 PM, Andrei Savu <[email protected]> >> wrote: >>> >>> That looks strange! It seems like it's trying to select a Windows AMI !? >>> >>> Is this Whirr 0.7.0 (just released yesterday)? >>> >>> Can you add the following two lines? >>> >>> whirr.hardware-id=m1.small >>> whirr.image-id=us-east-1/ami-1136fb78 >>> >>> I am also trying that now. >>> >>> -- Andrei Savu / andreisavu.ro >>> >>> >>> On Thu, Dec 22, 2011 at 10:04 PM, David Medinets >>> <[email protected]> wrote: >>>> >>>> Sure. I'll try as many times as needed. I'd love to help. This is my >>>> current hbase-ec2.properties file: >>>> >>>> whirr.cluster-name=hbase >>>> whirr.instance-templates=1 >>>> zookeeper+hadoop-namenode+hadoop-jobtracker+hbase-master,2 >>>> hadoop-datanode+hadoop-tasktracker+hbase-regionserver >>>> whirr.provider=aws-ec2 >>>> whirr.identity=${env:AWS_ACCESS_KEY_ID} >>>> whirr.credential=${env:AWS_SECRET_ACCESS_KEY} >>>> >>>> whirr.hbase.tarball.url=http://archive.apache.org/dist/hbase/hbase-0.89.20100924/hbase-0.89.20100924-bin.tar.gz >>>> >>>> This is the command I ran: >>>> >>>> $ whirr launch-cluster --config hbase-ec2.properties >>>> >>>> And this is the output: >>>> >>>> Bootstrapping cluster >>>> Configuring template >>>> Starting 1 node(s) with roles [zookeeper, hadoop-namenode, >>>> hadoop-jobtracker, hbase-master] >>>> Configuring template >>>> Unexpected error while starting 1 nodes, minimum 1 nodes for >>>> [zookeeper, hadoop-namenode, hadoop-jobtracker, hbase-master] of >>>> cluster hbase >>>> java.util.concurrent.ExecutionException: >>>> org.jclouds.http.HttpResponseException: command: POST >>>> https://ec2.us-east-1.amazonaws.com/ HTTP/1.1 failed with response: >>>> HTTP/1.1 400 Bad Request; content: [Non-Windows AMIs with a >>>> virtualization type of 'hvm' currently may only be used with Cluster >>>> Compute instance types.] >>>> at >>>> java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:252) >>>> at java.util.concurrent.FutureTask.get(FutureTask.java:111) >>>> at >>>> org.apache.whirr.actions.BootstrapClusterAction$StartupProcess.waitForOutcomes(BootstrapClusterAction.java:298) >>>> at >>>> org.apache.whirr.actions.BootstrapClusterAction$StartupProcess.call(BootstrapClusterAction.java:251) >>>> at >>>> org.apache.whirr.actions.BootstrapClusterAction$StartupProcess.call(BootstrapClusterAction.java:212) >>>> at >>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:166) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) >>>> at java.lang.Thread.run(Thread.java:679) >>>> Caused by: org.jclouds.http.HttpResponseException: command: POST >>>> https://ec2.us-east-1.amazonaws.com/ HTTP/1.1 failed with response: >>>> HTTP/1.1 400 Bad Request; content: [Non-Windows AMIs with a >>>> virtualization type of 'hvm' currently may only be used with Cluster >>>> Compute instance types.] >>>> at >>>> org.jclouds.aws.handlers.ParseAWSErrorFromXmlContent.handleError(ParseAWSErrorFromXmlContent.java:74) >>>> at >>>> org.jclouds.http.handlers.DelegatingErrorHandler.handleError(DelegatingErrorHandler.java:69) >>>> at >>>> org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.shouldContinue(BaseHttpCommandExecutorService.java:200) >>>> at >>>> org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:165) >>>> at >>>> org.jclouds.http.internal.BaseHttpCommandExecutorService$HttpResponseCallable.call(BaseHttpCommandExecutorService.java:134) >>>> ... 5 more >>>> >>>> It tries to start the master node twice. Then tries for the >>>> regionserver. The command seems to hang after several attempts. >>>> >>>> >>>> On Thu, Dec 22, 2011 at 2:44 PM, Andrei Savu <[email protected]> >>>> wrote: >>>> > Can you try one more time by removing both whirr.hardware-id & >>>> > whirr.image-id? >>>> > >>>> > Whirr should be able to pick good defaults. BTW some of the errors you >>>> > are >>>> > seeing >>>> > are harmless - wait for it to finish and check the remote machines (we >>>> > are >>>> > improving >>>> > jclouds to hide some of those messages that are not relevant to the >>>> > end >>>> > user) >>>> > >>>> > On Thu, Dec 22, 2011 at 9:39 PM, David Medinets >>>> > <[email protected]> >>>> > wrote: >>>> >> >>>> >> Setting whirr.hardware-id=m1.small produced the same large data dump. >>>> >> >>>> >> Setting whirr.hardware-id=c1.xlarge produced this error: >>>> >> >>>> >> org.jclouds.http.HttpResponseException: command: POST >>>> >> https://ec2.us-east-1.amazonaws.com/ HTTP/1.1 failed with response: >>>> >> HTTP/1.1 400 Bad Request; content: [Non-Windows AMIs with a >>>> >> virtualization type of 'hvm' currently may only be used with Cluster >>>> >> Compute instance types.] >>>> >> >>>> >> I'm sorry if I'm being dense here. Getting a basic hadoop cluster to >>>> >> work >>>> >> using >>>> >> >>>> >> whirr.hardware-id=c1.medium >>>> >> whirr.image-id=us-east-1/ami-d59d6bbc >>>> >> >>>> >> was so easy that I was excited to try hbase. I'm just not >>>> >> understanding the error messages enough to guess how to result these >>>> >> issues. >>>> >> >>>> >> On Thu, Dec 22, 2011 at 2:10 PM, Andrei Savu <[email protected]> >>>> >> wrote: >>>> >> > On Thu, Dec 22, 2011 at 9:08 PM, David Medinets >>>> >> > <[email protected]> >>>> >> >> Do I need use to use a larger instance type? >>>> >> > >>>> >> > Yes, t1.micro is not really suitable to run something like Hadoop & >>>> >> > HBase. >>>> >> > It should work with m1.small just fine. >>>> > >>>> > >>> >>> >> >
