Hi Tobe, I have removed HADOOP_HOME and HBASE_HOME , now i am getting different error like below \
Caused by: java.lang.IllegalArgumentException: Don't understand io engine name for cache - prefix with file:, heap or offheap at org.apache.hadoop.hbase.io.hfile.bucket.BucketCache.getIOEngineFromName(BucketCache.java:302) See below for complete startup logs : 2014-09-02 15:38:10,084 INFO [main] util.VersionInfo: Compiled by impadmin on Tue Sep 2 15:03:37 IST 2014 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:host.name=IMPETUS-NL147centos.impetus.co.in 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.version=1.7.0 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.vendor=Oracle Corporation 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.home=/usr/java/jdk1.7.0/jre 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.class.path=/usr/local/hbase/hbase-2.0.0-SNAPSHOT/bin/../conf:/usr/java/jdk1.7.0/lib/tools.jar:/usr/local/hbase/hbase-2.0.0-SNAPSHOT/bin/..:/usr/local/hbase/hbase-2.0.0- 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.io.tmpdir=/tmp 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:java.compiler=<NA> 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:os.name=Linux 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:os.arch=amd64 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:os.version=2.6.32-358.el6.x86_64 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:user.name=impadmin 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:user.home=/home/impadmin 2014-09-02 15:38:10,626 INFO [main] server.ZooKeeperServer: Server environment:user.dir=/usr/local/hbase/hbase-2.0.0-SNAPSHOT 2014-09-02 15:38:10,657 INFO [main] server.ZooKeeperServer: Created server with tickTime 2000 minSessionTimeout 4000 maxSessionTimeout 40000 datadir /tmp/hbase-impadmin/zookeeper/zookeeper_0/version-2 snapdir /tmp/hbase-impadmin/zookeeper/zookeeper_0/version-2 2014-09-02 15:38:10,704 INFO [main] server.NIOServerCnxnFactory: binding to port 0.0.0.0/0.0.0.0:2181 2014-09-02 15:38:11,291 INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181] server.NIOServerCnxnFactory: Accepted socket connection from / 127.0.0.1:49144 2014-09-02 15:38:11,296 INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181] server.NIOServerCnxn: Processing stat command from /127.0.0.1:49144 2014-09-02 15:38:11,301 INFO [Thread-2] server.NIOServerCnxn: Stat command output 2014-09-02 15:38:11,302 INFO [Thread-2] server.NIOServerCnxn: Closed socket connection for client /127.0.0.1:49144 (no session established for client) 2014-09-02 15:38:11,302 INFO [main] zookeeper.MiniZooKeeperCluster: Started MiniZK Cluster and connect 1 ZK server on client port: 2181 2014-09-02 15:38:12,066 INFO [main] regionserver.RSRpcServices: master/ IMPETUS-NL147centos.impetus.co.in/127.0.1.1:0 server-side HConnection retries=350 2014-09-02 15:38:12,517 INFO [main] ipc.SimpleRpcScheduler: Using deadline as user call queue, count=3 2014-09-02 15:38:12,543 INFO [main] ipc.RpcServer: master/ IMPETUS-NL147centos.impetus.co.in/127.0.1.1:0: started 10 reader(s). 2014-09-02 15:38:12,646 INFO [main] impl.MetricsConfig: loaded properties from hadoop-metrics2-hbase.properties 2014-09-02 15:38:12,710 INFO [main] impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2014-09-02 15:38:12,710 INFO [main] impl.MetricsSystemImpl: HBase metrics system started 2014-09-02 15:38:12,887 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2014-09-02 15:38:12,915 INFO [main] hfile.CacheConfig: Allocating LruBlockCache size=386.70 MB, blockSize=64 KB 2014-09-02 15:38:12,949 ERROR [main] master.HMasterCommandLine: Master exiting java.lang.RuntimeException: Failed construction of Master: class org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster at org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(JVMClusterUtil.java:145) at org.apache.hadoop.hbase.LocalHBaseCluster.addMaster(LocalHBaseCluster.java:214) at org.apache.hadoop.hbase.LocalHBaseCluster.<init>(LocalHBaseCluster.java:152) at org.apache.hadoop.hbase.master.HMasterCommandLine.startMaster(HMasterCommandLine.java:185) at org.apache.hadoop.hbase.master.HMasterCommandLine.run(HMasterCommandLine.java:139) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126) at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1764) Caused by: java.lang.IllegalArgumentException: Don't understand io engine name for cache - prefix with file:, heap or offheap at org.apache.hadoop.hbase.io.hfile.bucket.BucketCache.getIOEngineFromName(BucketCache.java:302) at org.apache.hadoop.hbase.io.hfile.bucket.BucketCache.<init>(BucketCache.java:218) at org.apache.hadoop.hbase.io.hfile.CacheConfig.getL2(CacheConfig.java:496) at org.apache.hadoop.hbase.io.hfile.CacheConfig.instantiateBlockCache(CacheConfig.java:517) at org.apache.hadoop.hbase.io.hfile.CacheConfig.<init>(CacheConfig.java:206) at org.apache.hadoop.hbase.regionserver.HRegionServer.<init>(HRegionServer.java:478) at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:264) at org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.<init>(HMasterCommandLine.java:266) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:525) at org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(JVMClusterUtil.java:141) ... 7 more Regards Sanjiv Singh Mob : +091 9990-447-339 On Tue, Sep 2, 2014 at 3:30 PM, tobe <tobeg3oo...@gmail.com> wrote: > I think you should NOT set HADOOP_HOME and HBASE_HOME, which have made > some trouble for me. > > > On Tue, Sep 2, 2014 at 5:48 PM, @Sanjiv Singh <sanjiv.is...@gmail.com> > wrote: > >> I just checked, following are defined for these. >> >> # pointing to hadoop 1.x >> HADOOP_HOME=/usr/local/hadoop/hadoop-1.2.1/ >> >> # directory which is created after >> extracting hbase-2.0.0-SNAPSHOT-bin.tar.gz >> >> HBASE_HOME=/usr/local/hbase/hbase-2.0.0-SNAPSHOT/ >> >> Regards >> Sanjiv Singh >> Mob : +091 9990-447-339 >> >> >> On Tue, Sep 2, 2014 at 2:35 PM, tobe <tobeg3oo...@gmail.com> wrote: >> >>> The default configuration should work well. Check if you set >>> $HADOOP_HOME and $HBASE_HOME or not. >>> >>> >>> On Tue, Sep 2, 2014 at 5:00 PM, @Sanjiv Singh <sanjiv.is...@gmail.com> >>> wrote: >>> >>>> Hi Dima, >>>> >>>> It's standalone mode where all daemons in one JVM. I have not changed >>>> any >>>> single configuration , i tried to start hbase with all default >>>> configuration. >>>> >>>> Let me know if need info to debug. >>>> >>>> >>>> >>>> >>>> Regards >>>> Sanjiv Singh >>>> Mob : +091 9990-447-339 >>>> >>>> >>>> On Tue, Sep 2, 2014 at 1:21 PM, Dima Spivak <dspi...@cloudera.com> >>>> wrote: >>>> >>>> > Hi Sanjiv, >>>> > >>>> > Are you looking to run HBase in standalone mode (all daemons in one >>>> JVM) >>>> > or pseudo-distributed mode (each with its own process, but on one >>>> host)? >>>> > Have you tried following the instructions on >>>> > http://hbase.apache.org/book.html regarding configurations to switch >>>> > between the two? >>>> > >>>> > All the best, >>>> > Dima >>>> > >>>> > >>>> > On Mon, Sep 1, 2014 at 10:53 PM, @Sanjiv Singh < >>>> sanjiv.is...@gmail.com> >>>> > wrote: >>>> > >>>> >> Any help on this ...issue is still not resolved. >>>> >> >>>> >> Regards >>>> >> Sanjiv Singh >>>> >> Mob : +091 9990-447-339 >>>> >> >>>> >> >>>> >> On Mon, Sep 1, 2014 at 5:31 PM, @Sanjiv Singh < >>>> sanjiv.is...@gmail.com> >>>> >> wrote: >>>> >> >>>> >> > Hi Matteo, >>>> >> > >>>> >> > Following your steps, I was able to build binary package hbase-2. >>>> >> > 0.0-SNAPSHOT-bin.tar.gz ..today I used the same to run hbase >>>> locally >>>> >> >>>> >> > it is giving error when i tried to start with start-hbase.sh >>>> >> > >>>> >> > I am totally lost with it , as don't have clue about the error. >>>> Please >>>> >> > help me on this. >>>> >> > >>>> >> > Following are logs generated .in logs directory : >>>> >> > >>>> >> > 2014-09-01 17:22:46,096 INFO [main] server.ZooKeeperServer: Server >>>> >> > environment:user.name=impadmin >>>> >> > 2014-09-01 17:22:46,096 INFO [main] server.ZooKeeperServer: Server >>>> >> > environment:user.home=/home/impadmin >>>> >> > 2014-09-01 17:22:46,096 INFO [main] server.ZooKeeperServer: Server >>>> >> > environment:user.dir=/usr/local/hbase/hbase-2.0.0-SNAPSHOT >>>> >> > 2014-09-01 17:22:46,109 INFO [main] server.ZooKeeperServer: >>>> Created >>>> >> > server with tickTime 2000 minSessionTimeout 4000 maxSessionTimeout >>>> 40000 >>>> >> > datadir /tmp/hbase-impad$ >>>> >> > 2014-09-01 17:22:46,121 INFO [main] server.NIOServerCnxnFactory: >>>> >> binding >>>> >> > to port 0.0.0.0/0.0.0.0:2181 >>>> >> > 2014-09-01 17:22:46,660 INFO [NIOServerCxn.Factory: >>>> >> 0.0.0.0/0.0.0.0:2181] >>>> >> > server.NIOServerCnxnFactory: Accepted socket connection from / >>>> >> > 127.0.0.1:34636 >>>> >> > 2014-09-01 17:22:46,667 INFO [NIOServerCxn.Factory: >>>> >> 0.0.0.0/0.0.0.0:2181] >>>> >> > server.NIOServerCnxn: Processing stat command from / >>>> 127.0.0.1:34636 >>>> >> > 2014-09-01 17:22:46,672 INFO [Thread-2] server.NIOServerCnxn: Stat >>>> >> > command output >>>> >> > 2014-09-01 17:22:46,673 INFO [Thread-2] server.NIOServerCnxn: >>>> Closed >>>> >> > socket connection for client /127.0.0.1:34636 (no session >>>> established >>>> >> for >>>> >> > client) >>>> >> > 2014-09-01 17:22:46,673 INFO [main] >>>> zookeeper.MiniZooKeeperCluster: >>>> >> > Started MiniZK Cluster and connect 1 ZK server on client port: 2181 >>>> >> > 2014-09-01 17:22:47,067 INFO [main] regionserver.RSRpcServices: >>>> master/ >>>> >> > IMPETUS-NL147centos.impetus.co.in/127.0.1.1:0 server-side >>>> HConnection >>>> >> > retries=350 >>>> >> > 2014-09-01 17:22:47,221 INFO [main] ipc.SimpleRpcScheduler: Using >>>> >> > deadline as user call queue, count=3 >>>> >> > 2014-09-01 17:22:47,234 INFO [main] ipc.RpcServer: master/ >>>> >> > IMPETUS-NL147centos.impetus.co.in/127.0.1.1:0: started 10 >>>> reader(s). >>>> >> > 2014-09-01 17:22:47,297 INFO [main] impl.MetricsConfig: loaded >>>> >> properties >>>> >> > from hadoop-metrics2-hbase.properties >>>> >> > 2014-09-01 17:22:47,324 INFO [main] impl.MetricsSystemImpl: >>>> Scheduled >>>> >> > snapshot period at 10 second(s). >>>> >> > 2014-09-01 17:22:47,324 INFO [main] impl.MetricsSystemImpl: HBase >>>> >> metrics >>>> >> > system started >>>> >> > 2014-09-01 17:22:47,428 ERROR [main] master.HMasterCommandLine: >>>> Master >>>> >> > exiting >>>> >> > java.lang.RuntimeException: Failed construction of Master: class >>>> >> > org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMasternull >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(JVMClusterUtil.java:145) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.LocalHBaseCluster.addMaster(LocalHBaseCluster.java:214) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.LocalHBaseCluster.<init>(LocalHBaseCluster.java:152) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.master.HMasterCommandLine.startMaster(HMasterCommandLine.java:185) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.master.HMasterCommandLine.run(HMasterCommandLine.java:139) >>>> >> > at >>>> org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126) >>>> >> > at >>>> >> org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1764) >>>> >> > Caused by: java.lang.RuntimeException: >>>> >> > java.lang.reflect.InvocationTargetException >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131) >>>> >> > at org.apache.hadoop.security.Groups.<init>(Groups.java:64) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:255) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.UserGroupInformation.isAuthenticationMethodEnabled(UserGroupInformation.java:309) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:303) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.security.User$SecureHadoopUser.isSecurityEnabled(User.java:349) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.security.User$SecureHadoopUser.login(User.java:340) >>>> >> > at >>>> org.apache.hadoop.hbase.security.User.login(User.java:208) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.security.UserProvider.login(UserProvider.java:116) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.regionserver.HRegionServer.login(HRegionServer.java:526) >>>> >> > at >>>> >> org.apache.hadoop.hbase.master.HMaster.login(HMaster.java:338) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.regionserver.HRegionServer.<init>(HRegionServer.java:475) >>>> >> > at >>>> >> org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:264) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.<init>(HMasterCommandLine.java:266) >>>> >> > at >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>> >> > Method) >>>> >> > at >>>> >> > >>>> >> >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >>>> >> > at >>>> >> > >>>> >> >>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>> >> > at >>>> >> java.lang.reflect.Constructor.newInstance(Constructor.java:525) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(JVMClusterUtil.java:141) >>>> >> > Caused by: java.lang.reflect.InvocationTargetException >>>> >> > at >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>> >> > Method) >>>> >> > at >>>> >> > >>>> >> >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >>>> >> > at >>>> >> > >>>> >> >>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>> >> > at >>>> >> java.lang.reflect.Constructor.newInstance(Constructor.java:525) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:129) >>>> >> > ... 27 more >>>> >> > Caused by: java.lang.UnsatisfiedLinkError: >>>> >> > >>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V >>>> >> > at >>>> >> > >>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native >>>> >> > Method) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49) >>>> >> > at >>>> >> > >>>> >> >>>> org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback.<init>(JniBasedUnixGroupsMappingWithFallback.java:38) >>>> >> > ... 32 more >>>> >> > >>>> >> > >>>> >> > >>>> >> > Regards >>>> >> > Sanjiv Singh >>>> >> > Mob : +091 9990-447-339 >>>> >> > >>>> >> > >>>> >> > On Fri, Aug 22, 2014 at 9:18 PM, @Sanjiv Singh < >>>> sanjiv.is...@gmail.com> >>>> >> > wrote: >>>> >> > >>>> >> >> Hi Matteo, >>>> >> >> >>>> >> >> I cleaned up all, started from scratch and followed all steps >>>> provided >>>> >> >> by you ...it worked this time without any issue. >>>> >> >> >>>> >> >> I really don't know what was the issue. thank you very much for >>>> the >>>> >> >> help. i will move forward with this. >>>> >> >> >>>> >> >> Regards >>>> >> >> Sanjiv Singh >>>> >> >> Mob : +091 9990-447-339 >>>> >> >> >>>> >> >> >>>> >> >> On Fri, Aug 22, 2014 at 8:51 PM, @Sanjiv Singh < >>>> sanjiv.is...@gmail.com >>>> >> > >>>> >> >> wrote: >>>> >> >> >>>> >> >>> my maven : >>>> >> >>> >>>> >> >>> $ mvn --version >>>> >> >>> >>>> >> >>> Apache Maven 3.2.1 (ea8b2b07643dbb1b84b6d16e1f08391b666bc1e9; >>>> >> >>> 2014-02-14T23:07:52+05:30) >>>> >> >>> Maven home: /usr/local/apache-maven/apache-maven-3.2.1 >>>> >> >>> Java version: 1.7.0, vendor: Oracle Corporation >>>> >> >>> Java home: /usr/java/jdk1.7.0/jre >>>> >> >>> Default locale: en_US, platform encoding: UTF-8 >>>> >> >>> OS name: "linux", version: "2.6.32-358.el6.x86_64", arch: >>>> "amd64", >>>> >> >>> family: "unix" >>>> >> >>> >>>> >> >>> >>>> >> >>> Regards >>>> >> >>> Sanjiv Singh >>>> >> >>> Mob : +091 9990-447-339 >>>> >> >>> >>>> >> >>> >>>> >> >>> On Fri, Aug 22, 2014 at 8:13 PM, Ted Yu <yuzhih...@gmail.com> >>>> wrote: >>>> >> >>> >>>> >> >>>> jamon-maven-plugin is used to generate the XXTmpl.java files >>>> >> >>>> >>>> >> >>>> In a successful build, you would see: >>>> >> >>>> >>>> >> >>>> [INFO] Source directory: >>>> >> >>>> /Users/tyu/trunk/hbase-server/target/generated-jamon added. >>>> >> >>>> [INFO] Source directory: >>>> >> >>>> /Users/tyu/trunk/hbase-server/target/generated-sources/java >>>> added. >>>> >> >>>> [INFO] >>>> >> >>>> [INFO] --- jamon-maven-plugin:2.3.4:translate (default) @ >>>> >> hbase-server >>>> >> >>>> --- >>>> >> >>>> [INFO] Translating 10 templates from >>>> >> >>>> /Users/tyu/trunk/hbase-server/src/main/jamon to >>>> >> >>>> /Users/tyu/trunk/hbase-server/target/generated-jamon >>>> >> >>>> [INFO] >>>> >> >>>> >>>> >> >>>> What maven version are you using ? >>>> >> >>>> Here is the version I use: >>>> >> >>>> >>>> >> >>>> $ mvn --version >>>> >> >>>> Apache Maven 3.0.5 (r01de14724cdef164cd33c7c8c2fe155faf9602da; >>>> >> >>>> 2013-02-19 05:51:28-0800) >>>> >> >>>> Maven home: /Users/tyu/apache-maven-3.0.5 >>>> >> >>>> >>>> >> >>>> Cheers >>>> >> >>>> >>>> >> >>>> >>>> >> >>>> On Fri, Aug 22, 2014 at 1:57 AM, @Sanjiv Singh < >>>> >> sanjiv.is...@gmail.com> >>>> >> >>>> wrote: >>>> >> >>>> >>>> >> >>>>> Thanks for quick response.. >>>> >> >>>>> >>>> >> >>>>> Please find attached "compile.log" for logs of command "mvn >>>> clean >>>> >> >>>>> package -DskipTests". >>>> >> >>>>> >>>> >> >>>>> Which clearly say "Building HBase 2.0.0-SNAPSHOT". >>>> >> >>>>> >>>> >> >>>>> Let me know if I am wrong. >>>> >> >>>>> >>>> >> >>>>> >>>> >> >>>>> Regards >>>> >> >>>>> Sanjiv Singh >>>> >> >>>>> Mob : +091 9990-447-339 >>>> >> >>>>> >>>> >> >>>>> >>>> >> >>>>> On Fri, Aug 22, 2014 at 2:20 PM, tobe <tobeg3oo...@gmail.com> >>>> >> wrote: >>>> >> >>>>> >>>> >> >>>>>> 2.0.0-SNAPSHOT should be the version of Hadoop, not HBase. >>>> >> >>>>>> >>>> >> >>>>>> Refer to the official guide >>>> >> http://hbase.apache.org/book/build.html, >>>> >> >>>>>> you should run `mvn clean package -DskipTests` to compile. >>>> >> >>>>>> >>>> >> >>>>>> >>>> >> >>>>>> On Fri, Aug 22, 2014 at 4:41 PM, @Sanjiv Singh < >>>> >> >>>>>> sanjiv.is...@gmail.com> wrote: >>>> >> >>>>>> >>>> >> >>>>>>> HI, >>>> >> >>>>>>> Here are details : >>>> >> >>>>>>> HBase - 2.0.0-SNAPSHOT (current hbase-master) >>>> >> >>>>>>> java version "1.7.0" >>>> >> >>>>>>> >>>> >> >>>>>>> Regards >>>> >> >>>>>>> Sanjiv Singh >>>> >> >>>>>>> Mob : +091 9990-447-339 >>>> >> >>>>>>> >>>> >> >>>>>>> >>>> >> >>>>>>> On Fri, Aug 22, 2014 at 12:54 PM, tobe < >>>> tobeg3oo...@gmail.com> >>>> >> >>>>>>> wrote: >>>> >> >>>>>>> >>>> >> >>>>>>>> What're the versions of java and hbase? >>>> >> >>>>>>>> >>>> >> >>>>>>>> >>>> >> >>>>>>>> On Fri, Aug 22, 2014 at 2:41 PM, @Sanjiv Singh < >>>> >> >>>>>>>> sanjiv.is...@gmail.com> wrote: >>>> >> >>>>>>>> >>>> >> >>>>>>>>> Hi All, >>>> >> >>>>>>>>> I just started Exploring HBase. I have downloaded Hbase >>>> master >>>> >> >>>>>>>>> source code. >>>> >> >>>>>>>>> While i am trying to compile and build it locally, it is >>>> >> giving >>>> >> >>>>>>>>> me error >>>> >> >>>>>>>>> missing some class and packages. >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> From the source it looks like missing classes are actually >>>> not >>>> >> >>>>>>>>> written in >>>> >> >>>>>>>>> java. but generated from other way around >>>> "RSStatusTmpl.jamon". >>>> >> >>>>>>>>> Please >>>> >> >>>>>>>>> help me on this to resolve the issue. >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >> mvn clean install -DskipTests >>>> >> >>>>>>>>> [INFO] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> ------------------------------------------------------------------------ >>>> >> >>>>>>>>> [INFO] BUILD FAILURE >>>> >> >>>>>>>>> [INFO] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> ------------------------------------------------------------------------ >>>> >> >>>>>>>>> [INFO] Total time: 31.110 s >>>> >> >>>>>>>>> [INFO] Finished at: 2014-08-22T12:00:25+05:30 >>>> >> >>>>>>>>> [INFO] Final Memory: 63M/408M >>>> >> >>>>>>>>> [INFO] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> ------------------------------------------------------------------------ >>>> >> >>>>>>>>> [ERROR] Failed to execute goal >>>> >> >>>>>>>>> >>>> org.apache.maven.plugins:maven-compiler-plugin:2.5.1:compile >>>> >> >>>>>>>>> (default-compile) on project hbase-server: Compilation >>>> failure: >>>> >> >>>>>>>>> Compilation >>>> >> >>>>>>>>> failure: >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java:[36,42] >>>> >> >>>>>>>>> error: package org.apache.hadoop.hbase.tmpl.master does not >>>> >> exist >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java:[29,48] >>>> >> >>>>>>>>> error: package org.apache.hadoop.hbase.tmpl.regionserver >>>> does >>>> >> not >>>> >> >>>>>>>>> exist >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java:[75,4] >>>> >> >>>>>>>>> error: cannot find symbol >>>> >> >>>>>>>>> [ERROR] symbol: class MasterStatusTmpl >>>> >> >>>>>>>>> [ERROR] location: class MasterStatusServlet >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java:[75,32] >>>> >> >>>>>>>>> error: cannot find symbol >>>> >> >>>>>>>>> [ERROR] symbol: class MasterStatusTmpl >>>> >> >>>>>>>>> [ERROR] location: class MasterStatusServlet >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java:[51,4] >>>> >> >>>>>>>>> error: cannot find symbol >>>> >> >>>>>>>>> [ERROR] symbol: class RSStatusTmpl >>>> >> >>>>>>>>> [ERROR] location: class RSStatusServlet >>>> >> >>>>>>>>> [ERROR] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>> /home/impadmin/source-code/hbase/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java:[51,28] >>>> >> >>>>>>>>> error: cannot find symbol >>>> >> >>>>>>>>> [ERROR] -> [Help 1] >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> >>>> >> >>>>>>>>> Regards >>>> >> >>>>>>>>> Sanjiv Singh >>>> >> >>>>>>>>> Mob : +091 9990-447-339 >>>> >> >>>>>>>>> >>>> >> >>>>>>>> >>>> >> >>>>>>>> >>>> >> >>>>>>> >>>> >> >>>>>> >>>> >> >>>>> >>>> >> >>>> >>>> >> >>> >>>> >> >> >>>> >> > >>>> >> >>>> > >>>> > >>>> >>> >>> >> >