Thanks Mahmoud, This is what I am using, but as the previous reply stated, I receiving an exception when starting the cluster. Thinking about it, it looks to be more of a build problem of my hbase mini cluster, as I am receiving the following error:
16/03/14 12:29:00 WARN datanode.DataNode: IOException in BlockReceiver.run(): java.io.IOException: Failed to move meta file for ReplicaBeingWritten, blk_1073741825_1001, RBW getNumBytes() = 7 getBytesOnDisk() = 7 getVisibleLength()= 7 getVolume() = C:\Users\unknown\Documents\trs\target\test-data\780d11ca-27b8-4004-bed8-480bc9903125\dfscluster_d292c05b-0190-43b1-83b2-bebf483c8b3c\dfs\data\data1\current getBlockFile() = C:\Users\unknown\Documents\trs\target\test-data\780d11ca-27b8-4004-bed8-480bc9903125\dfscluster_d292c05b-0190-43b1-83b2-bebf483c8b3c\dfs\data\data1\current\BP-1081755239-10.66.90.86-1457954925705\current\rbw\blk_1073741825 bytesAcked=7 bytesOnDisk=7 from C:\Users\unknown\Documents\trs\target\test-data\780d11ca-27b8-4004-bed8-480bc9903125\dfscluster_d292c05b-0190-43b1-83b2-bebf483c8b3c\dfs\data\data1\current\BP-1081755239-10.66.90.86-1457954925705\current\rbw\blk_1073741825_1001.meta to C:\Users\unknown\Documents\trs\target\test-data\780d11ca-27b8-4004-bed8-480bc9903125\dfscluster_d292c05b-0190-43b1-83b2-bebf483c8b3c\dfs\data\data1\current\BP-1081755239-10.66.90.86-1457954925705\current\finalized\subdir0\subdir0\blk_1073741825_1001.meta at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:615) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.BlockPoolSlice.addBlock(BlockPoolSlice.java:250) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsVolumeImpl.addBlock(FsVolumeImpl.java:229) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeReplica(FsDatasetImpl.java:1119) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.finalizeBlock(FsDatasetImpl.java:1100) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.finalizeBlock(BlockReceiver.java:1293) at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1233) at java.lang.Thread.run(Thread.java:745) Caused by: 3: The system cannot find the path specified. at org.apache.hadoop.io.nativeio.NativeIO.renameTo0(Native Method) at org.apache.hadoop.io.nativeio.NativeIO.renameTo(NativeIO.java:830) at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.moveBlockFiles(FsDatasetImpl.java:613) ... 7 more 16/03/14 12:29:00 INFO datanode.DataNode: Starting CheckDiskError Thread Thanks, Keech On 14 Mar 2016 6:10 pm, "Parsian, Mahmoud" <[email protected]> wrote: > Hi Keech, > > You may use the org.apache.hadoop.hbase.HBaseCommonTestingUtility class to > start a ZK, and an HBase cluster and then do your unit tests and > integration. > I am using this with junit and it works very well. But I am using Java > only. > > Best regards, > Mahmoud Parsian > > > On 3/13/16, 11:52 PM, "Nkechi Achara" <[email protected]> wrote: > > >Hi, > > > >I am trying to find an example of how to spin up a Hbase server in a mock > >or integration style, so I can test my code locally in my IDE. > >I have tried fake-hbase and hbase testing utility and receive errors > >especially when trying to start the cluster. > >Has anyone got any examples in scala to do this? > > > >Thanks, > > > >Keech > >
