Correct, Stack - command line failure is fixed now with the systemProperties change that you mentioned. Thanks for that.
But eclipse is still throwing the same error, the one mentioned by ramkrishna vasudevan or earlier mentioned by me. Thanks, Amit Kabra. On Wed, Oct 4, 2017 at 10:48 AM, ramkrishna vasudevan < [email protected]> wrote: > Hi Stack > I just took an update and tried running some test cases with the eclipse > IDE working on linux > > java.io.IOException: Shutting down > at org.apache.hadoop.hbase.MiniHBaseCluster.init( > MiniHBaseCluster.java:232) > at org.apache.hadoop.hbase.MiniHBaseCluster.<init>( > MiniHBaseCluster.java:94) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniHBaseCluster( > HBaseTestingUtility.java:1128) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster( > HBaseTestingUtility.java:1082) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster( > HBaseTestingUtility.java:953) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster( > HBaseTestingUtility.java:935) > at > org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster( > HBaseTestingUtility.java:917) > at > org.apache.hadoop.hbase.regionserver.TestRegionReplicasWithModifyTa > ble.before(TestRegionReplicasWithModifyTable.java:65) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: > 62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > .... > > Caused by: java.lang.UnsatisfiedLinkError: failed to load the required > native library > at > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > Epoll.ensureAvailability(Epoll.java:78) > at > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > EpollEventLoopGroup.<clinit>(EpollEventLoopGroup.java:38) > at > org.apache.hadoop.hbase.util.NettyEventLoopGroupConfig.<init>( > NettyEventLoopGroupConfig.java:61) > at > org.apache.hadoop.hbase.regionserver.HRegionServer.< > init>(HRegionServer.java:552) > at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:475) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:422) > at > org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread( > JVMClusterUtil.java:140) > ... 26 more > Caused by: java.lang.UnsatisfiedLinkError: > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > NativeStaticallyReferencedJniMethods.epollin()I > at > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > NativeStaticallyReferencedJniMethods.epollin(Native > Method) > at > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > Native.<clinit>(Native.java:66) > at > org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > Epoll.<clinit>(Epoll.java:33) > ... 35 more > > This also appears to come out of netty. > When I run the same test case using 'mvn test -Dtest' command then it works > fine. > > Regards > Ram > > On Wed, Oct 4, 2017 at 12:42 AM, Stack <[email protected]> wrote: > > > It just came in... Looks better on jenkins than my linux machine (4m > > total): > > https://issues.apache.org/jira/browse/HBASE-18606? > > focusedCommentId=16190154&page=com.atlassian.jira. > > plugin.system.issuetabpanels:comment-tabpanel#comment-16190154 > > > > Let me try pushing it. > > > > St.Ack > > > > > > > > On Tue, Oct 3, 2017 at 12:00 PM, Stack <[email protected]> wrote: > > > > > Yeah (12x the processors, 8x RAM, 12x the disks...). Lets see how patch > > > does one jenkins... (HBASE-18606). > > > S > > > > > > On Tue, Oct 3, 2017 at 11:48 AM, Sean Busbey <[email protected]> > wrote: > > > > > >> the linux box roughly as capable as your laptop? > > >> > > >> On Tue, Oct 3, 2017 at 12:51 PM, Stack <[email protected]> wrote: > > >> > Tests passed eventually for me: > > >> > > > >> > real 47m6.498s > > >> > user 5m29.671s > > >> > sys 0m41.885s > > >> > > > >> > ... which is a big diff from macosx run. > > >> > > > >> > Need to look into this. > > >> > > > >> > St.Ack > > >> > > > >> > > > >> > On Tue, Oct 3, 2017 at 10:03 AM, Stack <[email protected]> wrote: > > >> > > > >> >> The below gets us further but now I see that the spark tests take a > > >> really > > >> >> long time to run on linux but complete promptly on macosx (2m 55s). > > >> >> Looking.... > > >> >> > > >> >> St.Ack > > >> >> > > >> >> On Tue, Oct 3, 2017 at 9:13 AM, Stack <[email protected]> wrote: > > >> >> > > >> >>> This seems to work for me. Does it work for you? > > >> >>> > > >> >>> > > >> >>> diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml > > >> >>> index 594aa2a..6d191e3 100644 > > >> >>> --- a/hbase-spark/pom.xml > > >> >>> +++ b/hbase-spark/pom.xml > > >> >>> @@ -568,6 +568,9 @@ > > >> >>> <junitxml>.</junitxml> > > >> >>> <filereports>WDF TestSuite.txt</filereports> > > >> >>> <parallel>false</parallel> > > >> >>> + <systemProperties> > > >> >>> + <org.apache.hadoop.hbase.shade > > >> d.io.netty.packagePrefix>org. > > >> >>> apache.hadoop.hbase.shaded.</org.apache.hadoop.hbase. > > >> >>> shaded.io.netty.packagePrefix> > > >> >>> + </systemProperties> > > >> >>> </configuration> > > >> >>> <executions> > > >> >>> <execution> > > >> >>> > > >> >>> St.Ack > > >> >>> > > >> >>> On Tue, Oct 3, 2017 at 8:45 AM, Amit Kabra < > [email protected] > > > > > >> >>> wrote: > > >> >>> > > >> >>>> Thanks Stack / Sean Busbey for replying. > > >> >>>> > > >> >>>> OS : Ubuntu 16.04.2 , 64 bit. > > >> >>>> Eclipse : Version: Neon.3 Release (4.6.3) > > >> >>>> HBase branch : branch-2 > > >> >>>> Command line test to reproduce : mvn clean package > > >> >>>> -Dtest=TestIncrementalBackup > > >> >>>> Reproduce from eclipse , right click on TestIncBackupRestore and > > run > > >> as > > >> >>>> junit from test class TestIncrementalBackup. > > >> >>>> No I am not embedding hbase in my application. I have just > checked > > >> out > > >> >>>> hbase , switched to branch-2 and run the unit test from command > > line > > >> or > > >> >>>> from eclipse. Failing with same error in both cases. > > >> >>>> Yes the trailing period is also present. > > >> >>>> > > >> >>>> Thanks, > > >> >>>> Amit Kabra. > > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> On Tue, Oct 3, 2017 at 8:53 PM, Stack <[email protected]> wrote: > > >> >>>> > > >> >>>> > Thank you for the detail. > > >> >>>> > > > >> >>>> > Pardon the questions below asking for yet more detail. I am > > unable > > >> to > > >> >>>> > reproduce locally or on another os (though we see this issue up > > on > > >> our > > >> >>>> > build box). > > >> >>>> > > > >> >>>> > What is your OS when you see the below? > > >> >>>> > > > >> >>>> > On Tue, Oct 3, 2017 at 2:06 AM, Amit Kabra < > > >> [email protected]> > > >> >>>> > wrote: > > >> >>>> > > > >> >>>> > > Hello, > > >> >>>> > > > > >> >>>> > > I am using "branch-2" branch of hbase, when I run unit test I > > get > > >> >>>> > following > > >> >>>> > > error for netty "java.lang.UnsatisfiedLinkError: failed to > > load > > >> the > > >> >>>> > > required native library" > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > This is running a unit test in your eclipse environment? > > >> >>>> > > > >> >>>> > You are trying to run an hbase-spark unit test when you see the > > >> above? > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > *I already have following set in "maven-surefire-plugin" in > > >> pom.xml > > >> >>>> as > > >> >>>> > > per http://hbase.apache.org/book.html#thirdparty > > >> >>>> > > <http://hbase.apache.org/book.html#thirdparty>* > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > > >> >>>> > Are you embedding hbase into your application? > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > <systemPropertyVariables> > > >> >>>> > > <!-- > > >> >>>> > > <test.build.classes>${test.bu > > >> ild.classes}</test.build. > > >> >>>> > classe > > >> >>>> > > s> > > >> >>>> > > --> > > >> >>>> > > <!--For shaded netty, to find the relocated > .so. > > >> >>>> > > Trick from > > >> >>>> > > https://stackoverflow.com/que > > >> stions/33825743/rename- > > >> >>>> > > files-inside-a-jar-using-some-maven-plugin > > >> >>>> > > <https://stackoverflow.com/questions/33825743/rename- > > >> >>>> > files-inside-a-jar-using-some-maven-plugin> > > >> >>>> > > > > >> >>>> > > The netty jar has a .so in it. Shading > requires > > >> >>>> rename of > > >> >>>> > > the .so and then passing a system > > >> >>>> > > property so netty finds the renamed .so and > > >> >>>> associates it > > >> >>>> > > w/ the relocated netty files. > > >> >>>> > > > > >> >>>> > > The relocated netty is in hbase-thirdparty > > >> >>>> dependency. > > >> >>>> > Just > > >> >>>> > > set this propery globally rather > > >> >>>> > > than per module. > > >> >>>> > > --> > > >> >>>> > > <org.apache.hadoop.hbase.shad > > >> >>>> ed.io.netty.packagePrefix> > > >> >>>> > > org.apache.hadoop.hbase.shaded.</org.apache.hadoop. > > >> >>>> > hbase.shaded.io.netty. > > >> >>>> > > packagePrefix> > > >> >>>> > > </systemPropertyVariables> > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > *And I see in the code as per HBASE-18271, all io.netty is > > >> already > > >> >>>> > replaced > > >> >>>> > > with org.apache.hadoop.hbase.shaded.io.netty* > > >> >>>> > > > > >> >>>> > > > > >> >>>> > The trailing period is also present? > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > > > >> >>>> > > If I run a test from eclipse , I see the error immediately > and > > my > > >> >>>> test > > >> >>>> > > doesn't run, but when I run from command line , the test runs > > >> but I > > >> >>>> get > > >> >>>> > the > > >> >>>> > > error at the end when the mvn command finishes. > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > Is it any eclipse test? > > >> >>>> > > > >> >>>> > Thank you. Let me try and fix this this morning. > > >> >>>> > > > >> >>>> > S > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > *Here is the complete error output.* > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > > > >> >>>> > > [INFO] > > >> >>>> > > [INFO] --- maven-surefire-plugin:2.19.1:test (default-test) > @ > > >> >>>> > hbase-spark > > >> >>>> > > --- > > >> >>>> > > [INFO] > > >> >>>> > > [INFO] --- scalatest-maven-plugin:1.0:test (test) @ > > hbase-spark > > >> --- > > >> >>>> > > Discovery starting. > > >> >>>> > > Discovery completed in 1 second, 558 milliseconds. > > >> >>>> > > Run starting. Expected test count is: 79 > > >> >>>> > > HBaseDStreamFunctionsSuite: > > >> >>>> > > Formatting using clusterid: testClusterID > > >> >>>> > > *** RUN ABORTED *** > > >> >>>> > > java.io.IOException: Shutting down > > >> >>>> > > at org.apache.hadoop.hbase.MiniHBaseCluster.init( > > >> >>>> > > MiniHBaseCluster.java:232) > > >> >>>> > > at org.apache.hadoop.hbase.MiniHBaseCluster.<init>( > > >> >>>> > > MiniHBaseCluster.java:94) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniHBaseCl > > >> >>>> uster( > > >> >>>> > > HBaseTestingUtility.java:1124) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:1078) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:949) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:943) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:872) > > >> >>>> > > at org.apache.hadoop.hbase.spark. > > >> HBaseDStreamFunctionsSuite.bef > > >> >>>> oreAll( > > >> >>>> > > HBaseDStreamFunctionsSuite.scala:41) > > >> >>>> > > at org.scalatest.BeforeAndAfterAl > > >> l$class.beforeAll(BeforeAndAft > > >> >>>> erAll. > > >> >>>> > > scala:187) > > >> >>>> > > at org.apache.hadoop.hbase.spark. > > >> HBaseDStreamFunctionsSuite.bef > > >> >>>> oreAll( > > >> >>>> > > HBaseDStreamFunctionsSuite.scala:30) > > >> >>>> > > ... > > >> >>>> > > Cause: java.lang.RuntimeException: Failed construction of > > >> Master: > > >> >>>> class > > >> >>>> > > org.apache.hadoop.hbase.master.HMasterorg.apache. > > >> >>>> > > hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > NativeStaticallyReferencedJniM > > >> >>>> > > ethods.epollin()I > > >> >>>> > > at org.apache.hadoop.hbase.util.J > > >> VMClusterUtil.createMasterThread( > > >> >>>> > > JVMClusterUtil.java:145) > > >> >>>> > > at org.apache.hadoop.hbase.LocalHBaseCluster.addMaster( > > >> >>>> > > LocalHBaseCluster.java:217) > > >> >>>> > > at org.apache.hadoop.hbase.LocalHBaseCluster.<init>( > > >> >>>> > > LocalHBaseCluster.java:152) > > >> >>>> > > at org.apache.hadoop.hbase.MiniHBaseCluster.init( > > >> >>>> > > MiniHBaseCluster.java:214) > > >> >>>> > > at org.apache.hadoop.hbase.MiniHBaseCluster.<init>( > > >> >>>> > > MiniHBaseCluster.java:94) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniHBaseCl > > >> >>>> uster( > > >> >>>> > > HBaseTestingUtility.java:1124) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:1078) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:949) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:943) > > >> >>>> > > at org.apache.hadoop.hbase.HBaseT > > >> estingUtility.startMiniCluster( > > >> >>>> > > HBaseTestingUtility.java:872) > > >> >>>> > > ... > > >> >>>> > > Cause: java.lang.UnsatisfiedLinkError: failed to load the > > >> required > > >> >>>> > > native > > >> >>>> > > library > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > Epoll.ensureAvailability(Epoll.java:78) > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > EpollEventLoopGroup.<clinit>(EpollEventLoopGroup.java:38) > > >> >>>> > > at org.apache.hadoop.hbase.util.N > > >> ettyEventLoopGroupConfig.<init>( > > >> >>>> > > NettyEventLoopGroupConfig.java:61) > > >> >>>> > > at org.apache.hadoop.hbase.regionserver.HRegionServer.< > > >> >>>> > > init>(HRegionServer.java:552) > > >> >>>> > > at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster. > > java: > > >> 475) > > >> >>>> > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0( > > >> Native > > >> >>>> > Method) > > >> >>>> > > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > > >> >>>> > > NativeConstructorAccessorImpl.java:62) > > >> >>>> > > at sun.reflect.DelegatingConstructorAccessorI > > mpl.newInstance( > > >> >>>> > > DelegatingConstructorAccessorImpl.java:45) > > >> >>>> > > at java.lang.reflect.Constructor. > > >> newInstance(Constructor.java:423) > > >> >>>> > > at org.apache.hadoop.hbase.util.J > > >> VMClusterUtil.createMasterThread( > > >> >>>> > > JVMClusterUtil.java:140) > > >> >>>> > > ... > > >> >>>> > > Cause: java.lang.UnsatisfiedLinkError: > > >> org.apache.hadoop.hbase. > > >> >>>> > > shaded.io.netty.channel.epoll.NativeStaticallyReferencedJniM > > >> >>>> > > ethods.epollin()I > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > NativeStaticallyReferencedJniMethods.epollin(Native Method) > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > Native.<clinit>(Native.java:66) > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > Epoll.<clinit>(Epoll.java:33) > > >> >>>> > > at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll. > > >> >>>> > > EpollEventLoopGroup.<clinit>(EpollEventLoopGroup.java:38) > > >> >>>> > > at org.apache.hadoop.hbase.util.N > > >> ettyEventLoopGroupConfig.<init>( > > >> >>>> > > NettyEventLoopGroupConfig.java:61) > > >> >>>> > > at org.apache.hadoop.hbase.regionserver.HRegionServer.< > > >> >>>> > > init>(HRegionServer.java:552) > > >> >>>> > > at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster. > > java: > > >> 475) > > >> >>>> > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0( > > >> Native > > >> >>>> > Method) > > >> >>>> > > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > > >> >>>> > > NativeConstructorAccessorImpl.java:62) > > >> >>>> > > at sun.reflect.DelegatingConstructorAccessorI > > mpl.newInstance( > > >> >>>> > > DelegatingConstructorAccessorImpl.java:45) > > >> >>>> > > ... > > >> >>>> > > [INFO] ------------------------------ > > >> ------------------------------ > > >> >>>> > > ------------ > > >> >>>> > > [INFO] Reactor Summary: > > >> >>>> > > [INFO] > > >> >>>> > > [INFO] Apache HBase ....................................... > > >> SUCCESS > > >> >>>> [ > > >> >>>> > > 1.575 s] > > >> >>>> > > [INFO] Apache HBase - Checkstyle .......................... > > >> SUCCESS [ > > >> >>>> > > 0.317 s] > > >> >>>> > > [INFO] Apache HBase - Annotations ......................... > > >> SUCCESS [ > > >> >>>> > > 0.537 s] > > >> >>>> > > [INFO] Apache HBase - Build Configuration ................. > > >> SUCCESS [ > > >> >>>> > > 0.053 s] > > >> >>>> > > [INFO] Apache HBase - Shaded Protocol ..................... > > >> SUCCESS [ > > >> >>>> > > 15.410 s] > > >> >>>> > > [INFO] Apache HBase - Common .............................. > > >> SUCCESS [ > > >> >>>> > > 4.603 s] > > >> >>>> > > [INFO] Apache HBase - Metrics API ......................... > > >> SUCCESS [ > > >> >>>> > > 1.213 s] > > >> >>>> > > [INFO] Apache HBase - Hadoop Compatibility ................ > > >> SUCCESS [ > > >> >>>> > > 0.985 s] > > >> >>>> > > [INFO] Apache HBase - Metrics Implementation .............. > > >> SUCCESS [ > > >> >>>> > > 0.863 s] > > >> >>>> > > [INFO] Apache HBase - Hadoop Two Compatibility ............ > > >> SUCCESS [ > > >> >>>> > > 1.750 s] > > >> >>>> > > [INFO] Apache HBase - Protocol ............................ > > >> SUCCESS [ > > >> >>>> > > 4.880 s] > > >> >>>> > > [INFO] Apache HBase - Client .............................. > > >> SUCCESS [ > > >> >>>> > > 5.233 s] > > >> >>>> > > [INFO] Apache HBase - Replication ......................... > > >> SUCCESS [ > > >> >>>> > > 1.040 s] > > >> >>>> > > [INFO] Apache HBase - Prefix Tree ......................... > > >> SUCCESS [ > > >> >>>> > > 1.121 s] > > >> >>>> > > [INFO] Apache HBase - Procedure ........................... > > >> SUCCESS [ > > >> >>>> > > 1.084 s] > > >> >>>> > > [INFO] Apache HBase - Resource Bundle ..................... > > >> SUCCESS [ > > >> >>>> > > 0.092 s] > > >> >>>> > > [INFO] Apache HBase - Server .............................. > > >> SUCCESS [ > > >> >>>> > > 19.849 s] > > >> >>>> > > [INFO] Apache HBase - MapReduce ........................... > > >> SUCCESS [ > > >> >>>> > > 4.221 s] > > >> >>>> > > [INFO] Apache HBase - Testing Util ........................ > > >> SUCCESS [ > > >> >>>> > > 3.273 s] > > >> >>>> > > [INFO] Apache HBase - Thrift .............................. > > >> SUCCESS [ > > >> >>>> > > 5.519 s] > > >> >>>> > > [INFO] Apache HBase - RSGroup ............................. > > >> SUCCESS [ > > >> >>>> > > 3.408 s] > > >> >>>> > > [INFO] Apache HBase - Shell ............................... > > >> SUCCESS > > >> >>>> [ > > >> >>>> > > 3.859 s] > > >> >>>> > > [INFO] Apache HBase - Coprocessor Endpoint ................ > > >> SUCCESS [ > > >> >>>> > > 4.038 s] > > >> >>>> > > [INFO] Apache HBase - Backup .............................. > > >> SUCCESS > > >> >>>> > [01:13 > > >> >>>> > > min] > > >> >>>> > > [INFO] Apache HBase - Integration Tests ................... > > >> SUCCESS [ > > >> >>>> > > 4.229 s] > > >> >>>> > > [INFO] Apache HBase - Examples ............................ > > >> SUCCESS [ > > >> >>>> > > 3.471 s] > > >> >>>> > > [INFO] Apache HBase - Rest ................................ > > >> SUCCESS > > >> >>>> [ > > >> >>>> > > 4.448 s] > > >> >>>> > > [INFO] Apache HBase - External Block Cache ................ > > >> SUCCESS [ > > >> >>>> > > 2.040 s] > > >> >>>> > > [INFO] Apache HBase - Spark ............................... > > >> FAILURE > > >> >>>> [ > > >> >>>> > > 32.833 s] > > >> >>>> > > [INFO] Apache HBase - Spark Integration Tests ............. > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Assembly ............................ > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Shaded .............................. > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Shaded - Client ..................... > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Shaded - MapReduce .................. > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase Shaded Packaging Invariants ........... > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Archetypes .......................... > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Exemplar for hbase-client archetype . > > >> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Exemplar for hbase-shaded-client > > archetype > > >> >>>> SKIPPED > > >> >>>> > > [INFO] Apache HBase - Archetype builder ................... > > >> SKIPPED > > >> >>>> > > [INFO] ------------------------------ > > >> ------------------------------ > > >> >>>> > > ------------ > > >> >>>> > > [INFO] BUILD FAILURE > > >> >>>> > > [INFO] ------------------------------ > > >> ------------------------------ > > >> >>>> > > ------------ > > >> >>>> > > [INFO] Total time: 03:26 min > > >> >>>> > > [INFO] Finished at: 2017-09-27T19:34:35+05:30 > > >> >>>> > > [INFO] Final Memory: 345M/6055M > > >> >>>> > > [INFO] ------------------------------ > > >> ------------------------------ > > >> >>>> > > ------------ > > >> >>>> > > [ERROR] Failed to execute goal org.scalatest:scalatest-maven- > > >> >>>> > > plugin:1.0:test > > >> >>>> > > (test) on project hbase-spark: There are test failures -> > [Help > > >> 1] > > >> >>>> > > [ERROR] > > >> >>>> > > [ERROR] To see the full stack trace of the errors, re-run > Maven > > >> with > > >> >>>> the > > >> >>>> > -e > > >> >>>> > > switch. > > >> >>>> > > [ERROR] Re-run Maven using the -X switch to enable full debug > > >> >>>> logging. > > >> >>>> > > [ERROR] > > >> >>>> > > [ERROR] For more information about the errors and possible > > >> solutions, > > >> >>>> > > please read the following articles: > > >> >>>> > > [ERROR] [Help 1] http://cwiki.apache.org/conflu > > >> ence/display/MAVEN/ > > >> >>>> > > MojoFailureException > > >> >>>> > > [ERROR] > > >> >>>> > > [ERROR] After correcting the problems, you can resume the > build > > >> with > > >> >>>> the > > >> >>>> > > command > > >> >>>> > > [ERROR] mvn <goals> -rf :hbase-spark > > >> >>>> > > > > >> >>>> > > > >> >>>> > > >> >>> > > >> >>> > > >> >> > > >> > > > > > > > > >
