This seems to work for me. Does it work for you?

diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml
index 594aa2a..6d191e3 100644
--- a/hbase-spark/pom.xml
+++ b/hbase-spark/pom.xml
@@ -568,6 +568,9 @@
           <junitxml>.</junitxml>
           <filereports>WDF TestSuite.txt</filereports>
           <parallel>false</parallel>
+          <systemProperties>
+
 
<org.apache.hadoop.hbase.shaded.io.netty.packagePrefix>org.apache.hadoop.hbase.shaded.</org.apache.hadoop.hbase.shaded.io.netty.packagePrefix>
+          </systemProperties>
         </configuration>
         <executions>
           <execution>

St.Ack

On Tue, Oct 3, 2017 at 8:45 AM, Amit Kabra <amitkabrai...@gmail.com> wrote:

> Thanks Stack / Sean Busbey for replying.
>
> OS : Ubuntu 16.04.2 , 64 bit.
> Eclipse : Version: Neon.3 Release (4.6.3)
> HBase branch : branch-2
> Command line test to reproduce : mvn clean package
> -Dtest=TestIncrementalBackup
> Reproduce from eclipse , right click on TestIncBackupRestore and run as
> junit from test class TestIncrementalBackup.
> No I am not embedding hbase in my application. I have just checked out
> hbase , switched to branch-2 and run the unit test from command line or
> from eclipse. Failing with same error in both cases.
> Yes the trailing period is also present.
>
> Thanks,
> Amit Kabra.
>
>
>
>
>
> On Tue, Oct 3, 2017 at 8:53 PM, Stack <st...@duboce.net> wrote:
>
> > Thank you for the detail.
> >
> > Pardon the questions below asking for yet more detail. I am unable to
> > reproduce locally or on another os (though we see this issue up on our
> > build box).
> >
> > What is your OS when you see the below?
> >
> > On Tue, Oct 3, 2017 at 2:06 AM, Amit Kabra <amitkabrai...@gmail.com>
> > wrote:
> >
> > > Hello,
> > >
> > > I am using "branch-2" branch of hbase, when I run unit test I get
> > following
> > > error for netty "java.lang.UnsatisfiedLinkError: failed to load the
> > > required native library"
> > >
> > >
> > > This is running a unit test in your eclipse environment?
> >
> > You are trying to run an hbase-spark unit test when you see the above?
> >
> >
> >
> >
> > > *I already have following set in "maven-surefire-plugin" in pom.xml as
> > > per http://hbase.apache.org/book.html#thirdparty
> > > <http://hbase.apache.org/book.html#thirdparty>*
> > >
> > >
> > >
> >
> > Are you embedding hbase into your application?
> >
> >
> >
> > >             <systemPropertyVariables>
> > >                 <!--
> > >               <test.build.classes>${test.build.classes}</test.build.
> > classe
> > > s>
> > >                 -->
> > >               <!--For shaded netty, to find the relocated .so.
> > >                    Trick from
> > >                 https://stackoverflow.com/questions/33825743/rename-
> > > files-inside-a-jar-using-some-maven-plugin
> > > <https://stackoverflow.com/questions/33825743/rename-
> > files-inside-a-jar-using-some-maven-plugin>
> > >
> > >                 The netty jar has a .so in it. Shading requires rename
> of
> > > the .so and then passing a system
> > >                 property so netty finds the renamed .so and associates
> it
> > > w/ the relocated netty files.
> > >
> > >                 The relocated netty is in hbase-thirdparty dependency.
> > Just
> > > set this propery globally rather
> > >                 than per module.
> > >                -->
> > >               <org.apache.hadoop.hbase.shaded.io.netty.packagePrefix>
> > > org.apache.hadoop.hbase.shaded.</org.apache.hadoop.
> > hbase.shaded.io.netty.
> > > packagePrefix>
> > >             </systemPropertyVariables>
> > >
> > >
> > >
> > > *And I see in the code as per HBASE-18271, all io.netty is already
> > replaced
> > > with org.apache.hadoop.hbase.shaded.io.netty*
> > >
> > >
> > The trailing period is also present?
> >
> >
> >
> > >
> > > If I run a test from eclipse , I see the error immediately and my test
> > > doesn't run, but when I run from command line , the test runs but I get
> > the
> > > error at the end when the mvn command finishes.
> > >
> > >
> > > Is it any eclipse test?
> >
> > Thank you. Let me try and fix this this morning.
> >
> > S
> >
> >
> >
> >
> >
> > > *Here is the complete error output.*
> > >
> > >
> > >
> > > [INFO]
> > > [INFO] --- maven-surefire-plugin:2.19.1:test (default-test) @
> > hbase-spark
> > > ---
> > > [INFO]
> > > [INFO] --- scalatest-maven-plugin:1.0:test (test) @ hbase-spark ---
> > > Discovery starting.
> > > Discovery completed in 1 second, 558 milliseconds.
> > > Run starting. Expected test count is: 79
> > > HBaseDStreamFunctionsSuite:
> > > Formatting using clusterid: testClusterID
> > > *** RUN ABORTED ***
> > >   java.io.IOException: Shutting down
> > >   at org.apache.hadoop.hbase.MiniHBaseCluster.init(
> > > MiniHBaseCluster.java:232)
> > >   at org.apache.hadoop.hbase.MiniHBaseCluster.<init>(
> > > MiniHBaseCluster.java:94)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.
> startMiniHBaseCluster(
> > > HBaseTestingUtility.java:1124)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:1078)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:949)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:943)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:872)
> > >   at org.apache.hadoop.hbase.spark.HBaseDStreamFunctionsSuite.
> beforeAll(
> > > HBaseDStreamFunctionsSuite.scala:41)
> > >   at org.scalatest.BeforeAndAfterAll$class.
> beforeAll(BeforeAndAfterAll.
> > > scala:187)
> > >   at org.apache.hadoop.hbase.spark.HBaseDStreamFunctionsSuite.
> beforeAll(
> > > HBaseDStreamFunctionsSuite.scala:30)
> > >   ...
> > >   Cause: java.lang.RuntimeException: Failed construction of Master:
> class
> > > org.apache.hadoop.hbase.master.HMasterorg.apache.
> > > hadoop.hbase.shaded.io.netty.channel.epoll.
> > NativeStaticallyReferencedJniM
> > > ethods.epollin()I
> > >   at org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(
> > > JVMClusterUtil.java:145)
> > >   at org.apache.hadoop.hbase.LocalHBaseCluster.addMaster(
> > > LocalHBaseCluster.java:217)
> > >   at org.apache.hadoop.hbase.LocalHBaseCluster.<init>(
> > > LocalHBaseCluster.java:152)
> > >   at org.apache.hadoop.hbase.MiniHBaseCluster.init(
> > > MiniHBaseCluster.java:214)
> > >   at org.apache.hadoop.hbase.MiniHBaseCluster.<init>(
> > > MiniHBaseCluster.java:94)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.
> startMiniHBaseCluster(
> > > HBaseTestingUtility.java:1124)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:1078)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:949)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:943)
> > >   at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(
> > > HBaseTestingUtility.java:872)
> > >   ...
> > >   Cause: java.lang.UnsatisfiedLinkError: failed to load the required
> > > native
> > > library
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > Epoll.ensureAvailability(Epoll.java:78)
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > EpollEventLoopGroup.<clinit>(EpollEventLoopGroup.java:38)
> > >   at org.apache.hadoop.hbase.util.NettyEventLoopGroupConfig.<init>(
> > > NettyEventLoopGroupConfig.java:61)
> > >   at org.apache.hadoop.hbase.regionserver.HRegionServer.<
> > > init>(HRegionServer.java:552)
> > >   at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:475)
> > >   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> > Method)
> > >   at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> > > NativeConstructorAccessorImpl.java:62)
> > >   at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> > > DelegatingConstructorAccessorImpl.java:45)
> > >   at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> > >   at org.apache.hadoop.hbase.util.JVMClusterUtil.createMasterThread(
> > > JVMClusterUtil.java:140)
> > >   ...
> > >   Cause: java.lang.UnsatisfiedLinkError: org.apache.hadoop.hbase.
> > > shaded.io.netty.channel.epoll.NativeStaticallyReferencedJniM
> > > ethods.epollin()I
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > NativeStaticallyReferencedJniMethods.epollin(Native Method)
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > Native.<clinit>(Native.java:66)
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > Epoll.<clinit>(Epoll.java:33)
> > >   at org.apache.hadoop.hbase.shaded.io.netty.channel.epoll.
> > > EpollEventLoopGroup.<clinit>(EpollEventLoopGroup.java:38)
> > >   at org.apache.hadoop.hbase.util.NettyEventLoopGroupConfig.<init>(
> > > NettyEventLoopGroupConfig.java:61)
> > >   at org.apache.hadoop.hbase.regionserver.HRegionServer.<
> > > init>(HRegionServer.java:552)
> > >   at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:475)
> > >   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> > Method)
> > >   at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> > > NativeConstructorAccessorImpl.java:62)
> > >   at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> > > DelegatingConstructorAccessorImpl.java:45)
> > >   ...
> > > [INFO] ------------------------------------------------------------
> > > ------------
> > > [INFO] Reactor Summary:
> > > [INFO]
> > > [INFO] Apache HBase ....................................... SUCCESS [
> > > 1.575 s]
> > > [INFO] Apache HBase - Checkstyle .......................... SUCCESS [
> > > 0.317 s]
> > > [INFO] Apache HBase - Annotations ......................... SUCCESS [
> > > 0.537 s]
> > > [INFO] Apache HBase - Build Configuration ................. SUCCESS [
> > > 0.053 s]
> > > [INFO] Apache HBase - Shaded Protocol ..................... SUCCESS [
> > > 15.410 s]
> > > [INFO] Apache HBase - Common .............................. SUCCESS [
> > > 4.603 s]
> > > [INFO] Apache HBase - Metrics API ......................... SUCCESS [
> > > 1.213 s]
> > > [INFO] Apache HBase - Hadoop Compatibility ................ SUCCESS [
> > > 0.985 s]
> > > [INFO] Apache HBase - Metrics Implementation .............. SUCCESS [
> > > 0.863 s]
> > > [INFO] Apache HBase - Hadoop Two Compatibility ............ SUCCESS [
> > > 1.750 s]
> > > [INFO] Apache HBase - Protocol ............................ SUCCESS [
> > > 4.880 s]
> > > [INFO] Apache HBase - Client .............................. SUCCESS [
> > > 5.233 s]
> > > [INFO] Apache HBase - Replication ......................... SUCCESS [
> > > 1.040 s]
> > > [INFO] Apache HBase - Prefix Tree ......................... SUCCESS [
> > > 1.121 s]
> > > [INFO] Apache HBase - Procedure ........................... SUCCESS [
> > > 1.084 s]
> > > [INFO] Apache HBase - Resource Bundle ..................... SUCCESS [
> > > 0.092 s]
> > > [INFO] Apache HBase - Server .............................. SUCCESS [
> > > 19.849 s]
> > > [INFO] Apache HBase - MapReduce ........................... SUCCESS [
> > > 4.221 s]
> > > [INFO] Apache HBase - Testing Util ........................ SUCCESS [
> > > 3.273 s]
> > > [INFO] Apache HBase - Thrift .............................. SUCCESS [
> > > 5.519 s]
> > > [INFO] Apache HBase - RSGroup ............................. SUCCESS [
> > > 3.408 s]
> > > [INFO] Apache HBase - Shell ............................... SUCCESS [
> > > 3.859 s]
> > > [INFO] Apache HBase - Coprocessor Endpoint ................ SUCCESS [
> > > 4.038 s]
> > > [INFO] Apache HBase - Backup .............................. SUCCESS
> > [01:13
> > > min]
> > > [INFO] Apache HBase - Integration Tests ................... SUCCESS [
> > > 4.229 s]
> > > [INFO] Apache HBase - Examples ............................ SUCCESS [
> > > 3.471 s]
> > > [INFO] Apache HBase - Rest ................................ SUCCESS [
> > > 4.448 s]
> > > [INFO] Apache HBase - External Block Cache ................ SUCCESS [
> > > 2.040 s]
> > > [INFO] Apache HBase - Spark ............................... FAILURE [
> > > 32.833 s]
> > > [INFO] Apache HBase - Spark Integration Tests ............. SKIPPED
> > > [INFO] Apache HBase - Assembly ............................ SKIPPED
> > > [INFO] Apache HBase - Shaded .............................. SKIPPED
> > > [INFO] Apache HBase - Shaded - Client ..................... SKIPPED
> > > [INFO] Apache HBase - Shaded - MapReduce .................. SKIPPED
> > > [INFO] Apache HBase Shaded Packaging Invariants ........... SKIPPED
> > > [INFO] Apache HBase - Archetypes .......................... SKIPPED
> > > [INFO] Apache HBase - Exemplar for hbase-client archetype . SKIPPED
> > > [INFO] Apache HBase - Exemplar for hbase-shaded-client archetype
> SKIPPED
> > > [INFO] Apache HBase - Archetype builder ................... SKIPPED
> > > [INFO] ------------------------------------------------------------
> > > ------------
> > > [INFO] BUILD FAILURE
> > > [INFO] ------------------------------------------------------------
> > > ------------
> > > [INFO] Total time: 03:26 min
> > > [INFO] Finished at: 2017-09-27T19:34:35+05:30
> > > [INFO] Final Memory: 345M/6055M
> > > [INFO] ------------------------------------------------------------
> > > ------------
> > > [ERROR] Failed to execute goal org.scalatest:scalatest-maven-
> > > plugin:1.0:test
> > > (test) on project hbase-spark: There are test failures -> [Help 1]
> > > [ERROR]
> > > [ERROR] To see the full stack trace of the errors, re-run Maven with
> the
> > -e
> > > switch.
> > > [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> > > [ERROR]
> > > [ERROR] For more information about the errors and possible solutions,
> > > please read the following articles:
> > > [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/
> > > MojoFailureException
> > > [ERROR]
> > > [ERROR] After correcting the problems, you can resume the build with
> the
> > > command
> > > [ERROR]   mvn <goals> -rf :hbase-spark
> > >
> >
>

Reply via email to