>From your description, it seems the error was similar to the one in the
description of HADOOP-13866 where netty-all-4.0.2?.Final.jar (from hdfs or
Spark) is ahead of 4.1.1.Final jar (used by hbase) in the classpath.
In pom.xml of Spark 1.6, you can see:
<artifactId>netty-all</artifactId>
<version>4.0.29.Final</version>
Please check the classpath and confirm.
Cheers
On Sat, Dec 10, 2016 at 3:08 AM, Yiannis Gkoufas <[email protected]>
wrote:
> Hi there,
>
> I have been trying to do a bulk load in HBase 2.0.0-SNAPSHOT using Spark
> 1.6.3.
> I followed the example in BulkLoadSuite, while the test works ok, I cannot
> make it work in my cluster.
> The error I have been getting has to do with a netty dependency conflict I
> guess because I get:
>
> no such method bytebuf retainedDuplicate
>
> Would really appreciate any hints.
>
> Thanks!
>