Caused by: java.lang.IllegalAccessError: class
org.apache.hadoop.hdfs.web.HftpFileSystem cannot access its
superinterface
org.apache.hadoop.hdfs.web.TokenAspect$TokenManagementDelegator
This is the root cause, it seems that HBase 1.2 can't access interface
of Hadoop 3.1, so you should consider degrading Hadoop's version or
upgrading HBase's version.
----------------------------------------
Yun Zhang
Best regards!
2018-08-21 11:28 GMT+08:00 Mich Talebzadeh <mich.talebza...@gmail.com
<mailto:mich.talebza...@gmail.com>>:
Hi,
The Hadoop version is Hadoop 3.1.0. Hbase is 1.2.6 and Phoenix is
apache-phoenix-4.8.1-HBase-1.2-bin
In the past I had issues with Hbase 2 working with Hadoop 3.1 so I
had to use Hbase 1.2.6. The individual components work fine. In
other words I can do all operations on Hbase with Hadoop 3.1 and
Phoenix.
The issue I am facing is using both
org.apache.phoenix.mapreduce.CsvBulkLoadTool and
hbase.mapreduce.ImportTsv utilities.
So I presume the issue may be to do with both these command line
tools not working with Hadoop 3.1?
Thanks
Dr Mich Talebzadeh
LinkedIn
/https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>/
http://talebzadehmich.wordpress.com
<http://talebzadehmich.wordpress.com>
*Disclaimer:* Use it at your own risk.Any and all responsibility for
any loss, damage or destruction of data or any other property which
may arise from relying on this email's technical content is
explicitly disclaimed. The author will in no case be liable for any
monetary damages arising from such loss, damage or destruction.
On Tue, 21 Aug 2018 at 00:48, Sergey Soldatov
<sergey.solda...@gmail.com <mailto:sergey.solda...@gmail.com>> wrote:
If I read it correctly you are trying to use Phoenix and HBase
that were built against Hadoop 2 with Hadoop 3. Is HBase was the
only component you have upgraded?
Thanks,
Sergey
On Mon, Aug 20, 2018 at 1:42 PM Mich Talebzadeh
<mich.talebza...@gmail.com <mailto:mich.talebza...@gmail.com>>
wrote:
Here you go
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client
environment:java.library.path=/home/hduser/hadoop-3.1.0/lib
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:java.io.tmpdir=/tmp
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:java.compiler=<NA>
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:os.name <http://os.name>=Linux
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:os.arch=amd64
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:os.version=3.10.0-862.3.2.el7.x86_64
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:user.name <http://user.name>=hduser
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client environment:user.home=/home/hduser
2018-08-20 18:29:47,248 INFO [main] zookeeper.ZooKeeper:
Client
environment:user.dir=/data6/hduser/streaming_data/2018-08-20
2018-08-20 18:29:47,249 INFO [main] zookeeper.ZooKeeper:
Initiating client connection, connectString=rhes75:2181
sessionTimeout=90000 watcher=hconnection-0x493d44230x0,
quorum=rhes75:2181, baseZNode=/hbase
2018-08-20 18:29:47,261 INFO [main-SendThread(rhes75:2181)]
zookeeper.ClientCnxn: Opening socket connection to server
rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>.
Will not attempt to authenticate using SASL (unknown error)
2018-08-20 18:29:47,264 INFO [main-SendThread(rhes75:2181)]
zookeeper.ClientCnxn: Socket connection established to
rhes75/50.140.197.220:2181 <http://50.140.197.220:2181>,
initiating session
2018-08-20 18:29:47,281 INFO [main-SendThread(rhes75:2181)]
zookeeper.ClientCnxn: Session establishment complete on
server rhes75/50.140.197.220:2181
<http://50.140.197.220:2181>, sessionid = 0x1002ea99eed0077,
negotiated timeout = 40000
Exception in thread "main" java.sql.SQLException: ERROR 103
(08004): Unable to establish connection.
at
org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455)
at
org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:386)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl.access$300(ConnectionQueryServicesImpl.java:222)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2318)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2294)
at
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2294)
at
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:232)
at
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:147)
at
org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202)
at
java.sql.DriverManager.getConnection(DriverManager.java:664)
at
java.sql.DriverManager.getConnection(DriverManager.java:208)
at
org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:340)
at
org.apache.phoenix.util.QueryUtil.getConnection(QueryUtil.java:332)
at
org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(AbstractBulkLoadTool.java:209)
at
org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(AbstractBulkLoadTool.java:183)
at
org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
at
org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
at
org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:101)
at
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:308)
at org.apache.hadoop.util.RunJar.main(RunJar.java:222)
Caused by: java.io.IOException:
java.lang.reflect.InvocationTargetException
at
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:240)
at
org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:431)
at
org.apache.hadoop.hbase.client.ConnectionManager.createConnectionInternal(ConnectionManager.java:340)
at
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:144)
at
org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
at
org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:383)
... 23 more
Caused by: java.lang.reflect.InvocationTargetException
at
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at
java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:238)
... 28 more
Caused by: java.lang.IllegalAccessError: class
org.apache.hadoop.hdfs.web.HftpFileSystem cannot access its
superinterface
org.apache.hadoop.hdfs.web.TokenAspect$TokenManagementDelegator
at java.lang.ClassLoader.defineClass1(Native Method)
at
java.lang.ClassLoader.defineClass(ClassLoader.java:763)
at
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at
java.net.URLClassLoader.defineClass(URLClassLoader.java:467)
at
java.net.URLClassLoader.access$100(URLClassLoader.java:73)
at
java.net.URLClassLoader$1.run(URLClassLoader.java:368)
at
java.net.URLClassLoader$1.run(URLClassLoader.java:362)
at
java.security.AccessController.doPrivileged(Native Method)
at
java.net.URLClassLoader.findClass(URLClassLoader.java:361)
at
java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at
java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:348)
at
java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370)
at
java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
at
java.util.ServiceLoader$1.next(ServiceLoader.java:480)
at
org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:3268)
at
org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:3313)
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3352)
at
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403)
at
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371)
at
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477)
at
org.apache.hadoop.fs.Path.getFileSystem(Path.java:361)
at
org.apache.hadoop.hbase.util.DynamicClassLoader.initTempDir(DynamicClassLoader.java:120)
at
org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:98)
at
org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:242)
at
org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
at
org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
at
org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:105)
at
org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:905)
at
org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:648)
... 33 more
Thanks
Dr Mich Talebzadeh
LinkedIn
/https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>/
http://talebzadehmich.wordpress.com
<http://talebzadehmich.wordpress.com>
*Disclaimer:* Use it at your own risk.Any and all
responsibility for any loss, damage or destruction of data
or any other property which may arise from relying on this
email's technical content is explicitly disclaimed. The
author will in no case be liable for any monetary damages
arising from such loss, damage or destruction.
On Mon, 20 Aug 2018 at 21:24, Josh Elser <els...@apache.org
<mailto:els...@apache.org>> wrote:
(-cc user@hbase, +bcc user@hbase)
How about the rest of the stacktrace? You didn't share
the cause.
On 8/20/18 1:35 PM, Mich Talebzadeh wrote:
>
> This was working fine before my Hbase upgrade to 1.2.6
>
> I have Hbase version 1.2.6 and Phoenix
> version apache-phoenix-4.8.1-HBase-1.2-bin
>
> This command bulkloading into Hbase through phoenix
failsnow fails
>
>
HADOOP_CLASSPATH=${HOME}/jars/hbase-protocol-1.2.6.jar:${HBASE_HOME}/conf
hadoop
> jar ${HBASE_HOME}/lib/phoenix-4.8.1-HBase-1.2-client.jar
> org.apache.phoenix.mapreduce.CsvBulkLoadTool --table
${TABLE_NAME}
> --input
hdfs://rhes75:9000/${REFINED_HBASE_SUB_DIR}/${FILE_NAME}_${dir}.txt
>
> hadoop jar
>
/data6/hduser/hbase-1.2.6/lib/phoenix-4.8.1-HBase-1.2-client.jar
> org.apache.phoenix.mapreduce.CsvBulkLoadTool --table
> MARKETDATAHBASEBATCH --input
>
hdfs://rhes75:9000//data/prices/2018-08-20_refined/populate_Phoenix_table_MARKETDATAHBASEBATCH_2018-08-20.txt
> +
>
HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.6.jar:/data6/hduser/hbase-1.2.6/conf
>
>
> With the following error
>
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
>
environment:java.library.path=/home/hduser/hadoop-3.1.0/lib
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:java.io.tmpdir=/tmp
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:java.compiler=<NA>
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:os.name <http://os.name>
<http://os.name>=Linux
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:os.arch=amd64
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:os.version=3.10.0-862.3.2.el7.x86_64
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:user.name <http://user.name>
<http://user.name>=hduser
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
> environment:user.home=/home/hduser
> 2018-08-20 18:29:47,248 INFO [main]
zookeeper.ZooKeeper: Client
>
environment:user.dir=/data6/hduser/streaming_data/2018-08-20
> 2018-08-20 18:29:47,249 INFO [main]
zookeeper.ZooKeeper: Initiating
> client connection, connectString=rhes75:2181
sessionTimeout=90000
> watcher=hconnection-0x493d44230x0,
quorum=rhes75:2181, baseZNode=/hbase
> 2018-08-20 18:29:47,261 INFO
[main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Opening socket connection to
server
> rhes75/50.140.197.220:2181
<http://50.140.197.220:2181>
<http://50.140.197.220:2181>. Will not
> attempt to authenticate using SASL (unknown error)
> 2018-08-20 18:29:47,264 INFO
[main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Socket connection established to
> rhes75/50.140.197.220:2181
<http://50.140.197.220:2181>
<http://50.140.197.220:2181>, initiating session
> 2018-08-20 18:29:47,281 INFO
[main-SendThread(rhes75:2181)]
> zookeeper.ClientCnxn: Session establishment complete
on server
> rhes75/50.140.197.220:2181
<http://50.140.197.220:2181>
<http://50.140.197.220:2181>, sessionid =
> 0x1002ea99eed0077, negotiated timeout = 40000
> Exception in thread "main" java.sql.SQLException:
ERROR 103 (08004):
> Unable to establish connection.
> at
>
org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:455)
>
> Any thoughts?
>
> Thanks
>
> Dr Mich Talebzadeh
>
> LinkedIn
>
/https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw/>
>
> http://talebzadehmich.wordpress.com
<http://talebzadehmich.wordpress.com>
>
>
> *Disclaimer:* Use it at your own risk.Any and all
responsibility for any
> loss, damage or destruction of data or any other
property which may
> arise from relying on this email's technical content
is explicitly
> disclaimed. The author will in no case be liable for
any monetary
> damages arising from such loss, damage or destruction.
>