Re: spark 1.1.0 - hbase 0.98.6-hadoop2 version - py4j.protocol.Py4JJavaError java.lang.ClassNotFoundException

2014-10-04 Thread Nick Pentreath
forgot to copy user list

On Sat, Oct 4, 2014 at 3:12 PM, Nick Pentreath 
wrote:

> what version did you put in the pom.xml?
>
> it does seem to be in Maven central:
> http://search.maven.org/#artifactdetails%7Corg.apache.hbase%7Chbase%7C0.98.6-hadoop2%7Cpom
>
> 
> org.apache.hbase
> hbase
> 0.98.6-hadoop2
> 
>
> Note you shouldn't need to rebuild Spark, I think just the example project
> via sbt examples/assembly
>
> On Fri, Oct 3, 2014 at 10:55 AM, serkan.dogan 
> wrote:
>
>> Hi,
>> I installed hbase-0.98.6-hadoop2. It's working not any problem with that.
>>
>>
>> When i am try to run spark hbase  python examples, (wordcount examples
>> working - not python issue)
>>
>>  ./bin/spark-submit  --master local --driver-class-path
>> ./examples/target/spark-examples_2.10-1.1.0.jar
>> ./examples/src/main/python/hbase_inputformat.py localhost myhbasetable
>>
>> the process exit with ClassNotFoundException...
>>
>> I search lots of blogs, sites all says "spark 1.1 version built with hbase
>> 0.94.6 rebuild with own hbase version".
>>
>> I try first,
>> change hbase version number - in pom.xml  -- nothing found maven central
>>
>> I try second,
>> compile hbase from src and copy hbase/lib folder hbase jars to
>> spark/lib_managed folder and edit spark-defaults.conf
>>
>> my spark-defaults.conf
>>
>> spark.executor.extraClassPath
>>
>> /home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-server-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-protocol-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-hadoop2-compat-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-client-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-commont-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/htrace-core-2.04.jar
>>
>>
>> My question is how i can work with hbase 0.98.6-hadoop2 with spark 1.1.0
>>
>> Here is the exception message
>>
>>
>> Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 14/10/03 11:27:15 WARN Utils: Your hostname, xxx.yyy.com resolves to a
>> loopback address: 127.0.0.1; using 1.1.1.1 instead (on interface eth0)
>> 14/10/03 11:27:15 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to
>> another address
>> 14/10/03 11:27:15 INFO SecurityManager: Changing view acls to: root,
>> 14/10/03 11:27:15 INFO SecurityManager: Changing modify acls to: root,
>> 14/10/03 11:27:15 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users with view permissions: Set(root, );
>> users
>> with modify permissions: Set(root, )
>> 14/10/03 11:27:16 INFO Slf4jLogger: Slf4jLogger started
>> 14/10/03 11:27:16 INFO Remoting: Starting remoting
>> 14/10/03 11:27:16 INFO Remoting: Remoting started; listening on addresses
>> :[akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
>> 14/10/03 11:27:16 INFO Remoting: Remoting now listens on addresses:
>> [akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
>> 14/10/03 11:27:16 INFO Utils: Successfully started service 'sparkDriver'
>> on
>> port 49256.
>> 14/10/03 11:27:16 INFO SparkEnv: Registering MapOutputTracker
>> 14/10/03 11:27:16 INFO SparkEnv: Registering BlockManagerMaster
>> 14/10/03 11:27:16 INFO DiskBlockManager: Created local directory at
>> /tmp/spark-local-20141003112716-298d
>> 14/10/03 11:27:16 INFO Utils: Successfully started service 'Connection
>> manager for block manager' on port 35106.
>> 14/10/03 11:27:16 INFO ConnectionManager: Bound socket to port 35106 with
>> id
>> = ConnectionManagerId(1-1-1-1-1.rev.mydomain.io,35106)
>> 14/10/03 11:27:16 INFO MemoryStore: MemoryStore started with capacity
>> 267.3
>> MB
>> 14/10/03 11:27:16 INFO BlockManagerMaster: Trying to register BlockManager
>> 14/10/03 11:27:16 INFO BlockManagerMasterActor: Registering block manager
>> 1-1-1-1-1.rev.mydomain.io:35106 with 267.3 MB RAM
>> 14/10/03 11:27:16 INFO BlockManagerMaster: Registered BlockManager
>> 14/10/03 11:27:16 INFO HttpFileServer: HTTP File server directory is
>> /tmp/spark-f60b0533-998f-4af2-a208-d04c571eab82
>> 14/10/03 11:27:16 INFO HttpServer: Starting HTTP Server
>> 14/10/03 11:27:16 INFO Utils: Successfully started service 'HTTP file
>> server' on port 49611.
>> 14/10/03 11:27:16 INFO Utils: Successfully started service 'SparkUI' on
>> port
>> 4040.
>> 14/10/03 11:27:16 INFO SparkUI: Started SparkUI at
>> http://1-1-1-1-1.rev.mydomain.io:4040
>> 14/10/03 11:27:16 INFO Utils: Copying
>>
>> /home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
>> to /tmp/spark-7232227a-0547-454e-9f68-805fa7b0c2f0/hbase_inputformat.py
>> 14/10/03 11:27:16 INFO SparkContext: Added file
>>
>> file:/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
>> at http://1.1.1.1:49611/files/hbase_inputformat.py with timestamp
>> 1412324836837
>> 14/10/03 11:27:16 INFO AkkaUtils: Connecting to HeartbeatReceiver:
>> akka.tcp://
>> spa

spark 1.1.0 - hbase 0.98.6-hadoop2 version - py4j.protocol.Py4JJavaError java.lang.ClassNotFoundException

2014-10-03 Thread serkan.dogan
Hi,
I installed hbase-0.98.6-hadoop2. It's working not any problem with that. 


When i am try to run spark hbase  python examples, (wordcount examples
working - not python issue)

 ./bin/spark-submit  --master local --driver-class-path
./examples/target/spark-examples_2.10-1.1.0.jar
./examples/src/main/python/hbase_inputformat.py localhost myhbasetable

the process exit with ClassNotFoundException...

I search lots of blogs, sites all says "spark 1.1 version built with hbase
0.94.6 rebuild with own hbase version".

I try first,
change hbase version number - in pom.xml  -- nothing found maven central

I try second,
compile hbase from src and copy hbase/lib folder hbase jars to
spark/lib_managed folder and edit spark-defaults.conf

my spark-defaults.conf

spark.executor.extraClassPath
/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-server-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-protocol-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-hadoop2-compat-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-client-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-commont-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/htrace-core-2.04.jar


My question is how i can work with hbase 0.98.6-hadoop2 with spark 1.1.0

Here is the exception message


Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
14/10/03 11:27:15 WARN Utils: Your hostname, xxx.yyy.com resolves to a
loopback address: 127.0.0.1; using 1.1.1.1 instead (on interface eth0)
14/10/03 11:27:15 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to
another address
14/10/03 11:27:15 INFO SecurityManager: Changing view acls to: root,
14/10/03 11:27:15 INFO SecurityManager: Changing modify acls to: root,
14/10/03 11:27:15 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(root, ); users
with modify permissions: Set(root, )
14/10/03 11:27:16 INFO Slf4jLogger: Slf4jLogger started
14/10/03 11:27:16 INFO Remoting: Starting remoting
14/10/03 11:27:16 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
14/10/03 11:27:16 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
14/10/03 11:27:16 INFO Utils: Successfully started service 'sparkDriver' on
port 49256.
14/10/03 11:27:16 INFO SparkEnv: Registering MapOutputTracker
14/10/03 11:27:16 INFO SparkEnv: Registering BlockManagerMaster
14/10/03 11:27:16 INFO DiskBlockManager: Created local directory at
/tmp/spark-local-20141003112716-298d
14/10/03 11:27:16 INFO Utils: Successfully started service 'Connection
manager for block manager' on port 35106.
14/10/03 11:27:16 INFO ConnectionManager: Bound socket to port 35106 with id
= ConnectionManagerId(1-1-1-1-1.rev.mydomain.io,35106)
14/10/03 11:27:16 INFO MemoryStore: MemoryStore started with capacity 267.3
MB
14/10/03 11:27:16 INFO BlockManagerMaster: Trying to register BlockManager
14/10/03 11:27:16 INFO BlockManagerMasterActor: Registering block manager
1-1-1-1-1.rev.mydomain.io:35106 with 267.3 MB RAM
14/10/03 11:27:16 INFO BlockManagerMaster: Registered BlockManager
14/10/03 11:27:16 INFO HttpFileServer: HTTP File server directory is
/tmp/spark-f60b0533-998f-4af2-a208-d04c571eab82
14/10/03 11:27:16 INFO HttpServer: Starting HTTP Server
14/10/03 11:27:16 INFO Utils: Successfully started service 'HTTP file
server' on port 49611.
14/10/03 11:27:16 INFO Utils: Successfully started service 'SparkUI' on port
4040.
14/10/03 11:27:16 INFO SparkUI: Started SparkUI at
http://1-1-1-1-1.rev.mydomain.io:4040
14/10/03 11:27:16 INFO Utils: Copying
/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
to /tmp/spark-7232227a-0547-454e-9f68-805fa7b0c2f0/hbase_inputformat.py
14/10/03 11:27:16 INFO SparkContext: Added file
file:/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
at http://1.1.1.1:49611/files/hbase_inputformat.py with timestamp
1412324836837
14/10/03 11:27:16 INFO AkkaUtils: Connecting to HeartbeatReceiver:
akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256/user/HeartbeatReceiver
Traceback (most recent call last):
  File
"/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py",
line 70, in 
conf=conf)
  File "/home/downloads/spark/spark-1.1.0/python/pyspark/context.py", line
471, in newAPIHadoopRDD
jconf, batchSize)
  File
"/usr/lib/python2.6/site-packages/py4j-0.8.2.1-py2.6.egg/py4j/java_gateway.py",
line 538, in __call__
self.target_id, self.name)
  File
"/usr/lib/python2.6/site-packages/py4j-0.8.2.1-py2.6.egg/py4j/protocol.py",
line 300, in get_return_value
format(target_id, '.', name), value)
py4j.protocol.Py4JJavaError: An error occurred while calling
z:org.apache.spark.api.python.PythonRDD.newAPIHadoopRDD.
: java.l

spark 1.1.0 - hbase 0.98.6-hadoop2 version - py4j.protocol.Py4JJavaError java.lang.ClassNotFoundException

2014-10-03 Thread serkan.dogan
Hi,
I installed hbase-0.98.6-hadoop2. It's working not any problem with that. 


When i am try to run spark hbase  python examples, (wordcount examples
working - not python issue)

 ./bin/spark-submit  --master local --driver-class-path
./examples/target/spark-examples_2.10-1.1.0.jar
./examples/src/main/python/hbase_inputformat.py localhost myhbasetable

the process exit with ClassNotFoundException...

I search lots of blogs, sites all says "spark 1.1 version built with hbase
0.94.6 rebuild with own hbase version".

I try first,
change hbase version number - in pom.xml  -- nothing found maven central

I try second,
compile hbase from src and copy hbase/lib folder hbase jars to
spark/lib_managed folder and edit spark-defaults.conf

my spark-defaults.conf

spark.executor.extraClassPath
/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-server-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-protocol-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-hadoop2-compat-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-client-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/hbase-commont-0.98.6-hadoop2.jar:/home/downloads/spark/spark-1.1.0/lib_managed/jars/htrace-core-2.04.jar


My question is how i can work with hbase 0.98.6-hadoop2 with spark 1.1.0

Here is the exception message


Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
14/10/03 11:27:15 WARN Utils: Your hostname, testmachine1.tracescore.com
resolves to a loopback address: 127.0.0.1; using 1.1.1.1 instead (on
interface eth0)
14/10/03 11:27:15 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to
another address
14/10/03 11:27:15 INFO SecurityManager: Changing view acls to: root,
14/10/03 11:27:15 INFO SecurityManager: Changing modify acls to: root,
14/10/03 11:27:15 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(root, ); users
with modify permissions: Set(root, )
14/10/03 11:27:16 INFO Slf4jLogger: Slf4jLogger started
14/10/03 11:27:16 INFO Remoting: Starting remoting
14/10/03 11:27:16 INFO Remoting: Remoting started; listening on addresses
:[akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
14/10/03 11:27:16 INFO Remoting: Remoting now listens on addresses:
[akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256]
14/10/03 11:27:16 INFO Utils: Successfully started service 'sparkDriver' on
port 49256.
14/10/03 11:27:16 INFO SparkEnv: Registering MapOutputTracker
14/10/03 11:27:16 INFO SparkEnv: Registering BlockManagerMaster
14/10/03 11:27:16 INFO DiskBlockManager: Created local directory at
/tmp/spark-local-20141003112716-298d
14/10/03 11:27:16 INFO Utils: Successfully started service 'Connection
manager for block manager' on port 35106.
14/10/03 11:27:16 INFO ConnectionManager: Bound socket to port 35106 with id
= ConnectionManagerId(1-1-1-1-1.rev.mydomain.io,35106)
14/10/03 11:27:16 INFO MemoryStore: MemoryStore started with capacity 267.3
MB
14/10/03 11:27:16 INFO BlockManagerMaster: Trying to register BlockManager
14/10/03 11:27:16 INFO BlockManagerMasterActor: Registering block manager
1-1-1-1-1.rev.mydomain.io:35106 with 267.3 MB RAM
14/10/03 11:27:16 INFO BlockManagerMaster: Registered BlockManager
14/10/03 11:27:16 INFO HttpFileServer: HTTP File server directory is
/tmp/spark-f60b0533-998f-4af2-a208-d04c571eab82
14/10/03 11:27:16 INFO HttpServer: Starting HTTP Server
14/10/03 11:27:16 INFO Utils: Successfully started service 'HTTP file
server' on port 49611.
14/10/03 11:27:16 INFO Utils: Successfully started service 'SparkUI' on port
4040.
14/10/03 11:27:16 INFO SparkUI: Started SparkUI at
http://1-1-1-1-1.rev.mydomain.io:4040
14/10/03 11:27:16 INFO Utils: Copying
/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
to /tmp/spark-7232227a-0547-454e-9f68-805fa7b0c2f0/hbase_inputformat.py
14/10/03 11:27:16 INFO SparkContext: Added file
file:/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py
at http://1.1.1.1:49611/files/hbase_inputformat.py with timestamp
1412324836837
14/10/03 11:27:16 INFO AkkaUtils: Connecting to HeartbeatReceiver:
akka.tcp://sparkdri...@1-1-1-1-1.rev.mydomain.io:49256/user/HeartbeatReceiver
Traceback (most recent call last):
  File
"/home/downloads/spark/spark-1.1.0/./examples/src/main/python/hbase_inputformat.py",
line 70, in 
conf=conf)
  File "/home/downloads/spark/spark-1.1.0/python/pyspark/context.py", line
471, in newAPIHadoopRDD
jconf, batchSize)
  File
"/usr/lib/python2.6/site-packages/py4j-0.8.2.1-py2.6.egg/py4j/java_gateway.py",
line 538, in __call__
self.target_id, self.name)
  File
"/usr/lib/python2.6/site-packages/py4j-0.8.2.1-py2.6.egg/py4j/protocol.py",
line 300, in get_return_value
format(target_id, '.', name), value)
py4j.protocol.Py4JJavaError: An error occurred while calling
z:org.apache.spark.api.python.PythonRDD.newAPIHad