Still same problem my environment is now exact :

hadoop : 2.3
zeppelin : zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3
spark : spark-1.3.1-bin-hadoop2.3

while checking version in zeppelin using sc.version, it sowing 1.4.0

pls help I am stuck..

thanks

On Sun, Oct 11, 2015 at 10:58 AM, Ranveer kumar <ranveer.k.ku...@gmail.com>
wrote:

> I am using 0.5.0 now but earlier I was using own build with hadoop 2.6
> with yarn.
>
> Is 0.5.0 binary not compatible with hadoop 2.6?
>
> I am again going to build zeppelin with hadoop 2.6 also any pre build
> zeppelin binary for hadoop 2.6 so directly download.
>
> Regards
> On 11 Oct 2015 10:07, "Felix Cheung" <felixcheun...@hotmail.com> wrote:
>
>> Are you getting the last Zeppelin release binary from
>> https://zeppelin.incubator.apache.org/download.html
>> <https://zeppelin.incubator.apache.org/download.html?> or the Apache
>> release page?
>>
>> It looks like the 0.5.0 release of Zeppelin is only built with Hadoop
>> 2.3. We might be cutting another release soon or you might need to build
>> you own...
>>
>> _____________________________
>> From: Ranveer Kumar <ranveer.s...@gmail.com>
>> Sent: Saturday, October 10, 2015 9:00 PM
>> Subject: Re: Problem loading data from HDFS
>> To: <d...@zeppelin.incubator.apache.org>
>> Cc: <users@zeppelin.incubator.apache.org>
>>
>>
>> Hi Felix thanks for reply.
>>
>> I am using binary download from apache.
>>
>> I also tried using 2.6 hadoop, which build from source code.
>>
>> Which version of hadoop is compatible with zeppelin binary available on
>> apache site.
>> On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote:
>>
>>> Is your Zeppelin built with Hadoop 2.6?
>>>
>>>
>>>
>>>
>>>
>>>
>>> On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" <
>>> ranveer.k.ku...@gmail.com> wrote:
>>> Hi All,
>>>
>>> I am new in Zepplin and HDFS. I manage to install zeppelin and working
>>> fine
>>> while loading data from local directory . But when same I am trying to
>>> load
>>> from HDFS (install locally standalone mode).
>>>
>>> here is my code :
>>>
>>> val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
>>>
>>> above is working fine.
>>>
>>> but when trying form hdfs :
>>>
>>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
>>>
>>> not working and giving error :
>>>
>>> java.lang.VerifyError: class
>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto
>>>
>>> overrides final method
>>> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at
>>> java.lang.ClassLoader.defineClass1(Native Method) at
>>> java.lang.ClassLoader.defineClass(ClassLoader.java:800) at
>>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>>> at
>>> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at
>>> java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>>
>>> complete code is :
>>>
>>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv
>>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>")
>>> // val bankText =
>>> sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv")
>>> case class Bank(age:Integer, job:String, marital : String, education :
>>> String, balance : Integer)
>>>
>>> val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map(
>>>     s=>Bank(s(0).toInt,
>>>             s(1).replaceAll("\"", ""),
>>>             s(2).replaceAll("\"", ""),
>>>             s(3).replaceAll("\"", ""),
>>>             s(5).replaceAll("\"", "").toInt
>>>         )
>>> )
>>>
>>> // Below line works only in spark 1.3.0.
>>> // For spark 1.1.x and spark 1.2.x,
>>> // use bank.registerTempTable("bank") instead.
>>> bank.toDF().registerTempTable("bank")
>>> println(bankText.count())
>>>
>>> my environment are :
>>>
>>> spark version : 1.3.1 with hadoop 2.6
>>>
>>> zeppelin : binary from apache 0.5
>>>
>>> hadoop version : 2.6 binary from apache
>>>
>>> java : 1.8
>>>
>>> please help I am stuck here.
>>>
>>> thanks
>>>
>>> regards
>>>
>>> Ranveer
>>>
>>
>>
>>

Reply via email to