Still same problem my environment is now exact : hadoop : 2.3 zeppelin : zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3 spark : spark-1.3.1-bin-hadoop2.3
while checking version in zeppelin using sc.version, it sowing 1.4.0 pls help I am stuck.. thanks On Sun, Oct 11, 2015 at 10:58 AM, Ranveer kumar <ranveer.k.ku...@gmail.com> wrote: > I am using 0.5.0 now but earlier I was using own build with hadoop 2.6 > with yarn. > > Is 0.5.0 binary not compatible with hadoop 2.6? > > I am again going to build zeppelin with hadoop 2.6 also any pre build > zeppelin binary for hadoop 2.6 so directly download. > > Regards > On 11 Oct 2015 10:07, "Felix Cheung" <felixcheun...@hotmail.com> wrote: > >> Are you getting the last Zeppelin release binary from >> https://zeppelin.incubator.apache.org/download.html >> <https://zeppelin.incubator.apache.org/download.html?> or the Apache >> release page? >> >> It looks like the 0.5.0 release of Zeppelin is only built with Hadoop >> 2.3. We might be cutting another release soon or you might need to build >> you own... >> >> _____________________________ >> From: Ranveer Kumar <ranveer.s...@gmail.com> >> Sent: Saturday, October 10, 2015 9:00 PM >> Subject: Re: Problem loading data from HDFS >> To: <d...@zeppelin.incubator.apache.org> >> Cc: <users@zeppelin.incubator.apache.org> >> >> >> Hi Felix thanks for reply. >> >> I am using binary download from apache. >> >> I also tried using 2.6 hadoop, which build from source code. >> >> Which version of hadoop is compatible with zeppelin binary available on >> apache site. >> On 11 Oct 2015 09:05, "Felix Cheung" < felixcheun...@hotmail.com> wrote: >> >>> Is your Zeppelin built with Hadoop 2.6? >>> >>> >>> >>> >>> >>> >>> On Sat, Oct 10, 2015 at 7:35 PM -0700, "Ranveer kumar" < >>> ranveer.k.ku...@gmail.com> wrote: >>> Hi All, >>> >>> I am new in Zepplin and HDFS. I manage to install zeppelin and working >>> fine >>> while loading data from local directory . But when same I am trying to >>> load >>> from HDFS (install locally standalone mode). >>> >>> here is my code : >>> >>> val bankText = sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") >>> >>> above is working fine. >>> >>> but when trying form hdfs : >>> >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>") >>> >>> not working and giving error : >>> >>> java.lang.VerifyError: class >>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetAdditionalDatanodeRequestProto >>> >>> overrides final method >>> getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; at >>> java.lang.ClassLoader.defineClass1(Native Method) at >>> java.lang.ClassLoader.defineClass(ClassLoader.java:800) at >>> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) >>> at >>> java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at >>> java.net.URLClassLoader.access$100(URLClassLoader.java:71) >>> >>> complete code is : >>> >>> val bankText = sc.textFile("hdfs:// 127.0.0.1:9000/demo/ >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>csv >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>/ >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>bank-full.csv >>> < http://127.0.0.1:9000/demo/csv/bank-full.csv>") >>> // val bankText = >>> sc.textFile("/home/ranveer/Desktop/CSVs/bank-full.csv") >>> case class Bank(age:Integer, job:String, marital : String, education : >>> String, balance : Integer) >>> >>> val bank = bankText.map(s=>s.split(";")).filter(s=>s(0)!="\"age\"").map( >>> s=>Bank(s(0).toInt, >>> s(1).replaceAll("\"", ""), >>> s(2).replaceAll("\"", ""), >>> s(3).replaceAll("\"", ""), >>> s(5).replaceAll("\"", "").toInt >>> ) >>> ) >>> >>> // Below line works only in spark 1.3.0. >>> // For spark 1.1.x and spark 1.2.x, >>> // use bank.registerTempTable("bank") instead. >>> bank.toDF().registerTempTable("bank") >>> println(bankText.count()) >>> >>> my environment are : >>> >>> spark version : 1.3.1 with hadoop 2.6 >>> >>> zeppelin : binary from apache 0.5 >>> >>> hadoop version : 2.6 binary from apache >>> >>> java : 1.8 >>> >>> please help I am stuck here. >>> >>> thanks >>> >>> regards >>> >>> Ranveer >>> >> >> >>