Thanks, I just did it. Some progresses, but not yet there. Now in the paragraph:
--3-- %sql select * from sms I'm getting the following exception. I googled other people getting similar exception but did not find root cause. Any idea of what's going on? Note that my env is very simple, I just built zeppelin as per your suggestion, started it, and tried my supersimple notebook. java.lang.ClassNotFoundException: com.databricks.spark.csv.CsvRelation$$anonfun$tokenRdd$1$$anonfun$1 at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at java.lang.Class.forName0(Native Method) at java.lang.Class.forName(Class.java:274) at org.apache.spark.util.InnerClosureFinder$$anon$4.visitMethodInsn(ClosureCleaner.scala:455) at com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown Source) at com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown Source) at org.apache.spark.util.ClosureCleaner$.getInnerClosureClasses(ClosureCleaner.scala:101) at org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:197) at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:132) at org.apache.spark.SparkContext.clean(SparkContext.scala:1893) at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:683) at org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:682) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108) at org.apache.spark.rdd.RDD.withScope(RDD.scala:286) at org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:682) at com.databricks.spark.csv.CsvRelation.tokenRdd(CsvRelation.scala:90) at com.databricks.spark.csv.CsvRelation.buildScan(CsvRelation.scala:105) at org.apache.spark.sql.sources.DataSourceStrategy$.apply(DataSourceStrategy.scala:101) at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58) at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58) at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59) at org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54) at org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:314) at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58) at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58) at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59) at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:943) at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:941) at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:947) at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:947) at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1269) at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1203) at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1262) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.zeppelin.spark.ZeppelinContext.showDF(ZeppelinContext.java:300) at org.apache.zeppelin.spark.SparkSqlInterpreter.interpret(SparkSqlInterpreter.java:142) at org.apache.zeppelin.interpreter.ClassloaderInterpreter.interpret(ClassloaderInterpreter.java:57) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:276) at org.apache.zeppelin.scheduler.Job.run(Job.java:170) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask.run(FutureTask.java:262) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) On Mon, Oct 19, 2015 at 1:59 PM, moon soo Lee <m...@apache.org> wrote: > You'll need to build your self. > Please let me know if you have any problem on building Zeppelin. > > Thanks, > moon > > > On Mon, Oct 19, 2015 at 8:32 PM michele crudele <mic.crud...@gmail.com> > wrote: > >> Thanks moon, >> >> is there a repo where I can download the 0.6.0-SNAPHOT or do I have to >> build zeppelin myself ? Thanks again for your help. >> >> - michele >> >> >> On Mon, Oct 19, 2015 at 7:41 AM, moon soo Lee <m...@apache.org> wrote: >> >>> Hi Michele, >>> >>> Thanks for sharing the problem. >>> I have tested your code on both 0.5.0 and 0.6.0-SNAPSHOT. >>> I have the same problem on 0.5.0 but 0.6.0-SNAPSHOT runs smoothly. >>> >>> So, you can try 0.6.0-SNAPSHOT until next release is out. >>> Or if you want to see what's going on with 0.5.0, you'll need to apply >>> this commit >>> https://github.com/apache/incubator-zeppelin/commit/d0a30435414726e7fa6d8b8e106e4b6ddb46da67 >>> to >>> see exception in your notebook. >>> >>> Best, >>> moon >>> >>> On Sat, Oct 17, 2015 at 7:18 AM michele crudele <mic.crud...@gmail.com> >>> wrote: >>> >>>> I've created a supersimple notebook: >>>> >>>> -- 1 -- >>>> %dep >>>> >>>> z.reset >>>> z.load("com.databricks:spark-csv_2.10:1.2.0") >>>> >>>> --2-- >>>> %spark >>>> >>>> val smsFile = "/home/barabba/data/SMSSpamCollection.csv" >>>> sqlContext.load("com.databricks.spark.csv", Map("path" -> smsFile, >>>> "header" -> "true", "delimiter" -> "|")).registerTempTable("sms") >>>> >>>> --3-- >>>> %sql >>>> select * from sms >>>> >>>> 1 and 2 runs fine, while 3 displays the >>>> java.lang.reflect.InvocationTargetException. I cannot find anything useful >>>> in the logs. >>>> If I add the paragraph >>>> >>>> --4-- >>>> %spark >>>> sqlContext.sql("select * from sms").show >>>> >>>> it works correctly, showing the top rows of the table. >>>> +--+----+--------------------+ >>>> |id|type| text| >>>> +--+----+--------------------+ >>>> | 0| ham|Go until jurong p...| >>>> | 1| ham|Ok lar... Joking ...| >>>> | 2|spam|Free entry in 2 a...| >>>> | 3| ham|U dun say so earl...| >>>> | 4| ham|Nah I don't think...| >>>> | 5|spam|FreeMsg Hey there...| >>>> | 6| ham|Even my brother i...| >>>> | 7| ham|As per your reque...| >>>> ... >>>> >>>> I'm using zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3. >>>> >>>> >>>> Any idea of what's going on? Thanks >>>> - michele >>>> >>>> >>