Thanks, I just did it. Some progresses, but not yet there.
Now in the paragraph:

--3--
%sql
select * from sms

I'm getting the following exception. I googled other people getting similar
exception but did not find root cause. Any idea of what's going on?
Note that my env is very simple, I just built zeppelin as per your
suggestion, started it, and tried my supersimple notebook.

java.lang.ClassNotFoundException:
com.databricks.spark.csv.CsvRelation$$anonfun$tokenRdd$1$$anonfun$1 at
java.net.URLClassLoader$1.run(URLClassLoader.java:366) at
java.net.URLClassLoader$1.run(URLClassLoader.java:355) at
java.security.AccessController.doPrivileged(Native Method) at
java.net.URLClassLoader.findClass(URLClassLoader.java:354) at
java.lang.ClassLoader.loadClass(ClassLoader.java:425) at
sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at
java.lang.ClassLoader.loadClass(ClassLoader.java:358) at
java.lang.Class.forName0(Native Method) at
java.lang.Class.forName(Class.java:274) at
org.apache.spark.util.InnerClosureFinder$$anon$4.visitMethodInsn(ClosureCleaner.scala:455)
at
com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown
Source) at
com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.accept(Unknown
Source) at
org.apache.spark.util.ClosureCleaner$.getInnerClosureClasses(ClosureCleaner.scala:101)
at
org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:197)
at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:132) at
org.apache.spark.SparkContext.clean(SparkContext.scala:1893) at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:683) at
org.apache.spark.rdd.RDD$$anonfun$mapPartitions$1.apply(RDD.scala:682) at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:286) at
org.apache.spark.rdd.RDD.mapPartitions(RDD.scala:682) at
com.databricks.spark.csv.CsvRelation.tokenRdd(CsvRelation.scala:90) at
com.databricks.spark.csv.CsvRelation.buildScan(CsvRelation.scala:105) at
org.apache.spark.sql.sources.DataSourceStrategy$.apply(DataSourceStrategy.scala:101)
at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
at
org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:314)
at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
at
org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:943)
at
org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:941)
at
org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:947)
at
org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:947)
at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1269) at
org.apache.spark.sql.DataFrame.head(DataFrame.scala:1203) at
org.apache.spark.sql.DataFrame.take(DataFrame.scala:1262) at
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606) at
org.apache.zeppelin.spark.ZeppelinContext.showDF(ZeppelinContext.java:300)
at
org.apache.zeppelin.spark.SparkSqlInterpreter.interpret(SparkSqlInterpreter.java:142)
at
org.apache.zeppelin.interpreter.ClassloaderInterpreter.interpret(ClassloaderInterpreter.java:57)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:276)
at org.apache.zeppelin.scheduler.Job.run(Job.java:170) at
org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:118)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262) at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)


On Mon, Oct 19, 2015 at 1:59 PM, moon soo Lee <m...@apache.org> wrote:

> You'll need to build your self.
> Please let me know if you have any problem on building Zeppelin.
>
> Thanks,
> moon
>
>
> On Mon, Oct 19, 2015 at 8:32 PM michele crudele <mic.crud...@gmail.com>
> wrote:
>
>> Thanks moon,
>>
>> is there a repo where I can download the 0.6.0-SNAPHOT or do I have to
>> build zeppelin myself ? Thanks again for your help.
>>
>> - michele
>>
>>
>> On Mon, Oct 19, 2015 at 7:41 AM, moon soo Lee <m...@apache.org> wrote:
>>
>>> Hi Michele,
>>>
>>> Thanks for sharing the problem.
>>> I have tested your code on both 0.5.0 and 0.6.0-SNAPSHOT.
>>> I have the same problem on 0.5.0 but 0.6.0-SNAPSHOT runs smoothly.
>>>
>>> So, you can try 0.6.0-SNAPSHOT until next release is out.
>>> Or if you want to see what's going on with 0.5.0, you'll need to apply
>>> this commit
>>> https://github.com/apache/incubator-zeppelin/commit/d0a30435414726e7fa6d8b8e106e4b6ddb46da67
>>>  to
>>> see exception in your notebook.
>>>
>>> Best,
>>> moon
>>>
>>> On Sat, Oct 17, 2015 at 7:18 AM michele crudele <mic.crud...@gmail.com>
>>> wrote:
>>>
>>>> I've created a supersimple notebook:
>>>>
>>>> -- 1 --
>>>> %dep
>>>>
>>>> z.reset
>>>> z.load("com.databricks:spark-csv_2.10:1.2.0")
>>>>
>>>> --2--
>>>> %spark
>>>>
>>>> val smsFile = "/home/barabba/data/SMSSpamCollection.csv"
>>>> sqlContext.load("com.databricks.spark.csv", Map("path" -> smsFile,
>>>> "header" -> "true", "delimiter" -> "|")).registerTempTable("sms")
>>>>
>>>> --3--
>>>> %sql
>>>> select * from sms
>>>>
>>>> 1 and 2 runs fine, while 3 displays the
>>>> java.lang.reflect.InvocationTargetException. I cannot find anything useful
>>>> in the logs.
>>>> If I add the paragraph
>>>>
>>>> --4--
>>>> %spark
>>>> sqlContext.sql("select * from sms").show
>>>>
>>>> it works correctly, showing the top rows of the table.
>>>> +--+----+--------------------+
>>>> |id|type|                text|
>>>> +--+----+--------------------+
>>>> | 0| ham|Go until jurong p...|
>>>> | 1| ham|Ok lar... Joking ...|
>>>> | 2|spam|Free entry in 2 a...|
>>>> | 3| ham|U dun say so earl...|
>>>> | 4| ham|Nah I don't think...|
>>>> | 5|spam|FreeMsg Hey there...|
>>>> | 6| ham|Even my brother i...|
>>>> | 7| ham|As per your reque...|
>>>> ...
>>>>
>>>> I'm using zeppelin-0.5.0-incubating-bin-spark-1.3.1_hadoop-2.3.
>>>>
>>>>
>>>> Any idea of what's going on? Thanks
>>>> - michele
>>>>
>>>>
>>

Reply via email to