Re: How to disable creating a new notebook for few users

2021-07-02 Thread Sebastian Albrecht
Am Fr., 2. Juli 2021 um 01:23 Uhr schrieb Jeff Zhang :

> I don't think zeppelin has such a feature yet.
>
> Great Info  于2021年7月2日周五 上午2:42写道:
>
>> I am using Zeppelin  0.8.2, different kinds of users use our zeppelin.
>> There is one category of users wherein they just need to run the notebook.
>> I know how to set the authorization to disable 'edit or running' the
>> notebook but I did not find the option to disable creating a new notebook
>> for these users.
>>
>
Fyi, i already created a ticket as feature request for this
https://issues.apache.org/jira/browse/ZEPPELIN-4716



>
>> Thanks,
>> Cod
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>


Re: Question: Adding Dependencies in with the Spark Interpreter with Kubernetes

2020-05-14 Thread Sebastian Albrecht
Am Mi., 13. Mai 2020 um 21:59 Uhr schrieb Hetul Patel :

>
> Are dependency downloads supported with zeppelin and spark over
> kubernetes? Or am I required to add the dependency jars directly to my
> spark docker image and add them to the classpath?
>
>
Hi Hetu,
i don't use docker but to connect to my cassandra from the spark cluster i
have to set SPARK_SUBMIT_OPTIONS='--packages
com.datastax.spark:spark-cassandra-connector_2.11:2.4.3'

HTH,
Sebastian.


> Thanks,
> Hetu
>


Notebook permissions

2020-04-02 Thread Sebastian Albrecht
Hi,
i want logged in users allow to only view and use existing notebooks, but
not to create new ones. Is that possible somehow?

Thx+BR,
Sebastian.


Re: Spark performance after update to 0.8.2

2020-03-24 Thread Sebastian Albrecht
Hi Jeff,

Am Di., 24. März 2020 um 09:39 Uhr schrieb Jeff Zhang :

> Do you use the same version of spark for zeppelin 0.7 & 0.8 ? What kind of
> job do you run ? Is it easy to reproduce it ?
>

Zeppelin comes with spark 2.1.0 in 0.7.3 and spark 2.2.1 in 0.8.2, so
different version.
In this case it is a 'cache table ...' statement in the %spark.sql
interpreter and i see the larger delay in task executions everytime i
execute it with the different zeppelin versions.

BR,
Sebastian.


>
> Sebastian Albrecht  于2020年3月24日周二
> 下午4:24写道:
>
>> Hi,
>> i just upgraded from zeppelin 0.7.3 to 0.8.2. I am using zeppelin for
>> spark in local mode and 0.8.2 came with spark updated to 2.2.1. I now
>> experience slower spark job executions. I can show it in these log
>> snippets: task executions with 0.7.3: milliseconds, 0.8.2: seconds
>>
>> 0.7.3
>> INFO [2020-03-12 00:01:22,105] ({Executor task launch worker-2}
>> Logging.scala[logInfo]:54) - Code generated in 24.742647 ms
>>  INFO [2020-03-12 00:01:22,139] ({Executor task launch worker-2}
>> Logging.scala[logInfo]:54) - Code generated in 23.640976 ms
>>  INFO [2020-03-12 00:01:22,174] ({Executor task launch worker-2}
>> Logging.scala[logInfo]:54) - Code generated in 11.667064 ms
>>  INFO [2020-03-12 00:01:22,193] ({Executor task launch worker-0}
>> Logging.scala[logInfo]:54) - Code generated in 14.959374 ms
>>  INFO [2020-03-12 00:01:22,872] ({Executor task launch worker-6}
>> Logging.scala[logInfo]:54) - Finished task 6.0 in stage 8.0 (TID 14). 2593
>> bytes result sent to driver
>>  INFO [2020-03-12 00:01:22,873] ({dispatcher-event-loop-2}
>> Logging.scala[logInfo]:54) - Starting task 8.0 in stage 8.0 (TID 16,
>> localhost, executor driver, partition 8, NODE_LOCAL, 9086 bytes)
>>  INFO [2020-03-12 00:01:22,875] ({Executor task launch worker-6}
>> Logging.scala[logInfo]:54) - Running task 8.0 in stage 8.0 (TID 16)
>>  INFO [2020-03-12 00:01:22,876] ({task-result-getter-0}
>> Logging.scala[logInfo]:54) - Finished task 6.0 in stage 8.0 (TID 14) in
>> 2061 ms on localhost (executor driver) (1/97)
>>  INFO [2020-03-12 00:01:22,938] ({Executor task launch worker-2}
>> Logging.scala[logInfo]:54) - Finished task 2.0 in stage 8.0 (TID 10). 2593
>> bytes result sent to driver
>>  INFO [2020-03-12 00:01:22,939] ({dispatcher-event-loop-7}
>> Logging.scala[logInfo]:54) - Starting task 9.0 in stage 8.0 (TID 17,
>> localhost, executor driver, partition 9, NODE_LOCAL, 8838 bytes)
>>  INFO [2020-03-12 00:01:22,946] ({task-result-getter-1}
>> Logging.scala[logInfo]:54) - Finished task 2.0 in stage 8.0 (TID 10) in
>> 2135 ms on localhost (executor driver) (2/97)
>>  INFO [2020-03-12 00:01:22,962] ({Executor task launch worker-8}
>> Logging.scala[logInfo]:54) - Running task 9.0 in stage 8.0 (TID 17)
>>  INFO [2020-03-12 00:01:22,983] ({Executor task launch worker-3}
>> Logging.scala[logInfo]:54) - Finished task 3.0 in stage 8.0 (TID 11). 2506
>> bytes result sent to driver
>>  INFO [2020-03-12 00:01:22,984] ({dispatcher-event-loop-3}
>> Logging.scala[logInfo]:54) - Starting task 10.0 in stage 8.0 (TID 18,
>> localhost, executor driver, partition 10, NODE_LOCAL, 8839 bytes)
>>  INFO [2020-03-12 00:01:22,985] ({task-result-getter-2}
>> Logging.scala[logInfo]:54) - Finished task 3.0 in stage 8.0 (TID 11) in
>> 2173 ms on localhost (executor driver) (3/97)
>>  INFO [2020-03-12 00:01:22,986] ({Executor task launch worker-2}
>> Logging.scala[logInfo]:54) - Running task 10.0 in stage 8.0 (TID 18)
>>  INFO [2020-03-12 00:01:23,068] ({Executor task launch worker-0}
>> Logging.scala[logInfo]:54) - Finished task 0.0 in stage 8.0 (TID 8). 2506
>> bytes result sent to driver
>>  INFO [2020-03-12 00:01:23,069] ({dispatcher-event-loop-6}
>> Logging.scala[logInfo]:54) - Starting task 11.0 in stage 8.0 (TID 19,
>> localhost, executor driver, partition 11, NODE_LOCAL, 9333 bytes)
>>  INFO [2020-03-12 00:01:23,070] ({Executor task launch worker-3}
>> Logging.scala[logInfo]:54) - Running task 11.0 in stage 8.0 (TID 19)
>>  INFO [2020-03-12 00:01:23,070] ({task-result-getter-3}
>> Logging.scala[logInfo]:54) - Finished task 0.0 in stage 8.0 (TID 8) in 2262
>> ms on localhost (executor driver) (4/97)
>>  INFO [2020-03-12 00:01:23,088] ({Executor task launch worker-7}
>> Logging.scala[logInfo]:54) - Finished task 7.0 in stage 8.0 (TID 15). 2506
>> bytes result sent to driver
>>  INFO [2020-03-12 00:01:23,089] ({dispatcher-event-loop-2}
>> Logging.scala[logInfo]:54) - Starting task 12.0 in stage 8.0 (TID 20,
>> localhost, executor driver, partition 12, NODE_LOCAL, 8591 bytes)
>>  INFO [2020-03-1

Spark performance after update to 0.8.2

2020-03-24 Thread Sebastian Albrecht
Hi,
i just upgraded from zeppelin 0.7.3 to 0.8.2. I am using zeppelin for spark
in local mode and 0.8.2 came with spark updated to 2.2.1. I now experience
slower spark job executions. I can show it in these log snippets: task
executions with 0.7.3: milliseconds, 0.8.2: seconds

0.7.3
INFO [2020-03-12 00:01:22,105] ({Executor task launch worker-2}
Logging.scala[logInfo]:54) - Code generated in 24.742647 ms
 INFO [2020-03-12 00:01:22,139] ({Executor task launch worker-2}
Logging.scala[logInfo]:54) - Code generated in 23.640976 ms
 INFO [2020-03-12 00:01:22,174] ({Executor task launch worker-2}
Logging.scala[logInfo]:54) - Code generated in 11.667064 ms
 INFO [2020-03-12 00:01:22,193] ({Executor task launch worker-0}
Logging.scala[logInfo]:54) - Code generated in 14.959374 ms
 INFO [2020-03-12 00:01:22,872] ({Executor task launch worker-6}
Logging.scala[logInfo]:54) - Finished task 6.0 in stage 8.0 (TID 14). 2593
bytes result sent to driver
 INFO [2020-03-12 00:01:22,873] ({dispatcher-event-loop-2}
Logging.scala[logInfo]:54) - Starting task 8.0 in stage 8.0 (TID 16,
localhost, executor driver, partition 8, NODE_LOCAL, 9086 bytes)
 INFO [2020-03-12 00:01:22,875] ({Executor task launch worker-6}
Logging.scala[logInfo]:54) - Running task 8.0 in stage 8.0 (TID 16)
 INFO [2020-03-12 00:01:22,876] ({task-result-getter-0}
Logging.scala[logInfo]:54) - Finished task 6.0 in stage 8.0 (TID 14) in
2061 ms on localhost (executor driver) (1/97)
 INFO [2020-03-12 00:01:22,938] ({Executor task launch worker-2}
Logging.scala[logInfo]:54) - Finished task 2.0 in stage 8.0 (TID 10). 2593
bytes result sent to driver
 INFO [2020-03-12 00:01:22,939] ({dispatcher-event-loop-7}
Logging.scala[logInfo]:54) - Starting task 9.0 in stage 8.0 (TID 17,
localhost, executor driver, partition 9, NODE_LOCAL, 8838 bytes)
 INFO [2020-03-12 00:01:22,946] ({task-result-getter-1}
Logging.scala[logInfo]:54) - Finished task 2.0 in stage 8.0 (TID 10) in
2135 ms on localhost (executor driver) (2/97)
 INFO [2020-03-12 00:01:22,962] ({Executor task launch worker-8}
Logging.scala[logInfo]:54) - Running task 9.0 in stage 8.0 (TID 17)
 INFO [2020-03-12 00:01:22,983] ({Executor task launch worker-3}
Logging.scala[logInfo]:54) - Finished task 3.0 in stage 8.0 (TID 11). 2506
bytes result sent to driver
 INFO [2020-03-12 00:01:22,984] ({dispatcher-event-loop-3}
Logging.scala[logInfo]:54) - Starting task 10.0 in stage 8.0 (TID 18,
localhost, executor driver, partition 10, NODE_LOCAL, 8839 bytes)
 INFO [2020-03-12 00:01:22,985] ({task-result-getter-2}
Logging.scala[logInfo]:54) - Finished task 3.0 in stage 8.0 (TID 11) in
2173 ms on localhost (executor driver) (3/97)
 INFO [2020-03-12 00:01:22,986] ({Executor task launch worker-2}
Logging.scala[logInfo]:54) - Running task 10.0 in stage 8.0 (TID 18)
 INFO [2020-03-12 00:01:23,068] ({Executor task launch worker-0}
Logging.scala[logInfo]:54) - Finished task 0.0 in stage 8.0 (TID 8). 2506
bytes result sent to driver
 INFO [2020-03-12 00:01:23,069] ({dispatcher-event-loop-6}
Logging.scala[logInfo]:54) - Starting task 11.0 in stage 8.0 (TID 19,
localhost, executor driver, partition 11, NODE_LOCAL, 9333 bytes)
 INFO [2020-03-12 00:01:23,070] ({Executor task launch worker-3}
Logging.scala[logInfo]:54) - Running task 11.0 in stage 8.0 (TID 19)
 INFO [2020-03-12 00:01:23,070] ({task-result-getter-3}
Logging.scala[logInfo]:54) - Finished task 0.0 in stage 8.0 (TID 8) in 2262
ms on localhost (executor driver) (4/97)
 INFO [2020-03-12 00:01:23,088] ({Executor task launch worker-7}
Logging.scala[logInfo]:54) - Finished task 7.0 in stage 8.0 (TID 15). 2506
bytes result sent to driver
 INFO [2020-03-12 00:01:23,089] ({dispatcher-event-loop-2}
Logging.scala[logInfo]:54) - Starting task 12.0 in stage 8.0 (TID 20,
localhost, executor driver, partition 12, NODE_LOCAL, 8591 bytes)
 INFO [2020-03-12 00:01:23,090] ({task-result-getter-0}
Logging.scala[logInfo]:54) - Finished task 7.0 in stage 8.0 (TID 15) in
2275 ms on localhost (executor driver) (5/97)
 INFO [2020-03-12 00:01:23,090] ({Executor task launch worker-7}
Logging.scala[logInfo]:54) - Running task 12.0 in stage 8.0 (TID 20)
 INFO [2020-03-12 00:01:23,266] ({Executor task launch worker-1}
Logging.scala[logInfo]:54) - Finished task 1.0 in stage 8.0 (TID 9). 2506
bytes result sent to driver
 INFO [2020-03-12 00:01:23,267] ({dispatcher-event-loop-7}
Logging.scala[logInfo]:54) - Starting task 13.0 in stage 8.0 (TID 21,
localhost, executor driver, partition 13, NODE_LOCAL, 9458 bytes)
 INFO [2020-03-12 00:01:23,268] ({task-result-getter-1}
Logging.scala[logInfo]:54) - Finished task 1.0 in stage 8.0 (TID 9) in 2457
ms on localhost (executor driver) (6/97)
 INFO [2020-03-12 00:01:23,269] ({Executor task launch worker-1}
Logging.scala[logInfo]:54) - Running task 13.0 in stage 8.0 (TID 21)
 INFO [2020-03-12 00:01:23,278] ({Executor task launch worker-4}
Logging.scala[logInfo]:54) - Finished task 4.0 in stage 8.0 (TID 12). 2506
bytes result sent to driver
 INFO [2020-03-12 

Re: use more than one host with spark cassandra connector

2019-09-09 Thread Sebastian Albrecht
Hi Jeff,
thank you. Is it possible to change the spark property in a %spark
paragraph which was set interpreter UI or has the interpreter to be
restarted for that?

BR,
Sebastian.

Am Sa., 7. Sept. 2019 um 04:36 Uhr schrieb Jeff Zhang :

> Sebastian,
>
> I don't think spark interpreter of zeppelin does anything for cassandra.
> IIUC spark.cassandra.connection.host is spark-cassandra connector
> property, as long as spark support it, it should work in zeppelin. So I
> would suggest to you to figure out how to set more than one host in spark
> cassandra connector first.
>
> Sebastian Albrecht  于2019年9月7日周六
> 上午12:25写道:
>
>>
>> Am Fr., 6. Sept. 2019 um 17:04 Uhr schrieb Jeff Zhang :
>>
>>> It is better to ask this kind of question in spark community. As long as
>>> spark support it, you can do it in zeppelin.
>>>
>>>
>> Hi Jeff,
>> thank you for your answer. Zeppelin already does is in the spark
>> interpreter: set the property spark.cassandra.connection.host , add the
>> cassandra-connector to the dependencies and go. But where do i configure an
>> additional host? In the same interpreter. Do i have to add a separate spark
>> interpreter?
>>
>> Thank you,
>> Sebastian.
>>
>>
>>> Sebastian Albrecht  于2019年9月6日周五
>>> 下午9:27写道:
>>>
>>>> Hi,
>>>> is there a way to connect a zeppelin instance to more than one
>>>> cassandra hosts via the spark cassandra connector to use the data from them
>>>> in spark?
>>>>
>>>> BR,
>>>> Sebastian.
>>>>
>>>
>>>
>>> --
>>> Best Regards
>>>
>>> Jeff Zhang
>>>
>>
>
> --
> Best Regards
>
> Jeff Zhang
>


Re: use more than one host with spark cassandra connector

2019-09-06 Thread Sebastian Albrecht
Am Fr., 6. Sept. 2019 um 17:04 Uhr schrieb Jeff Zhang :

> It is better to ask this kind of question in spark community. As long as
> spark support it, you can do it in zeppelin.
>
>
Hi Jeff,
thank you for your answer. Zeppelin already does is in the spark
interpreter: set the property spark.cassandra.connection.host , add the
cassandra-connector to the dependencies and go. But where do i configure an
additional host? In the same interpreter. Do i have to add a separate spark
interpreter?

Thank you,
Sebastian.


> Sebastian Albrecht  于2019年9月6日周五
> 下午9:27写道:
>
>> Hi,
>> is there a way to connect a zeppelin instance to more than one cassandra
>> hosts via the spark cassandra connector to use the data from them in spark?
>>
>> BR,
>> Sebastian.
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>


use more than one host with spark cassandra connector

2019-09-06 Thread Sebastian Albrecht
Hi,
is there a way to connect a zeppelin instance to more than one cassandra
hosts via the spark cassandra connector to use the data from them in spark?

BR,
Sebastian.