But if you don't set it in interpreter setting, it would get spark ui url
dynamically.



ankit jain <ankitjain....@gmail.com>于2018年1月25日周四 下午3:03写道:

> That method is just reading it from a config defined in interpreter
> settings called "uiWebUrl" which makes it configurable but still static.
>
> On Wed, Jan 24, 2018 at 10:58 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>>
>> IIRC, spark interpreter can get web ui url at runtime instead of static
>> url.
>>
>>
>> https://github.com/apache/zeppelin/blob/master/spark/src/main/java/org/apache/zeppelin/spark/SparkInterpreter.java#L940
>>
>>
>> ankit jain <ankitjain....@gmail.com>于2018年1月25日周四 下午2:55写道:
>>
>>> Issue with Spark UI when running on AWS EMR is it requires ssh tunneling
>>> to be setup which requires private aws keys.
>>>
>>> Our team is building a analytic platform on zeppelin for end-users who
>>> we obviously can't hand out these keys.
>>>
>>> Another issue is setting up correct port - Zeppelin tries to use 4040
>>> for spark but during an interpreter restart 4040 could be used by an old
>>> still stuck paragraph. In that case Zeppelin simply tries the next port and
>>> so on.
>>>
>>> Static url for Spark can't handle this and hence requires some dynamic
>>> implementation.
>>>
>>> PS - As I write this a lightbulb goes on in my head. I guess we could
>>> also modify Zeppelin restart script to kill those rogue processes and make
>>> sure 4040 is always available?
>>>
>>> Thanks
>>> Ankit
>>>
>>> On Wed, Jan 24, 2018 at 6:10 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>
>>>>
>>>> If Spark interpreter didn't give you the correct spark UI, this should
>>>> be a bug, you can file a ticket to fix it. Although you can make a custom
>>>> interpreter by extending the current spark interpreter, it is not a trivial
>>>> work.
>>>>
>>>>
>>>> ankit jain <ankitjain....@gmail.com>于2018年1月25日周四 上午8:07写道:
>>>>
>>>>> Hi fellow Zeppelin users,
>>>>> Has anyone tried to write a custom Spark Interpreter perhaps extending
>>>>> from the one that ships currently with zeppelin -
>>>>> spark/src/main/java/org/apache/zeppelin/spark/*SparkInterpreter.java?*
>>>>>
>>>>> We are coming across cases where we need the interpreter to do "more",
>>>>> eg change getSparkUIUrl() to directly load Yarn
>>>>> ResourceManager/proxy/application_id123 rather than a fixed web ui.
>>>>>
>>>>> If we directly modify Zeppelin source code, upgrading to new zeppelin
>>>>> versions will be a mess.
>>>>>
>>>>> Before we get too deep into it, wanted to get thoughts of the
>>>>> community.
>>>>>
>>>>> What is a "clean" way to do such changes?
>>>>>
>>>>> --
>>>>> Thanks & Regards,
>>>>> Ankit.
>>>>>
>>>>
>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Ankit.
>>>
>>
>
>
> --
> Thanks & Regards,
> Ankit.
>

Reply via email to