it seems that the job block whene we call newAPIHadoopRDD to get data from
Hbase. it may be the issue !!
is there another api to load date from hbase ?


<https://mailtrack.io/> Sent with Mailtrack
<https://chrome.google.com/webstore/detail/mailtrack-for-gmail-inbox/ndnaehgpjlnokgebbaldlmgkapkpjkkb?utm_source=gmail&utm_medium=signature&utm_campaign=signaturevirality>
<#>

2017-10-11 14:45 GMT+02:00 Sebastian Piu <sebastian....@gmail.com>:

> We do have this issue randomly too, so interested in hearing if someone
> was able to get to the bottom of it
>
> On Wed, 11 Oct 2017, 13:40 amine_901, <cherifimohamedam...@gmail.com>
> wrote:
>
>> We encounter a problem on a Spark job 1.6(on yarn) that never ends, whene
>> several jobs launched simultaneously.
>> We found that by launching the job spark in yarn-client mode we do not
>> have
>> this problem, unlike launching it in yarn-cluster mode.
>> it could be a trail to find the cause.
>>
>> we changed the code to add a sparkContext.stop ()
>> Indeed, the SparkContext was created (val sparkContext =
>> createSparkContext)
>> but not stopped. this solution has allowed us to decrease the number of
>> jobs
>> that remains blocked but nevertheless we still have some jobs blocked.
>>
>> by analyzing the logs we have found this log that repeats without
>> stopping:
>> /17/09/29 11:04:37 DEBUG SparkEventPublisher: Enqueue
>> SparkListenerExecutorMetricsUpdate(1,WrappedArray())
>> 17/09/29 11:04:41 DEBUG ApplicationMaster: Sending progress
>> 17/09/29 11:04:41 DEBUG ApplicationMaster: Number of pending allocations
>> is
>> 0. Sleeping for 5000. /
>>
>> Does someone have an idea about this issue ?
>> Thank you in advance
>>
>>
>>
>> --
>> Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>>
>>


-- 
CHERIFI Mohamed Amine
Développeur Big data/Data scientist
07 81 65 17 03

Reply via email to