strange behavior of spark 2.1.0

2017-04-01 Thread Jiang Jacky
Hello, Guys
I am running the spark streaming in 2.1.0, the scala version is tried on
2.11.7 and 2.11.4. And it is consuming from JMS. Recently, I have get the
following error
*"ERROR scheduler.ReceiverTracker: Deregistered receiver for stream 0:
Stopped by driver"*

*This error can be occurred randomly, it might be couple hours or couple
days. besides this error, everything is perfect.*
When the error happens, my job is stopped completely. There is no any other
error can be found.
I am running on top of yarn, and tried to look up the error through yarn
logs, container, no any further information appears there. The job is just
stopped from driver gracefully. BTW I have customized receiver, I either do
not think it is happened from receiver, there is no any error exception
from receiver, and I can also track the stop command is sent from "onStop"
function in receiver.

FYI, the driver is not consuming any large memory, there is no any RDD
"collect" command in the driver. I have also checked container log for each
executor, and cannot find any further error.




The following is my conf for the spark context
val conf = new SparkConf().setAppName(jobName).setMaster(master)
  .set("spark.hadoop.validateOutputSpecs", "false")
  .set("spark.driver.allowMultipleContexts", "true")
  .set("spark.streaming.receiver.maxRate", "500")
  .set("spark.streaming.backpressure.enabled", "true")
  .set("spark.streaming.stopGracefullyOnShutdown", "true")
  .set("spark.eventLog.enabled", "true");

If you have any idea or suggestion, please let me know. Appreciate on the
solution.

Thank you so much


Re: strange behavior of spark 2.1.0

2017-04-02 Thread Jiang Jacky
Thank you for replying. 
Actually there is no any message coming during the exception. And there is no 
OOME in any executor. What I am suspecting it might be caused by AWL. 

> On Apr 2, 2017, at 5:22 AM, Timur Shenkao  wrote:
> 
> Hello,
> It's difficult to tell without details.
> I believe one of the executors dies because of OOM or some Runtime Exception 
> (some unforeseen dirty data row).
> Less probable is GC stop-the-world pause when incoming message rate increases 
> drastically.
> 
> 
>> On Saturday, April 1, 2017, Jiang Jacky  wrote:
>> Hello, Guys
>> I am running the spark streaming in 2.1.0, the scala version is tried on 
>> 2.11.7 and 2.11.4. And it is consuming from JMS. Recently, I have get the 
>> following error
>> "ERROR scheduler.ReceiverTracker: Deregistered receiver for stream 0: 
>> Stopped by driver"
>> 
>> This error can be occurred randomly, it might be couple hours or couple 
>> days. besides this error, everything is perfect.
>> When the error happens, my job is stopped completely. There is no any other 
>> error can be found.
>> I am running on top of yarn, and tried to look up the error through yarn 
>> logs, container, no any further information appears there. The job is just 
>> stopped from driver gracefully. BTW I have customized receiver, I either do 
>> not think it is happened from receiver, there is no any error exception from 
>> receiver, and I can also track the stop command is sent from "onStop" 
>> function in receiver.
>> 
>> FYI, the driver is not consuming any large memory, there is no any RDD 
>> "collect" command in the driver. I have also checked container log for each 
>> executor, and cannot find any further error.
>> 
>> 
>> 
>> 
>> The following is my conf for the spark context
>> val conf = new SparkConf().setAppName(jobName).setMaster(master)
>>   .set("spark.hadoop.validateOutputSpecs", "false")
>>   .set("spark.driver.allowMultipleContexts", "true")
>>   .set("spark.streaming.receiver.maxRate", "500")
>>   .set("spark.streaming.backpressure.enabled", "true")
>>   .set("spark.streaming.stopGracefullyOnShutdown", "true")
>>   .set("spark.eventLog.enabled", "true");
>> 
>> If you have any idea or suggestion, please let me know. Appreciate on the 
>> solution.
>> 
>> Thank you so much
>> 


Re: convert JavaRDD> to JavaRDD

2017-04-05 Thread Jiang Jacky
There is no flattop just flatMap 

> On Apr 5, 2017, at 12:24 PM, Vinay Parekar  wrote:
> 
> I think flattop() will be helpful in this case . Correct me if I am wrong.
> 
> From: Hamza HACHANI 
> Date: Wednesday, April 5, 2017 at 3:43 AM
> To: "user@spark.apache.org" 
> Subject: convert JavaRDD> to JavaRDD
> 
> I want to convert a JavaRDD> to JavaRDD. For example if 
> there is 3 elment in List 3 Object would be created in my new 
> JavaRDD.
> 
> Does any one have an idea ?
> 
> 
> 
> CONFIDENTIALITY NOTICE: This message is intended only for the use and review 
> of the individual or entity to which it is addressed and may contain 
> information that is privileged and confidential. If the reader of this 
> message is not the intended recipient, or the employee or agent responsible 
> for delivering the message solely to the intended recipient, you are hereby 
> notified that any dissemination, distribution or copying of this 
> communication is strictly prohibited. If you have received this communication 
> in error, please notify sender immediately by telephone or return email. 
> Thank you.