also this one in warning log:

E0702 11:35:08.869998 17840 slave.cpp:2310] Container
'af557235-2d5f-4062-aaf3-a747cb3cd0d1' for executor
'20140616-104524-1694607552-5050-26919-1' of framework
'20140702-113428-1694607552-5050-17766-0000' failed to start: Failed to
fetch URIs for container 'af557235-2d5f-4062-aaf3-a747cb3cd0d1': exit
status 32512


2014-07-02 11:46 GMT+08:00 qingyang li <liqingyang1...@gmail.com>:

> Here is the log:
>
> E0702 10:32:07.599364 14915 slave.cpp:2686] Failed to unmonitor container
> for executor 20140616-104524-1694607552-5050-26919-1 of framework
> 20140702-102939-1694607552-5050-14846-0000: Not monitored
>
>
> 2014-07-02 1:45 GMT+08:00 Aaron Davidson <ilike...@gmail.com>:
>
> Can you post the logs from any of the dying executors?
>>
>>
>> On Tue, Jul 1, 2014 at 1:25 AM, qingyang li <liqingyang1...@gmail.com>
>> wrote:
>>
>> > i am using mesos0.19 and spark0.9.0 ,  the mesos cluster is started,
>> when I
>> > using spark-shell to submit one job, the tasks always lost.  here is the
>> > log:
>> > ----------
>> > 14/07/01 16:24:27 INFO DAGScheduler: Host gained which was in lost list
>> > earlier: bigdata005
>> > 14/07/01 16:24:27 INFO TaskSetManager: Starting task 0.0:1 as TID 4042
>> on
>> > executor 20140616-143932-1694607552-5050-4080-2: bigdata005
>> (PROCESS_LOCAL)
>> > 14/07/01 16:24:27 INFO TaskSetManager: Serialized task 0.0:1 as 1570
>> bytes
>> > in 0 ms
>> > 14/07/01 16:24:28 INFO TaskSetManager: Re-queueing tasks for
>> > 20140616-104524-1694607552-5050-26919-1 from TaskSet 0.0
>> > 14/07/01 16:24:28 WARN TaskSetManager: Lost TID 4041 (task 0.0:0)
>> > 14/07/01 16:24:28 INFO DAGScheduler: Executor lost:
>> > 20140616-104524-1694607552-5050-26919-1 (epoch 3427)
>> > 14/07/01 16:24:28 INFO BlockManagerMasterActor: Trying to remove
>> executor
>> > 20140616-104524-1694607552-5050-26919-1 from BlockManagerMaster.
>> > 14/07/01 16:24:28 INFO BlockManagerMaster: Removed
>> > 20140616-104524-1694607552-5050-26919-1 successfully in removeExecutor
>> > 14/07/01 16:24:28 INFO TaskSetManager: Re-queueing tasks for
>> > 20140616-143932-1694607552-5050-4080-2 from TaskSet 0.0
>> > 14/07/01 16:24:28 WARN TaskSetManager: Lost TID 4042 (task 0.0:1)
>> > 14/07/01 16:24:28 INFO DAGScheduler: Executor lost:
>> > 20140616-143932-1694607552-5050-4080-2 (epoch 3428)
>> > 14/07/01 16:24:28 INFO BlockManagerMasterActor: Trying to remove
>> executor
>> > 20140616-143932-1694607552-5050-4080-2 from BlockManagerMaster.
>> > 14/07/01 16:24:28 INFO BlockManagerMaster: Removed
>> > 20140616-143932-1694607552-5050-4080-2 successfully in removeExecutor
>> > 14/07/01 16:24:28 INFO DAGScheduler: Host gained which was in lost list
>> > earlier: bigdata005
>> > 14/07/01 16:24:28 INFO DAGScheduler: Host gained which was in lost list
>> > earlier: bigdata001
>> > 14/07/01 16:24:28 INFO TaskSetManager: Starting task 0.0:1 as TID 4043
>> on
>> > executor 20140616-143932-1694607552-5050-4080-2: bigdata005
>> (PROCESS_LOCAL)
>> > 14/07/01 16:24:28 INFO TaskSetManager: Serialized task 0.0:1 as 1570
>> bytes
>> > in 0 ms
>> > 14/07/01 16:24:28 INFO TaskSetManager: Starting task 0.0:0 as TID 4044
>> on
>> > executor 20140616-104524-1694607552-5050-26919-1: bigdata001
>> > (PROCESS_LOCAL)
>> > 14/07/01 16:24:28 INFO TaskSetManager: Serialized task 0.0:0 as 1570
>> bytes
>> > in 0 ms
>> >
>> >
>> > it seems other guy has also encountered such problem,
>> >
>> >
>> http://mail-archives.apache.org/mod_mbox/incubator-mesos-dev/201305.mbox/%3c201305161047069952...@nfs.iscas.ac.cn%3E
>> >
>>
>
>

Reply via email to