Hadoop does not run Spark programs, Spark does. How or why would something,
what, modify the byte code? No

On Mon, Jan 24, 2022, 9:07 AM sam smith <qustacksm2123...@gmail.com> wrote:

> My point is could Hadoop go wrong about one Spark execution ? meaning that
> it gets confused (given the concurrent distributed tasks) and then adds
> wrong instruction to the program, or maybe does execute an instruction not
> at its right order (shuffling the order of execution by executing previous
> ones, while it shouldn't) ? Before finishing and returning the results from
> one node it returns the results of the other in a wrong way for example.
>
> Le lun. 24 janv. 2022 à 15:31, Sean Owen <sro...@gmail.com> a écrit :
>
>> Not clear what you mean here. A Spark program is a program, so what are
>> the alternatives here? program execution order is still program execution
>> order. You are not guaranteed anything about order of concurrent tasks.
>> Failed tasks can be reexecuted so should be idempotent. I think the answer
>> is 'no' but not sure what you are thinking of here.
>>
>> On Mon, Jan 24, 2022 at 7:10 AM sam smith <qustacksm2123...@gmail.com>
>> wrote:
>>
>>> Hello guys,
>>>
>>> I hope my question does not sound weird, but could a Spark execution on
>>> Hadoop cluster give different output than the program actually does ? I
>>> mean by that, the execution order is messed by hadoop, or an instruction
>>> executed twice..; ?
>>>
>>> Thanks for your enlightenment
>>>
>>

Reply via email to