The reads are from a non-IGFS source, but the writes are through IGFS. Spark
uses Hadoop's FileOutputCommitter to write the output to IGFS. I think what
happens is essentially:

- During processing, temporary files are written by each of n executors
running on different nodes to some /data/path/output/_temporary/part-n...
- When the job completes, each of the executor performs the final "commit"
by renaming the files under /data/path/output/_temporary/part-n... to
/data/path/output/part-n... and deletes the _temporary directory.



--
View this message in context: 
http://apache-ignite-users.70518.x6.nabble.com/igfs-meta-behavior-when-node-restarts-tp13155p13322.html
Sent from the Apache Ignite Users mailing list archive at Nabble.com.

Reply via email to