Hey Kevin, It is a empty directory, It is able to write part files to the directory but while merging those part files we are getting above error.
Regards On Tue, Sep 20, 2016 at 7:46 PM, Kevin Mellott <kevin.r.mell...@gmail.com> wrote: > Have you checked to see if any files already exist at > /nfspartition/sankar/banking_l1_v2.csv? If so, you will need to delete > them before attempting to save your DataFrame to that location. > Alternatively, you may be able to specify the "mode" setting of the > df.write operation to "overwrite", depending on the version of Spark you > are running. > > *ERROR (from log)* > 16/09/17 08:03:28 WARN FileUtil: Failed to delete file or > dir[/nfspartition/sankar/banking_l1_v2.csv/_temporary/0/ > task_201609170802_0013_m_000000/.part-r-00000-46a7f178-2490- > 444e-9110-510978eaaecb.csv.crc]: > it still exists. > 16/09/17 08:03:28 WARN FileUtil: Failed to delete file or > dir[/nfspartition/sankar/banking_l1_v2.csv/_temporary/0/ > task_201609170802_0013_m_000000/part-r-00000-46a7f178-2490- > 444e-9110-510978eaaecb.csv]: > it still exists. > > *df.write Documentation* > http://spark.apache.org/docs/latest/api/R/write.df.html > > Thanks, > Kevin > > On Tue, Sep 20, 2016 at 12:16 AM, sankarmittapally <sankar.mittapally@ > creditvidya.com> wrote: > >> We have setup a spark cluster which is on NFS shared storage, there is no >> permission issues with NFS storage, all the users are able to write to NFS >> storage. When I fired write.df command in SparkR, I am getting below. Can >> some one please help me to fix this issue. >> >> >> 16/09/17 08:03:28 ERROR InsertIntoHadoopFsRelationCommand: Aborting job. >> java.io.IOException: Failed to rename DeprecatedRawLocalFileStatus >> {path=file:/nfspartition/sankar/banking_l1_v2.csv/_temporary >> /0/task_201609170802_0013_m_000000/part-r-00000-46a7f178- >> 2490-444e-9110-510978eaaecb.csv; >> isDirectory=false; length=436486316; replication=1; blocksize=33554432; >> modification_time=1474099400000; access_time=0; owner=; group=; >> permission=rw-rw-rw-; isSymlink=false} >> to >> file:/nfspartition/sankar/banking_l1_v2.csv/part-r-00000- >> 46a7f178-2490-444e-9110-510978eaaecb.csv >> at >> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.m >> ergePaths(FileOutputCommitter.java:371) >> at >> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.m >> ergePaths(FileOutputCommitter.java:384) >> at >> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter. >> commitJob(FileOutputCommitter.java:326) >> at >> org.apache.spark.sql.execution.datasources.BaseWriterContain >> er.commitJob(WriterContainer.scala:222) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoo >> pFsRelationCommand.scala:144) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRela >> tionCommand.scala:115) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRela >> tionCommand.scala:115) >> at >> org.apache.spark.sql.execution.SQLExecution$.withNewExecutio >> nId(SQLExecution.scala:57) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115) >> at >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> ideEffectResult$lzycompute(commands.scala:60) >> at >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> ideEffectResult(commands.scala:58) >> at >> org.apache.spark.sql.execution.command.ExecutedCommandExec. >> doExecute(commands.scala:74) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. >> apply(SparkPlan.scala:115) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1. >> apply(SparkPlan.scala:115) >> at >> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQue >> ry$1.apply(SparkPlan.scala:136) >> at >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperati >> onScope.scala:151) >> at >> org.apache.spark.sql.execution.SparkPlan.executeQuery( >> SparkPlan.scala:133) >> at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114) >> at >> org.apache.spark.sql.execution.QueryExecution.toRdd$ >> lzycompute(QueryExecution.scala:86) >> at >> org.apache.spark.sql.execution.QueryExecution.toRdd( >> QueryExecution.scala:86) >> at >> org.apache.spark.sql.execution.datasources.DataSource.write( >> DataSource.scala:487) >> at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211) >> at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBac >> kendHandler.scala:141) >> at >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackend >> Handler.scala:86) >> at >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackend >> Handler.scala:38) >> at >> io.netty.channel.SimpleChannelInboundHandler.channelRead(Sim >> pleChannelInboundHandler.java:105) >> at >> io.netty.channel.AbstractChannelHandlerContext.invokeChannel >> Read(AbstractChannelHandlerContext.java:308) >> at >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRe >> ad(AbstractChannelHandlerContext.java:294) >> at >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(M >> essageToMessageDecoder.java:103) >> at >> io.netty.channel.AbstractChannelHandlerContext.invokeChannel >> Read(AbstractChannelHandlerContext.java:308) >> at >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRe >> ad(AbstractChannelHandlerContext.java:294) >> at >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(Byte >> ToMessageDecoder.java:244) >> at >> io.netty.channel.AbstractChannelHandlerContext.invokeChannel >> Read(AbstractChannelHandlerContext.java:308) >> at >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRe >> ad(AbstractChannelHandlerContext.java:294) >> at >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(Defa >> ultChannelPipeline.java:846) >> at >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe. >> read(AbstractNioByteChannel.java:131) >> at >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEven >> tLoop.java:511) >> at >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimiz >> ed(NioEventLoop.java:468) >> at >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEve >> ntLoop.java:382) >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) >> at >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run( >> SingleThreadEventExecutor.java:111) >> at >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnabl >> eDecorator.run(DefaultThreadFactory.java:137) >> at java.lang.Thread.run(Thread.java:745) >> 16/09/17 08:03:28 WARN FileUtil: Failed to delete file or >> dir[/nfspartition/sankar/banking_l1_v2.csv/_temporary/0/ >> task_201609170802_0013_m_000000/.part-r-00000-46a7f178-2490- >> 444e-9110-510978eaaecb.csv.crc]: >> it still exists. >> 16/09/17 08:03:28 WARN FileUtil: Failed to delete file or >> dir[/nfspartition/sankar/banking_l1_v2.csv/_temporary/0/ >> task_201609170802_0013_m_000000/part-r-00000-46a7f178-2490- >> 444e-9110-510978eaaecb.csv]: >> it still exists. >> 16/09/17 08:03:28 ERROR DefaultWriterContainer: Job job_201609170803_0000 >> aborted. >> 16/09/17 08:03:28 ERROR RBackendHandler: save on 625 failed >> Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : >> org.apache.spark.SparkException: Job aborted. >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoo >> pFsRelationCommand.scala:149) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRela >> tionCommand.scala:115) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand$$anonfun$run$1.apply(InsertIntoHadoopFsRela >> tionCommand.scala:115) >> at >> org.apache.spark.sql.execution.SQLExecution$.withNewExecutio >> nId(SQLExecution.scala:57) >> at >> org.apache.spark.sql.execution.datasources.InsertIntoHadoopF >> sRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:115) >> at >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> ideEffectResult$lzycompute(commands.scala:60) >> at >> org.apache.spark.sql.execution.command.ExecutedCommandExec.s >> ideEffectResult(commands.scala:58) >> at org.apache.spark.sql.execution.command.ExecutedCommandExec.doE >> >> >> >> >> -- >> View this message in context: http://apache-spark-user-list. >> 1001560.n3.nabble.com/write-df-is-failing-on-Spark-Cluster-tp27761.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >> > -- Regards Sankar Mittapally Senior Software Engineer