Github user steveloughran commented on a diff in the pull request:

    https://github.com/apache/spark/pull/19269#discussion_r144822829
  
    --- Diff: 
sql/core/src/test/scala/org/apache/spark/sql/sources/v2/SimpleWritableDataSource.scala
 ---
    @@ -0,0 +1,254 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *    http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +
    +package org.apache.spark.sql.sources.v2
    +
    +import java.io.{BufferedReader, InputStreamReader}
    +import java.text.SimpleDateFormat
    +import java.util.{Collections, Date, List => JList, Locale, Optional, UUID}
    +
    +import scala.collection.JavaConverters._
    +
    +import org.apache.hadoop.conf.Configuration
    +import org.apache.hadoop.fs.{FileSystem, Path}
    +
    +import org.apache.spark.SparkContext
    +import org.apache.spark.sql.{Row, SaveMode}
    +import org.apache.spark.sql.catalyst.InternalRow
    +import org.apache.spark.sql.sources.v2.reader.{DataReader, 
DataSourceV2Reader, ReadTask}
    +import org.apache.spark.sql.sources.v2.writer._
    +import org.apache.spark.sql.types.{DataType, StructType}
    +import org.apache.spark.util.SerializableConfiguration
    +
    +/**
    + * A HDFS based transactional writable data source.
    + * Each task writes data to 
`target/_temporary/jobId/$jobId-$partitionId-$attemptNumber`.
    + * Each job moves files from `target/_temporary/jobId/` to `target`.
    + */
    +class SimpleWritableDataSource extends DataSourceV2 with ReadSupport with 
WriteSupport {
    +
    +  private val schema = new StructType().add("i", "long").add("j", "long")
    +
    +  class Reader(path: String, conf: Configuration) extends 
DataSourceV2Reader {
    +    override def readSchema(): StructType = schema
    +
    +    override def createReadTasks(): JList[ReadTask[Row]] = {
    +      val dataPath = new Path(path)
    +      val fs = dataPath.getFileSystem(conf)
    +      if (fs.exists(dataPath)) {
    +        
fs.listStatus(dataPath).filter(!_.getPath.getName.startsWith("_")).map { f =>
    +          val serializableConf = new SerializableConfiguration(conf)
    +          new SimpleCSVReadTask(f.getPath.toUri.toString, 
serializableConf): ReadTask[Row]
    +        }.toList.asJava
    +      } else {
    +        Collections.emptyList()
    +      }
    +    }
    +  }
    +
    +  class Writer(path: String, conf: Configuration) extends 
DataSourceV2Writer {
    +    // We can't get the real spark job id here, so we use a timestamp and 
random UUID to simulate
    +    // a unique job id.
    +    private val jobId = new SimpleDateFormat("yyyyMMddHHmmss", 
Locale.US).format(new Date()) +
    +      "-" + UUID.randomUUID()
    +
    +    override def createWriterFactory(): DataWriterFactory[Row] = {
    +      new SimpleCSVDataWriterFactory(path, jobId, new 
SerializableConfiguration(conf))
    +    }
    +
    +    override def commit(messages: Array[WriterCommitMessage]): Unit = {
    +      val finalPath = new Path(path)
    +      val jobPath = new Path(new Path(finalPath, "_temporary"), jobId)
    +      val fs = jobPath.getFileSystem(conf)
    +      try {
    +        for (file <- fs.listStatus(jobPath).map(_.getPath)) {
    +          fs.rename(file, new Path(finalPath, file.getName))
    +        }
    +      } finally {
    +        fs.delete(jobPath, true)
    +      }
    +    }
    +
    +    override def abort(messages: Array[WriterCommitMessage]): Unit = {
    +      val jobPath = new Path(new Path(path, "_temporary"), jobId)
    +      val fs = jobPath.getFileSystem(conf)
    +      fs.delete(jobPath, true)
    +    }
    +  }
    +
    +  class InternalRowWriter(path: String, conf: Configuration)
    +    extends DataSourceV2Writer with SupportsWriteInternalRow {
    +
    +    private val jobId = new SimpleDateFormat("yyyyMMddHHmmss", 
Locale.US).format(new Date())
    +
    +    override def createInternalRowWriterFactory(): 
DataWriterFactory[InternalRow] = {
    +      new InternalRowCSVDataWriterFactory(path, jobId, new 
SerializableConfiguration(conf))
    +    }
    +
    +    override def commit(messages: Array[WriterCommitMessage]): Unit = {
    +      val finalPath = new Path(path)
    +      val jobPath = new Path(new Path(finalPath, "_temporary"), jobId)
    +      val fs = jobPath.getFileSystem(conf)
    +      try {
    +        for (file <- fs.listStatus(jobPath).map(_.getPath)) {
    +          fs.rename(file, new Path(finalPath, file.getName))
    --- End diff --
    
    again, see commentary on rename()


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to