[ 
https://issues.apache.org/jira/browse/SPARK-16408?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15365668#comment-15365668
 ] 

zenglinxi commented on SPARK-16408:
-----------------------------------

I think we should add an parameter (spark.input.dir.recursive) to control the 
value of recursive, and make this parameter works by modify some code, like:
{noformat}
diff --git 
a/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/commands.scala 
b/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/commands.scala
index 6b16d59..3be8553 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/commands.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/commands.scala
@@ -113,8 +113,9 @@ case class AddFile(path: String) extends RunnableCommand {
 
   override def run(sqlContext: SQLContext): Seq[Row] = {
     val hiveContext = sqlContext.asInstanceOf[HiveContext]
+    val recursive = 
sqlContext.sparkContext.getConf.getBoolean("spark.input.dir.recursive", false)
     hiveContext.runSqlHive(s"ADD FILE $path")
-    hiveContext.sparkContext.addFile(path)
+    hiveContext.sparkContext.addFile(path, recursive)
     Seq.empty[Row]
   }
 }
{noformat}


> SparkSQL Added file get Exception: is a directory and recursive is not turned 
> on
> --------------------------------------------------------------------------------
>
>                 Key: SPARK-16408
>                 URL: https://issues.apache.org/jira/browse/SPARK-16408
>             Project: Spark
>          Issue Type: Task
>          Components: SQL
>    Affects Versions: 1.6.2
>            Reporter: zenglinxi
>
> when use Spark-sql to execute sql like:
> {quote}
> add file hdfs://xxx/user/test;
> {quote}
> if the HDFS path( hdfs://xxx/user/test) is a directory, then we will get an 
> exception like:
> {quote}
> org.apache.spark.SparkException: Added file hdfs://xxx/user/test is a 
> directory and recursive is not turned on.
>        at org.apache.spark.SparkContext.addFile(SparkContext.scala:1372)
>        at org.apache.spark.SparkContext.addFile(SparkContext.scala:1340)
>        at org.apache.spark.sql.hive.execution.AddFile.run(commands.scala:117)
>        at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
>        at 
> org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
>        at 
> org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to