[ https://issues.apache.org/jira/browse/SPARK-20263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15973157#comment-15973157 ]
Ott Toomet commented on SPARK-20263: ------------------------------------ for instance, by rbinding in a loop. I have several pieces of code that work along the lines: for(file in datafiles) { df <- read.parquet(file) result <- do_something_with(df) final <- rbind(final, result) } Sure, one can easily add logic to check if we have initialized 'final' but it seems so much simpler and logical just to initialize empty 'final' before the loop. > create empty dataframes in sparkR > --------------------------------- > > Key: SPARK-20263 > URL: https://issues.apache.org/jira/browse/SPARK-20263 > Project: Spark > Issue Type: Wish > Components: SparkR > Affects Versions: 2.1.0 > Reporter: Ott Toomet > Priority: Minor > > SparkR 2.1 does not support creating empty dataframes, nor conversion of > empty R dataframes to spark ones: > createDataFrame(data.frame(a=integer())) > gives > Error in takeRDD(x, 1)[[1]] : subscript out of bounds -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org