[ 
https://issues.apache.org/jira/browse/SPARK-20263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15973157#comment-15973157
 ] 

Ott Toomet commented on SPARK-20263:
------------------------------------

for instance, by rbinding in a loop.  I have several pieces of code that work 
along the lines:
for(file in datafiles) {
   df <- read.parquet(file)
   result <- do_something_with(df)
   final <- rbind(final, result) 
}

Sure, one can easily add logic to check if we have initialized 'final' but it 
seems so much simpler and logical just to initialize empty 'final' before the 
loop.

> create empty dataframes in sparkR
> ---------------------------------
>
>                 Key: SPARK-20263
>                 URL: https://issues.apache.org/jira/browse/SPARK-20263
>             Project: Spark
>          Issue Type: Wish
>          Components: SparkR
>    Affects Versions: 2.1.0
>            Reporter: Ott Toomet
>            Priority: Minor
>
> SparkR 2.1 does not support creating empty dataframes, nor conversion of 
> empty R dataframes to spark ones:
> createDataFrame(data.frame(a=integer()))
> gives 
> Error in takeRDD(x, 1)[[1]] : subscript out of bounds



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to