[
https://issues.apache.org/jira/browse/SPARK-20263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15973157#comment-15973157
]
Ott Toomet commented on SPARK-20263:
------------------------------------
for instance, by rbinding in a loop. I have several pieces of code that work
along the lines:
for(file in datafiles) {
df <- read.parquet(file)
result <- do_something_with(df)
final <- rbind(final, result)
}
Sure, one can easily add logic to check if we have initialized 'final' but it
seems so much simpler and logical just to initialize empty 'final' before the
loop.
> create empty dataframes in sparkR
> ---------------------------------
>
> Key: SPARK-20263
> URL: https://issues.apache.org/jira/browse/SPARK-20263
> Project: Spark
> Issue Type: Wish
> Components: SparkR
> Affects Versions: 2.1.0
> Reporter: Ott Toomet
> Priority: Minor
>
> SparkR 2.1 does not support creating empty dataframes, nor conversion of
> empty R dataframes to spark ones:
> createDataFrame(data.frame(a=integer()))
> gives
> Error in takeRDD(x, 1)[[1]] : subscript out of bounds
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]