Hey, from what I know you can try to Union them df.union(df2)

Not sure if this is what you need 

> On 25. May 2020, at 13:53, Tanveer Ahmad - EWI <t.ah...@tudelft.nl> wrote:
> 
> Hi all,
> 
> I need some help regarding Arrow RecordBatches/Pandas Dataframes to (Arrow 
> enabled) Spark Dataframe conversions.
> Here the example explains very well how to convert a single Pandas Dataframe 
> to Spark Dataframe [1]. 
> 
> But in my case, some external applications are generating Arrow RecordBatches 
> in my PySpark application in streaming fashion. Each time I receive an Arrow 
> RB, I want to transfer/append it to a Spark Dataframe. So is it possible to 
> create a Spark Dataframe initially from one Arrow RecordBatch and then start 
> appending many other in-coming Arrow RecordBatches to that Spark Dataframe 
> (like in streaming fashion)? Thanks!
> 
> I saw another example [2] in which all the Arrow RB are being converted to 
> Spark Dataframe but my case is little bit different than this.  
> 
> [1] https://spark.apache.org/docs/latest/sql-pyspark-pandas-with-arrow.html 
>  <https://spark.apache.org/docs/latest/sql-pyspark-pandas-with-arrow.html>
> [2] https://gist.github.com/linar-jether/7dd61ed6fa89098ab9c58a1ab428b2b5 
> <https://gist.github.com/linar-jether/7dd61ed6fa89098ab9c58a1ab428b2b5>
> 
> ---
> Regards,
> Tanveer Ahmad

Reply via email to