Hi all,
I need some help regarding Arrow RecordBatches/Pandas Dataframes to (Arrow enabled) Spark Dataframe conversions. Here the example explains very well how to convert a single Pandas Dataframe to Spark Dataframe [1]. But in my case, some external applications are generating Arrow RecordBatches in my PySpark application in streaming fashion. Each time I receive an Arrow RB, I want to transfer/append it to a Spark Dataframe. So is it possible to create a Spark Dataframe initially from one Arrow RecordBatch and then start appending many other in-coming Arrow RecordBatches to that Spark Dataframe (like in streaming fashion)? Thanks! I saw another example [2] in which all the Arrow RB are being converted to Spark Dataframe but my case is little bit different than this. [1] https://spark.apache.org/docs/latest/sql-pyspark-pandas-with-arrow.html [2] https://gist.github.com/linar-jether/7dd61ed6fa89098ab9c58a1ab428b2b5 --- Regards, Tanveer Ahmad