Is it correct to say that Spark Dataframe APIs are implemented using same execution as SparkSQL ? In other words, while the dataframe API is different than SparkSQL, the runtime performance of equivalent constructs in Dataframe and SparkSQL should be same. So one should be able to choose whichever of the two (DF v/s SQL) suite the use cases and not worry about runtime performance.
Pl comment ... Thanks,