Hello list,

I'm having a little performance issue, with different Spark versions.

I've a spark embedded application written in scala, Initially I've use
Spark 2.0.2, and works fine, with good speed response, but when I updated
to 2.3.2 , with no any code changes It becomes slower.

Mainly what the application do is to gather information from a source,
apply transformations with filters and performs aggregation over it. It's
source is mainly parquet and no write is done just a serialization from the
result.

Maybe I'm using  deprecated api functions or the order of the operations
are not generating a good plan...

Can someone give me some idea of any change on the versions that could
generate this behavior?

Regards,

-- 
Ing. Ivaldi Andres

Reply via email to