Re: Migration to Spark 3.2

2022-01-31 Thread Aurélien Mazoyer
Hi Stephen, I managed to solve my issue, I had a conflicting version of jackson databind that came from parent pom. Thank you, Aurelien Le dim. 30 janv. 2022 à 23:28, Aurélien Mazoyer a écrit : > Hi Stephen, > > Thank you for your answer. Yes, I changed the scope to "pro

Re: Migration to Spark 3.2

2022-01-30 Thread Aurélien Mazoyer
look fine. > > What happens if you change the scope of your Jackson dependencies to > “provided”? > > This should result in your application using the versions provided by > Spark and avoid this potential collision. > > Cheers, > > Steve C > > On 27 Jan 2022, at 9:48 p

Re: Migration to Spark 3.2

2022-01-27 Thread Aurélien Mazoyer
ependencies. > > Feel free to respond with the output if you have any questions about it. > > Cheers, > > Steve C > > > On 22 Jan 2022, at 10:49 am, Aurélien Mazoyer > wrote: > > > > Hello, > > > > I migrated my code to Spark 3.2 and I am

Re: Migration to Spark 3.2

2022-01-25 Thread Aurélien Mazoyer
Hello, Sorry for asking twice, but anyone has any idea which issue I could be facing with this depencency problem :-/? Thank you, Aurelien Le sam. 22 janv. 2022 à 00:49, Aurélien Mazoyer a écrit : > Hello, > > I migrated my code to Spark 3.2 and I am facing some issues. When I run

Migration to Spark 3.2

2022-01-21 Thread Aurélien Mazoyer
Hello, I migrated my code to Spark 3.2 and I am facing some issues. When I run my unit tests via Maven, I get this error: java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.rdd.RDDOperationScope$ which is not super nice. However, when I run my test via Intellij, I get

Re: Get application metric from Spark job

2021-09-09 Thread Aurélien Mazoyer
spark-metrics.html > > > http://mail-archives.us.apache.org/mod_mbox/spark-user/201501.mbox/%3CCAE50=dq+6tdx9VNVM3ctBMWPLDPbUAacO3aN3L8x38zg=xb...@mail.gmail.com%3E > > > > I hope these help. > > -- > > Thanks & Regards, > > Akshay Haryani > > > &g

Detailed information about corrupted record

2021-09-09 Thread Aurélien Mazoyer
Hello, I use a DataFrameReader and I use permissive mode for corrupted records. I would like to have more information about corrupted records: for example, in the case of a schema mismatch, something like: "Invalid type. Expected Integer but got String for column number 12 (in case of csv). By

Re: Get application metric from Spark job

2021-09-06 Thread Aurélien Mazoyer
t; If you want to build your own custom metrics, you can explore spark custom > plugins. Using a custom plugin, you can track your own custom metrics and > plug it into the spark metrics system. Please note plugins are supported > on spark versions above 3.0. > > > > > > --

Get application metric from Spark job

2021-09-02 Thread Aurélien Mazoyer
Hi community, I would like to collect information about the execution of a Spark job while it is running. Could I define some kind of application metrics (such as a counter that would be incremented in my code) that I could retrieve regularly while the job is running? Thank you for help,