Github user sancyx commented on the issue: https://github.com/apache/zeppelin/pull/2637 Hi @naveenkumargp, there was a refactor around the interpreter packaging which caused the ClassNotFound problems. Previously there was a big jar containing the interpreter class as well, which doesn't exists anymore. We've updated the PR, so that all jar files from local repo are enumerated with --jar option to spark-submit, which probably is a better approach. With regards to deploy mode: just tried to add `spark.submit.deployMode` property via UI and it worked for me, we do set this by default to `cluster` in our custom interpreter.json, as well as the master property to `k8s://https://kubernetes:443` however intentionally didn't included in the patch, since this is only optional. We are using this patch about a half year now, in our deployments and works fine for us, would be glad to contribute to community. Please try this latest version if works for you.
---