Re: Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
Thank you Jeff! I would certainly give it a try. Best, Rahul On 2021/09/26 22:49:03, Jeff Zhang wrote: > Hi kumar, > > You can try Zeppelin which support the udf sharing across languages > > http://zeppelin.apache.org/ > > > > > rahul kumar 于2021年9月27日周一 上午4:20写道: > > > I'm trying to

Re: Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
Thanks Sean, - I have tried executing it without wrapping into DataFrame constructor, but got the same error " AttributeError: 'StructField' object has no attribute '_get_object_id' - I have also tried using udf scala udf: class PythonUtil() extends UDF5[DataFrame, String,String,

Re: Unable to use scala function in pyspark

2021-09-26 Thread Sean Owen
You can also call a Scala UDF from Python in Spark - this doesn't need Zeppelin or relate to the front-end. This may indeed be much easier as a proper UDF; depends on what this function does. However I think the issue may be that you're trying to wrap the resulting DataFrame in a DataFrame or

Re: Unable to use scala function in pyspark

2021-09-26 Thread Jeff Zhang
You can first try it via docker http://zeppelin.apache.org/download.html#using-the-official-docker-image Jeff Zhang 于2021年9月27日周一 上午6:49写道: > Hi kumar, > > You can try Zeppelin which support the udf sharing across languages > > http://zeppelin.apache.org/ > > > > > rahul kumar 于2021年9月27日周一

Re: Unable to use scala function in pyspark

2021-09-26 Thread Jeff Zhang
Hi kumar, You can try Zeppelin which support the udf sharing across languages http://zeppelin.apache.org/ rahul kumar 于2021年9月27日周一 上午4:20写道: > I'm trying to use a function defined in scala jar in pyspark ( spark > 3.0.2). > > --scala --- > > Object PythonUtil { > > def

Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
I'm trying to use a function defined in scala jar in pyspark ( spark 3.0.2). --scala --- Object PythonUtil { def customedf(dataFrame: DataFrame, keyCol: String, table: String, outputSchema: StructType, database: String):

[ANNOUNCE] Release Apache Kyuubi(Incubating) 1.3.0-incubating

2021-09-26 Thread Cheng Pan
Hello Spark Community, The Apache Kyuubi(Incubating) community is pleased to announce that Apache Kyuubi(Incubating) 1.3.0-incubating has been released! Apache Kyuubi(Incubating) is a distributed multi-tenant JDBC server for large-scale data processing and analytics, built on top of Apache Spark