Re: Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
Thank you Jeff! I would certainly give it a try. Best, Rahul On 2021/09/26 22:49:03, Jeff Zhang wrote: > Hi kumar, > > You can try Zeppelin which support the udf sharing across languages > > http://zeppelin.apache.org/ > > > > > rahul kumar 于2021年9月27日

Re: Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
ch support the udf sharing across languages > > > > http://zeppelin.apache.org/ > > > > > > > > > > rahul kumar 于2021年9月27日周一 上午4:20写道: > > > >> I'm trying to use a function defined in scala jar in pyspark ( spark > >> 3.0.2). >

Unable to use scala function in pyspark

2021-09-26 Thread rahul kumar
I'm trying to use a function defined in scala jar in pyspark ( spark 3.0.2). --scala --- Object PythonUtil { def customedf(dataFrame: DataFrame, keyCol: String, table: String, outputSchema: StructType, database: String):

Spark 3 datasource v2: Can't extract user provided schema Dataframewriter save operation

2021-02-02 Thread Rahul Kumar
Dear friends,I'm implementing datasource v2 for a custom NoSql database. I'm facing following issuea) It seems while doing save operation, there is no way to access user specified schema on dataframe. There is an existing unresolved ongoing conversation here

Re: Implementing TableProvider in Spark 3.0

2021-02-01 Thread Rahul Kumar
Did you guys find a way to retrieve schema while saving into external database? I'm also stuck at the same place without any clear path forward. Thanks, Rahul -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/

Re: Insertable records in Datasource v2.

2021-01-14 Thread Rahul Kumar
Thank you Jacek, for trying it out and clarifying. Appreciate it. Best, Rahul -- Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ - To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Insertable records in Datasource v2.

2021-01-13 Thread Rahul Kumar
I'm implementing V2 datasource for a custom datasource. I'm trying to insert a record into a temp view, in following fashion. insertDFWithSchema.createOrReplaceTempView(sqlView) spark.sql(s”insert into $sqlView values (2, ‘insert_record1’, 200, 23000), (20001, ‘insert_record2’, 201,

[Structured spak streaming] How does cassandra connector readstream deals with deleted record

2020-06-24 Thread Rahul Kumar
Hello everyone, I was wondering, how Cassandra spark connector deals with deleted/updated record while readstream operation. If the record was already fetched in spark memory, and it got updated or deleted in database, does it get reflected in streaming join? Thanks, Rahul -- Sent from:

[Spark Streaming] predicate pushdown in custom connector source.

2020-06-23 Thread Rahul Kumar
I'm trying to implement structured spark streaming source for a custom connector. I'm wondering if it is possible to do predicate pushdown in the streaming source? I'm aware this may be something native to the datastore in question. However, I would really appreciate if someone can redirect me

Spark Context not getting initialized in local mode

2016-01-07 Thread Rahul Kumar
– Successfully stopped SparkContext Rahul Kumar *Software Engineer- I (Search Snapdeal)* *M*: +91 9023542950*EXT: *14226 362-363, ASF CENTRE , UDYOG VIHAR , PHASE - IV , GURGAON 122 016 , INDIA

Re: SparkContext.wholeTextFiles() java.io.FileNotFoundException: File does not exist:

2014-10-09 Thread Rahul Kumar Singh
I faced similar issue with wholeTextFiles function due to version compatibility. Spark 1.0 with Hadoop 2.4.1 worked. Did you try other function such as textFile to check if the issue is specific to wholeTextFiles? Spark needs to be re-compiled for different hadoop versions. However, you can keep