Kafka deserialization to Structured Streaming SQL - Encoders.bean result doesn't match itself?

2018-01-25 Thread Iain Cundy
Hi All I'm trying to move from MapWithState to Structured Streaming v2.2.1, but I've run into a problem. To convert from Kafka data with a binary (protobuf) value to SQL I'm taking the dataset from readStream and doing Dataset s = dataset.selectExpr("timestamp", "CAST(key as string)",

RE: mapWithState not compacting removed state

2016-04-07 Thread Iain Cundy
which you can monitor in the task data in ApplicationMaster web UI. Cheers Iain From: Ofir Kerker [mailto:ofir.ker...@gmail.com] Sent: 07 April 2016 17:06 To: Iain Cundy; user@spark.apache.org Subject: Re: mapWithState not compacting removed state Hi Iain, Did you manage to solve this issue

RE: [MARKETING] Timeout in mapWithState

2016-04-05 Thread Iain Cundy
Hi Abhi The concept is what you want – if you set StateSpec timeout to a Duration of 10 minutes then any keys no seen for more than 10 minutes will be deleted. However you did say “exactly” and specifically mention “removed from memory” in which case you may be interested in the much more

RE: [MARKETING] Re: mapwithstate Hangs with Error cleaning broadcast

2016-03-15 Thread Iain Cundy
Hi Manas I saw a very similar problem while using mapWithState. Timeout on BlockManager remove leading to a stall. In my case it only occurred when there was a big backlog of micro-batches, combined with a shortage of memory. The adding and removing of blocks between new and old tasks was

mapWithState not compacting removed state

2016-03-03 Thread Iain Cundy
r some reason? Thank you for your attention. Cheers Iain Cundy This message and the information contained herein is proprietary and confidential and subject to the Amdocs policy statement, you may review at http://www.amdocs.com/email_disclaimer.asp