i implemented a sink using foreach it was indeed straightforward thanks

On Fri, Jan 13, 2017 at 6:30 PM, Tathagata Das <tathagata.das1...@gmail.com>
wrote:

> Structured Streaming has a foreach sink, where you can essentially do what
> you want with your data. Its easy to create a Kafka producer, and write the
> data out to kafka.
> http://spark.apache.org/docs/latest/structured-streaming-
> programming-guide.html#using-foreach
>
> On Fri, Jan 13, 2017 at 8:28 AM, Koert Kuipers <ko...@tresata.com> wrote:
>
>> how do you do this with structured streaming? i see no mention of writing
>> to kafka
>>
>> On Fri, Jan 13, 2017 at 10:30 AM, Peyman Mohajerian <mohaj...@gmail.com>
>> wrote:
>>
>>> Yes, it is called Structured Streaming: https://docs.databr
>>> icks.com/_static/notebooks/structured-streaming-kafka.html
>>> http://spark.apache.org/docs/latest/structured-streaming-pro
>>> gramming-guide.html
>>>
>>> On Fri, Jan 13, 2017 at 3:32 AM, Senthil Kumar <senthilec...@gmail.com>
>>> wrote:
>>>
>>>> Hi Team ,
>>>>
>>>>      Sorry if this question already asked in this forum..
>>>>
>>>> Can we ingest data to Apache Kafka Topic from Spark SQL DataFrame ??
>>>>
>>>> Here is my Code which Reads Parquet File :
>>>>
>>>> *val sqlContext = new org.apache.spark.sql.SQLContext(sc);*
>>>>
>>>> *val df = sqlContext.read.parquet("..../temp/*.parquet")*
>>>>
>>>> *df.registerTempTable("beacons")*
>>>>
>>>>
>>>> I want to directly ingest df DataFrame to Kafka ! Is there any way to
>>>> achieve this ??
>>>>
>>>>
>>>> Cheers,
>>>>
>>>> Senthil
>>>>
>>>
>>>
>>
>

Reply via email to