Hi JB,Hope all is great.I am very new to "Beam". Trying to dive into it.Am 
having a lot of trouble to read from a Kafka topic using google.cloud.dataflow  
APis & FlinkPipelineRunner.Could you point me to some docs and/or forums where 
I can find a solution for my problem pls?I really appreciate it.
In case you are curious, given the following 
line:p.apply(Read.named("ReadFromKafka").from(UnboundedFlinkSource.of(kafkaConsumer))).
that connects to my kafka consumer (and I confirm it at the server side), it 
fails & reports:The transform ReadFromKafka [Read(UnboundedFlinkSource)] is 
currently not supported.

I am not sure where/how I need to specify "ReadFromKafka" part.I am sending 
stream data from my laptop to kafka in  a linux box.The default Kafka consumer 
reports thats its received.
I really appreciate your help. I know this is not a conventional way to ask 
such questions but have been spending a lot of time to figure it out.Have a 
wonderful day JB.
Amir-
      From: Jean-Baptiste Onofré <j...@nanthrax.net>
 To: dev@beam.incubator.apache.org 
 Sent: Thursday, April 28, 2016 5:41 AM
 Subject: [DISCUSS] Beam IO &runners native IO
   
Hi all,

regarding the recent threads on the mailing list, I would like to start 
a format discussion around the IO.
As we can expect the first contributions on this area (I already have 
some work in progress around this ;)), I think it's a fair discussion to 
have.

Now, we have two kinds of IO: the one "generic" to Beam, the one "local" 
to the runners.

For example, let's take Kafka: we have the KafkaIO (in IO), and for 
instance, we have the spark-streaming kafka connector (in Spark Runner).

Right now, we have two approaches for the user:
1. In the pipeline, we use KafkaIO from Beam: it's the preferred 
approach for sure. However, the user may want to use the runner specific 
IO for two reasons:
    * Beam doesn't provide the IO yet (for instance, spark cassandra 
connector is available whereas we don't have yet any CassandraIO (I'm 
working on it anyway ;))
    * The runner native IO is optimized or contain more features that the 
Beam native IO
2. So, for the previous reasons, the user could want to use the native 
runner IO. The drawback of this approach is that the pipeline will be 
tight to a specific runner, which is completely against the Beam design.

I wonder if it wouldn't make sense to add flag on the IO API (and 
related on Runner API) like .useNative().

For instance, the user would be able to do:

pipeline.apply(KafkaIO.read().withBootstrapServers("...").withTopics("...").useNative(true);

then, if the runner has a "native" IO, it will use it, else, if 
useNative(false) (the default), it won't use any runner native IO.

The point there is for the configuration: assuming the Beam IO and the 
runner IO can differ, it means that the "Beam IO" would have to populate 
all runner specific IO configuration.

Of course, it's always possible to use a PTransform to wrap the runner 
native IO, but we are back on the same concern: the pipeline will be 
couple to a specific runner.

The purpose of the useNative() flag is to "automatically" inform the 
runner to use a specific IO if it has one: the pipeline stays decoupled 
from the runners.

Thoughts ?

Thanks
Regards
JB
-- 
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com


  

Reply via email to