Hi JB,Hope all is great.I am very new to "Beam". Trying to dive into it.Am having a lot of trouble to read from a Kafka topic using google.cloud.dataflow APis & FlinkPipelineRunner.Could you point me to some docs and/or forums where I can find a solution for my problem pls?I really appreciate it. In case you are curious, given the following line:p.apply(Read.named("ReadFromKafka").from(UnboundedFlinkSource.of(kafkaConsumer))). that connects to my kafka consumer (and I confirm it at the server side), it fails & reports:The transform ReadFromKafka [Read(UnboundedFlinkSource)] is currently not supported.
I am not sure where/how I need to specify "ReadFromKafka" part.I am sending stream data from my laptop to kafka in a linux box.The default Kafka consumer reports thats its received. I really appreciate your help. I know this is not a conventional way to ask such questions but have been spending a lot of time to figure it out.Have a wonderful day JB. Amir- From: Jean-Baptiste Onofré <j...@nanthrax.net> To: dev@beam.incubator.apache.org Sent: Thursday, April 28, 2016 5:41 AM Subject: [DISCUSS] Beam IO &runners native IO Hi all, regarding the recent threads on the mailing list, I would like to start a format discussion around the IO. As we can expect the first contributions on this area (I already have some work in progress around this ;)), I think it's a fair discussion to have. Now, we have two kinds of IO: the one "generic" to Beam, the one "local" to the runners. For example, let's take Kafka: we have the KafkaIO (in IO), and for instance, we have the spark-streaming kafka connector (in Spark Runner). Right now, we have two approaches for the user: 1. In the pipeline, we use KafkaIO from Beam: it's the preferred approach for sure. However, the user may want to use the runner specific IO for two reasons: * Beam doesn't provide the IO yet (for instance, spark cassandra connector is available whereas we don't have yet any CassandraIO (I'm working on it anyway ;)) * The runner native IO is optimized or contain more features that the Beam native IO 2. So, for the previous reasons, the user could want to use the native runner IO. The drawback of this approach is that the pipeline will be tight to a specific runner, which is completely against the Beam design. I wonder if it wouldn't make sense to add flag on the IO API (and related on Runner API) like .useNative(). For instance, the user would be able to do: pipeline.apply(KafkaIO.read().withBootstrapServers("...").withTopics("...").useNative(true); then, if the runner has a "native" IO, it will use it, else, if useNative(false) (the default), it won't use any runner native IO. The point there is for the configuration: assuming the Beam IO and the runner IO can differ, it means that the "Beam IO" would have to populate all runner specific IO configuration. Of course, it's always possible to use a PTransform to wrap the runner native IO, but we are back on the same concern: the pipeline will be couple to a specific runner. The purpose of the useNative() flag is to "automatically" inform the runner to use a specific IO if it has one: the pipeline stays decoupled from the runners. Thoughts ? Thanks Regards JB -- Jean-Baptiste Onofré jbono...@apache.org http://blog.nanthrax.net Talend - http://www.talend.com