Judging from the code, you should separate different jars with a colon
":", i.e. "—addclasspath jar1:jar2"


Nico

On 26/02/18 10:36, kant kodali wrote:
> Hi Gordon,
> 
> Thanks for the response!! How do I add multiple jars to the classpaths?
> Are they separated by a semicolon and still using one flag like
> "—addclasspath jar1; jar2" or specify the flag multiple times like
> "—addclasspath jar1 —addclasspath jar2" or specify just the directory
> "—addclasspath ./opt" so it adds all the jars in that directory!
> 
> Thanks!
> 
> On Sun, Feb 25, 2018 at 11:29 PM, Tzu-Li (Gordon) Tai
> <tzuli...@apache.org <mailto:tzuli...@apache.org>> wrote:
> 
>     Hi,
> 
>     Good to see that you have it working! Yes, each of the Kafka
>     version-specific connectors also have a dependency on the base Kafka
>     connector module.
> 
>     Note that it is usually not recommended to put optional dependencies
>     (such as the connectors) under the lib folder.
>     To add additional dependencies when using the Scala shell, there is
>     a “—addclasspath” option which allows you to specify paths to the
>     dependency jars.
> 
>     Cheers,
>     Gordon
> 
> 
>     On 25 February 2018 at 12:22:28 PM, kant kodali (kanth...@gmail.com
>     <mailto:kanth...@gmail.com>) wrote:
> 
>>     Exception went away after
>>     downloading flink-connector-kafka-base_2.11-1.4.1.jar to lib folder
>>
>>     On Sat, Feb 24, 2018 at 6:36 PM, kant kodali <kanth...@gmail.com
>>     <mailto:kanth...@gmail.com>> wrote:
>>
>>         Hi,
>>
>>         I couldn't get flink and kafka working together. It looks like
>>         all examples I tried from web site fails with the following
>>         Exception.
>>
>>         Caused by: java.lang.ClassNotFoundException:
>>         org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase
>>
>>
>>         *or when I do something like this like it is in the website*
>>
>>
>>          val stream = senv.addSource(new
>>         FlinkKafkaConsumer08[String]("join_test", new
>>         SimpleStringSchema(), properties)).print()
>>
>>         *I get the following exception*
>>
>>         <console>:73: error: overloaded method value addSource with
>>         alternatives:
>>
>>           [T](function:
>>         
>> org.apache.flink.streaming.api.functions.source.SourceFunction.SourceContext[T]
>>         => Unit)(implicit evidence$10:
>>         
>> org.apache.flink.api.common.typeinfo.TypeInformation[T])org.apache.flink.streaming.api.scala.DataStream[T]
>>         <and>
>>
>>           [T](function:
>>         
>> org.apache.flink.streaming.api.functions.source.SourceFunction[T])(implicit
>>         evidence$9:
>>         
>> org.apache.flink.api.common.typeinfo.TypeInformation[T])org.apache.flink.streaming.api.scala.DataStream[T]
>>
>>          cannot be applied to (org.apache.flink.streaming.co
>>         
>> <http://org.apache.flink.streaming.co>nnectors.kafka.FlinkKafkaConsumer08[String])
>>
>>                val stream = senv.addSource(new
>>         FlinkKafkaConsumer08[String]("join_test", new
>>         SimpleStringSchema(), properties)).print()
>>
>>
>>         can anyone share a simple example of how to get Kafka Stream
>>         as a Table using scala shell? No need for any fancy schema
>>         just needs to print the value. I am using the latest version
>>         of flink 1.41 and my lib folder
>>         containers flink-connector-kafka-0.8_2.11-1.4.1.jar 
>>
>>         I wanted to use Kafka 0.9 but that didn't work so I thought
>>         let me just get something working first and downgraded to 0.8
>>         but 0.8 examples on the website also don't seem to work using
>>         scala shell. 
>>
>>         Thanks!!
>>
>>
>>
>>
> 

Attachment: signature.asc
Description: OpenPGP digital signature

Reply via email to