supposedly gmail wont strip \n so we'll try again with non-html mail composer
noob with Scala so Im looking for an experienced answer ConsumerGroupCommand.scala //private def createNewConsumer(): KafkaConsumer[String, String] = { //private def createNewConsumer(): KafkaConsumer[K extends // java.util.ArrayList[Byte],V extends java.util.ArrayList[Byte]] = { private def createNewConsumer(): KafkaConsumer[K <: java.util.ArrayList[Byte],V <: java.util.ArrayList[Byte]] = { val properties = new java.util.Properties() val deserializer = (new StringDeserializer).getClass.getName val brokerUrl = opts.options.valueOf(opts.bootstrapServerOpt) properties.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, brokerUrl) properties.put(ConsumerConfig.GROUP_ID_CONFIG,opts.options.valueOf(opts.groupOpt)) properties.put(ConsumerConfig.ENABLE_AUTO_COMMIT_CONFIG, "false") properties.put(ConsumerConfig.SESSION_TIMEOUT_MS_CONFIG, "30000") properties.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, deserializer) properties.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, deserializer) if (opts.options.has(opts.commandConfigOpt)) properties.putAll(Utils.loadProps(opts.options.valueOf(opts.commandConfigOpt))) new KafkaConsumer(properties).asInstanceOf[KafkaConsumer[K,V]] } scala-compiler displays: [ERROR] \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:309: error: ']' expected but '<:' found. [ERROR] private def createNewConsumer(): KafkaConsumer[? <: java.util.ArrayList[Byte],? <: java.util.ArrayList[Byte]] = { [ERROR] ^ [ERROR] \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:309: error: '=' expected but ',' found. [ERROR] private def createNewConsumer(): KafkaConsumer[? <: java.util.ArrayList[Byte],? <: java.util.ArrayList[Byte]] = { [ERROR] ^ [ERROR] \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:322: error: illegal start of simple expression i need 2 datatype parameter types extending java.util.ArrayList<Byte> in regular java this would be: public KafkaConsumer<K extends java.util.ArrayList<Byte>,V extends java.util.ArrayList<Byte>> createNewConsumer() { ....} this works in java but how do I setup a function or class declaration K, V whose parameter datatype extends java.util.ArrayList<Byte> ? Martin >> >> From: mgai...@hotmail.com >> To: mathieu.fenn...@replicon.com; users@kafka.apache.org >> Subject: RE: handling generics in Kafka Scala >> Date: Tue, 30 Aug 2016 23:00:29 -0400 >> >> >> >> >> noob with Scala so Im looking for an experienced answer >> ConsumerGroupCommand.scala >> //private def createNewConsumer(): KafkaConsumer[String, String] = >> {//private def createNewConsumer(): KafkaConsumer[K extends >> java.util.ArrayList[Byte],V extends java.util.ArrayList[Byte]] = { >> private def createNewConsumer(): KafkaConsumer[K <: >> java.util.ArrayList[Byte],V <: java.util.ArrayList[Byte]] = { val >> properties = new java.util.Properties() val deserializer = (new >> StringDeserializer).getClass.getName val brokerUrl = >> opts.options.valueOf(opts.bootstrapServerOpt) >> properties.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, brokerUrl) >> properties.put(ConsumerConfig.GROUP_ID_CONFIG, >> opts.options.valueOf(opts.groupOpt)) >> properties.put(ConsumerConfig.ENABLE_AUTO_COMMIT_CONFIG, "false") >> properties.put(ConsumerConfig.SESSION_TIMEOUT_MS_CONFIG, "30000") >> properties.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, >> deserializer) >> properties.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, >> deserializer) if (opts.options.has(opts.commandConfigOpt)) >> properties.putAll(Utils.loadProps(opts.options.valueOf(opts.commandConfigOpt))) >> new KafkaConsumer(properties).asInstanceOf[KafkaConsumer[K,V]] } >> scala-compiler displays: >> [ERROR] >> \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:309: >> error: ']' expected but '<:' found.[ERROR] private def >> createNewConsumer(): KafkaConsumer[? <: java.util.ArrayList[Byte],? <: >> java.util.ArrayList[Byte]] = {[ERROR] >> ^[ERROR] >> \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:309: >> error: '=' expected but ',' found.[ERROR] private def >> createNewConsumer(): KafkaConsumer[? <: java.util.ArrayList[Byte],? <: >> java.util.ArrayList[Byte]] = {[ERROR] >> ^[ERROR] >> \kafka\kafka-trunk\core\src\main\scala\kafka\admin\ConsumerGroupCommand.scala:322: >> error: illegal start of simple expression >> i want 2 datatype parameter types extending java.util.ArrayList<Byte> in >> regular java this would be: >> public KafkaConsumer<K extends java.util.ArrayList<Byte>,V extends >> java.util.ArrayList<Byte>> createNewConsumer() {} >> how do I setup a function or class declaration K, V whose parameter >> datatype >> extends java.util.ArrayList<Byte> ? >> Martin >> ______________________________________________ >> >> >> >>> From: mathieu.fenn...@replicon.com >>> Date: Wed, 17 Aug 2016 18:06:38 -0600 >>> Subject: Re: DLL Hell >>> To: mgai...@hotmail.com >>> >>> Hi Martin, >>> >>> I'm sorry, this is way outside my Kafka knowledge. I'm just a new >>> Kafka user who wanted to help with your Windows questions because I >>> had just faced the same hurdle. :-) Wish I could help, but I wouldn't >>> know where to start with this. >>> >>> Mathieu >>> >>> >>> On Wed, Aug 17, 2016 at 6:00 PM, Martin Gainty <mgai...@hotmail.com> >>> wrote: >>> > Hi Matthieu >>> > Many Thanks for attaching the binary >>> > >>> > running scala->java generator plugin I see: >>> > >>> > [ERROR] >>> > C:\Maven-plugin\kafka\kafka-trunk\core\src\main\scala\kafka\admin\AdminUtils.scala:639: >>> > error: type PartitionMetadata is not a member of object >>> > org.apache.kafka.common.requests.MetadataResponse >>> > >>> > yet when I look at >>> > org.apache.kafka.common.requests.MetadataResponse.java I >>> > see inner class >>> > >>> > public static class PartitionMetadata { >>> > >>> > inner static java classes are not visible to the converter for some >>> > reason >>> > the workaround seems to be birth inner static classes (e.g. >>> > PartitionMetadata) >>> > treating inner class as standalone works >>> > >>> > Advice? >>> > Martin >>> > ______________________________________________ >>> > >>> > >>> > >>> > >>> > ________________________________ >>> > From: mathieu.fenn...@replicon.com >>> > Date: Tue, 16 Aug 2016 08:04:52 -0600 >>> > Subject: Re: DLL Hell >>> > To: mgai...@hotmail.com >>> > >>> > >>> > Hey Martin, >>> > >>> > Attached is the native .dll that I was able to build for rocksdb. If >>> > you >>> > unzip this, and include the contained .dll into your >>> > rocksdbjni-4.8.0.jar at >>> > the root, it should be possible to use Kafka Streams in Windows. But >>> > this >>> > is just a minimal debug build; wouldn't be appropriate for production >>> > use. >>> > Might save you some time if you're just trying to get a dev >>> > environment >>> > working though. >>> > >>> > Mathieu >>> > >>> > >>> > On Tue, Aug 16, 2016 at 7:40 AM, Martin Gainty <mgai...@hotmail.com> >>> > wrote: >>> > >>> > >>> > >>> > >>> >> From: mathieu.fenn...@replicon.com >>> >> Date: Tue, 16 Aug 2016 06:57:16 -0600 >>> >> Subject: Re: DLL Hell >>> >> To: users@kafka.apache.org >>> >> >>> >> Hey Martin, >>> >> >>> >> I had to modify the -G argument to that command to include the visual >>> >> studio year. If you run "cmake /?", it will output all the available >>> >> generators. My cmake looked like: >>> >> >>> >> cmake -G "Visual Studio 12 2013 Win64" -DJNI=1 .. >>> >> >>> >> I think this is probably a change in cmake since the rocksdb doc was >>> >> written ( >>> >> >>> >> https://cmake.org/cmake/help/v3.0/generator/Visual%20Studio%2012%202013.html >>> >> ). >>> >> MG>same "informative error" >>> >>C:\cygwin64\bin\cmake -G "Visual Studio 12 2013 Win64" -DJNI=1 >>> > CMake Error: Could not create named generator Visual Studio 12 2013 >>> > Win64 >>> > Generators Unix Makefiles = Generates standard UNIX >>> > makefiles. Ninja = Generates build.ninja >>> > files. >>> > CodeBlocks - Ninja = Generates CodeBlocks project files. >>> > CodeBlocks - Unix Makefiles = Generates CodeBlocks project files. >>> > CodeLite >>> > - Ninja = Generates CodeLite project files. CodeLite - >>> > Unix >>> > Makefiles = Generates CodeLite project files. Eclipse CDT4 - Ninja >>> > = Generates Eclipse CDT 4.0 project files. Eclipse CDT4 - Unix >>> > Makefiles= >>> > Generates Eclipse CDT 4.0 project files. KDevelop3 >>> > = >>> > Generates KDevelop 3 project files. KDevelop3 - Unix Makefiles = >>> > Generates KDevelop 3 project files. Kate - Ninja = >>> > Generates Kate project files. Kate - Unix Makefiles = >>> > Generates >>> > Kate >>> > project files. Sublime Text 2 - Ninja = Generates Sublime Text >>> > 2 >>> > project files. Sublime Text 2 - Unix Makefiles >>> > = Generates Sublime Text 2 project files. >>> > MG>I am thinking if I want to automate this native build..I could more >>> > easily create binary thru maven-nar-plugin ? >>> > MG>as I do not have any MS VS or DotNet installed..maybe I need to >>> > install >>> > many gigs of MS specific VS? >>> > MG>Please advise >>> >> Mathieu >>> >> >>> >> >>> >> On Tue, Aug 16, 2016 at 5:03 AM, Martin Gainty <mgai...@hotmail.com> >>> >> wrote: >>> >> >>> >> > havent used cmake in over 10 years so Im a bit lost.. >>> >> > cmake -G "Visual Studio 12 Win64" -DGFLAGS=1 -DSNAPPY=1 >>> >> > -DJEMALLOC=1 >>> >> > -DJNI=1 >>> >> > CMake Error: Could not create named generator Visual Studio 12 >>> >> > Win64 >>> >> > ?Please advise >>> >> > Martin >>> >> > ______________________________________________ >>> >> > >>> >> > >>> >> > >>> >> > > From: mathieu.fenn...@replicon.com >>> >> > > Date: Mon, 15 Aug 2016 13:43:47 -0600 >>> >> > > Subject: Re: DLL Hell >>> >> > > To: users@kafka.apache.org >>> >> > > >>> >> > > Hi Martin, >>> >> > > >>> >> > > rocksdb does not currently distribute a Windows-compatible build >>> >> > > of >>> >> > > their >>> >> > > rocksdbjni library. I recently wrote up some instructions on how >>> >> > > to >>> >> > > produce a local build, which you can find here: >>> >> > > http://mail-archives.apache.org/mod_mbox/kafka-users/ >>> >> > 201608.mbox/%3CCAHoiPjweo-xSj3TiodcDVf4wNnnJ8u6PcwWDPF7L >>> >> > T5ps%2BxQ3eA%40mail.gmail.com%3E >>> >> > > >>> >> > > I'd also suggest tracking this issue in GitHub, which is likely >>> >> > > to >>> >> > > be >>> >> > > updated if this ever changes: https://github.com/facebook/ >>> >> > rocksdb/issues/703 >>> >> > > >>> >> > > Mathieu >>> >> > > >>> >> > > >>> >> > > On Mon, Aug 15, 2016 at 1:34 PM, Martin Gainty >>> >> > > <mgai...@hotmail.com> >>> >> > wrote: >>> >> > > >>> >> > > > kafka-trunk\streams>gradle buildCaused by: >>> >> > > > java.lang.RuntimeException: >>> >> > > > librocksdbjni-win64.dll was not found inside JAR. at >>> >> > org.rocksdb. >>> >> > > > NativeLibraryLoader.loadLibraryFromJarToTemp( >>> >> > NativeLibraryLoader.java:106) >>> >> > > > at org.rocksdb.NativeLibraryLoader.loadLibraryFromJar( >>> >> > NativeLibraryLoader.java:78) >>> >> > > > at org.rocksdb.NativeLibraryLoader.loadLibrary( >>> >> > NativeLibraryLoader.java:56) >>> >> > > > at org.rocksdb.RocksDB.loadLibrary(RocksDB.java:47) at >>> >> > > > org.rocksdb.RocksDB.<clinit>(RocksDB.java:23) >>> >> > > > any idea where I can locale librocksdbjni-win64.dll ? >>> >> > > > /thanks/ >>> >> > > > Martin >>> >> > > > ______________________________________________ >>> >> > > > >>> >> > > > >>> >> > >>> >> > >>> > >>> > >>> > >> >