Hi guys. Thanks for helping out.
We downgraded to HBase 0.98 and resolved some classpath issues and then it worked. /Palle ----- Original meddelelse ----- > Fra: Stephan Ewen <se...@apache.org> > Til: user@flink.apache.org > Dato: Ons, 11. maj 2016 17:19 > Emne: Re: HBase write problem > > Just to narrow down the problem:The insertion into HBase actually > works, but the job does not finish after that?And the same job (same > source of data) that writes to a file, or prints, finishes?If that is > the case, can you check what status each task is in, via the web > dashboard? Are all tasks still in "running"? > > On Wed, May 11, 2016 at 4:53 PM, Flavio Pompermaier < > pomperma...@okkam.it [mailto:pomperma...@okkam.it]> wrote: > > I can't help you with the choice of the db storage, as always the > answer is "it depends" on a lot of factors :) > > For what I can tell you the problem could be that Flink support > HBase 0.98, so it could worth to update Flink connectors to a > more recent version (that should be backward compatible > hopefully..) or maybe create two separte hbase connectors (one > for hbase-0.9x and one for 0.1x). Let me know about your attempts > :) > > On Wed, May 11, 2016 at 4:47 PM, Palle < pa...@sport.dk > [mailto:pa...@sport.dk]> > wrote: > > Hadoop 2.7.2 > HBase 1.2.1 > > I have this running from a Hadoop job, but just not from > Flink. > > I will look into your suggestions, but would I be better off > choosing another DB for storage? I can see that Cassandra > gets some attention in this mailing list. I need to store app > 2 bio key value pairs consisting of 100 bytes for each pair. > > ----- Original meddelelse ----- > > Fra: Flavio Pompermaier < pomperma...@okkam.it > [mailto:pomperma...@okkam.it]> > Til: user < user@flink.apache.org [mailto:user@flink.apache.org]> > Dato: Ons, 11. maj 2016 16:29 > Emne: Re: HBase write problem > > And which version of HBase and Hadoop are you running? > Did you try to put the hbase-site.xml in the jar?Moreover, > I don't know how much reliable is at the moment the web > client UI..my experience is that the command line client > is much more reliable.You just need to run from the flink > dir something like: bin/flink run -c xxx.yyy.MyMainClass > /path/to/shadedJar.jar > On Wed, May 11, 2016 at 4:19 PM, Palle < pa...@sport.dk > [mailto:pa...@sport.dk]> > wrote: > > I run the job from the cluster. I run it through the > web UI. > The jar file submitted does not contain the > hbase-site.xml file. > > ----- Original meddelelse ----- > > Fra: Flavio Pompermaier < pomperma...@okkam.it > [mailto:pomperma...@okkam.it]> > Til: user < user@flink.apache.org [mailto:user@flink.apache.org]> > Dato: Ons, 11. maj 2016 09:36 > Emne: Re: HBase write problem > > Do you run the job from your IDE or from the > cluster? > > On Wed, May 11, 2016 at 9:22 AM, Palle < > pa...@sport.dk [mailto:pa...@sport.dk]> wrote: > > Thanks for the response, but I don't think > the problem is the classpath - hbase-site.xml > should be added. This is what it looks like > (hbase conf is added at the end): > > 2016-05-11 09:16:45,831 INFO > org.apache.zookeeper.ZooKeeper - Client > > environment:java.class.path=C:\systems\packages\flink-1.0.2\lib\flink-dist_2.11-1.0.2.jar;C:\systems\packages\flink-1.0.2\lib\flink-python_2.11-1.0.2.jar;C:\systems\packages\flink-1.0.2\lib\guava-11.0.2.jar;C:\systems\packages\flink-1.0.2\lib\hbase-annotations-1.2.1-tests.jar;C:\systems\packages\flink-1.0.2\lib\hbase-annotations-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-client-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-common-1.2.1-tests.jar;C:\systems\packages\flink-1.0.2\lib\hbase-common-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-examples-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-external-blockcache-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-hadoop-compat-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-hadoop2-compat-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-it-1.2.1-tests.jar;C:\systems\packages\flink-1.0.2\lib\hbase-it-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-prefix-tree-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-procedure-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-protocol-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-resource-bundle-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-rest-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-server-1.2.1-tests.jar;C:\systems\packages\flink-1.0.2\lib\hbase-server-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-shell-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\hbase-thrift-1.2.1.jar;C:\systems\packages\flink-1.0.2\lib\log4j-1.2.17.jar;C:\systems\packages\flink-1.0.2\lib\slf4j-log4j12-1.7.7.jar;C:\systems\master_flink\bin;C:\systems\packages\flink-1.0.2\lib;;C:\systems\packages\hbase-1.2.1\lib;C:\systems\hbase\conf;C:\systems\hbase\conf\hbase-site.xml; > 2016-05-11 09:16:45,831 INFO > org.apache.zookeeper.ZooKeeper - Client > > environment:java.library.path=C:\systems\packages\jre-1.8.0_74_x64\bin;C:\Windows\Sun\Java\bin;C:\Windows\system32;C:\Windows;C:\systems\master_flink\bin;C:\systems\packages\appsync-1.0.6\bin;C:\systems\packages\flink-1.0.2\bin;C:\systems\packages\jre-1.8.0_74_x64\bin;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Program > Files\Microsoft SQL > Server\120\DTS\Binn\;C:\Program > Files\Microsoft SQL Server\Client > SDK\ODBC\110\Tools\Binn\;C:\Program Files > (x86)\Microsoft SQL > Server\120\Tools\Binn\;C:\Program > Files\Microsoft SQL > Server\120\Tools\Binn\;C:\Program Files > (x86)\Microsoft SQL > Server\120\Tools\Binn\ManagementStudio\;C:\Program > Files (x86)\Microsoft SQL > > Server\120\DTS\Binn\;C:\systems\packages\hadoop-2.7.2\bin;C:\systems\packages\jdk-1.8.0_74_x64\bin;C:\systems\packages\apache-maven-3.3.9\bin;C:\systems\packages\protoc-2.5.0-win32\;C:\systems\packages\cygwin64\bin\;C:\systems\packages\cmake-3.5.2-win32-x86\bin;C:\Program > Files\Microsoft Windows Performance > > Toolkit\;C:\systems\packages\perl-5.6.0-win\bin;C:\systems\hbase\conf;. > ----- Original meddelelse ----- > > Fra: Flavio Pompermaier < > pomperma...@okkam.it [mailto:pomperma...@okkam.it]> > Til: user < user@flink.apache.org > [mailto:user@flink.apache.org]> > Dato: Ons, 11. maj 2016 00:05 > Emne: Re: HBase write problem > > Do you have the hbase-site.xml available > in the classpath? > > On 10 May 2016 23:10, "Palle" < > pa...@sport.dk [mailto:pa...@sport.dk]> > wrote: > > HBase write problem > > Hi all. > > I have a problem writing to HBase. > > I am using a slightly modified > example of this class to proof the > concept: > > https://github.com/apache/flink/blob/master/flink-batch-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase/example/HBaseWriteExample.java > > [https://github.com/apache/flink/blob/master/flink-batch-connectors/flink-hbase/src/test/java/org/apache/flink/addons/hbase/example/HBaseWriteExample.java] > > However all the HBase-specific stuff > is exactly the same as in the > HBaseWriteExample. > > The problem I see is that the job > never completes (been running for > more than an hour now) and it is only > 13 key/value pairs that is to be > written to HBase :-) > I have tested the map/reduce stuff > works if I replace the HBase > connection stuff with just a write to > a text file - works OK. I have also > tested that I can insert data in > HBase from a similar Hadoop MapReduce > job. > > Here is the part of the code where I > guess the problem is: > > @Override > public Tuple2<Text, Mutation> > map(Tuple2<String, Integer> t) throws > Exception { > LOG.info("Tuple2 map() called"); > reuse.f0 = new Text(t.f0); > Put put = new Put(t.f0.getBytes()); > put.add(MasterConstants.CF_SOME, > MasterConstants.COUNT, > Bytes.toBytes(t.f1)); > reuse.f1 = put; > return reuse; > } > }).output(new > HadoopOutputFormat<Text, > Mutation>(new > TableOutputFormat<Text>(), job)); > > env.execute("Flink HBase Event Count > Hello World Test"); > > This code matches the code in the > HBaseWriteExample.java I should > think. > > The "Tuple2" log line I see exactly > the 13 times I expect, and the last > log line I see is this: > 2016-05-10 21:48:42,715 INFO > org.apache.hadoop.hbase.mapreduce.TableOutputFormat > - Created table instance for > event_type_count > > Any suggestions to what the problem > could be? > > Thanks, > Palle