PIG-1769 has been created sorry I lost track of this :( On Nov 22, 2010, at 2:30 PM, Dmitriy Ryaboy wrote:
> Hm, good point. Can you create a JIRA for this? > > On Mon, Nov 22, 2010 at 1:16 PM, Corbin Hoenes <cor...@tynt.com> wrote: > >> One comment on the HBaseStorage store func. In our load statement we >> are allowed to prefix the table name with "hbase://" but when we call >> store it throws an exception unless we remove hbase:// from the table >> name: >> >> this works: >> store raw into 'piggytest2' USING >> org.apache.pig.backend.hadoop.hbase.HBaseStorage('content2:field1 >> anchor2:field1a anchor2:field2a'); >> >> this won't >> store raw into 'hbase://piggytest2' >> >> Exception: >> Caused by: java.lang.IllegalArgumentException: >> java.net.URISyntaxException: Relative path in absolute URI: >> hbase://piggytest2_logs >> >> Would be nice to be able to prefix the store with hbase:// as well. >> >> >> On Mon, Nov 22, 2010 at 12:10 PM, Dmitriy Ryaboy <dvrya...@gmail.com> >> wrote: >>> >>> Why is it connecting to localhost? >>> Sounds like you don't have the appropriate config files on the path. >>> Hm, maybe we should serialize those in the constructor so that you don't >>> have to have them on the JT classpath (I have them on the JT classpath so >>> this never came up). Can you confirm that this is the problem? >>> >>> D >>> >>> On Fri, Nov 19, 2010 at 10:33 PM, Corbin Hoenes <cor...@tynt.com> wrote: >>> >>>> Hey Jeff, >>>> >>>> It wasn't starting a job but I got a bit further by registering the >> pig8 >>>> jar in my pig script. It seemed to have a bunch of dependencies on >> google >>>> common collections; zookeeper etc... built into that jar. >>>> >>>> Now I am seeing this in the web ui logs: >>>> 2010-11-19 23:19:44,200 INFO org.apache.zookeeper.ClientCnxn: >> Attempting >>>> connection to server localhost/127.0.0.1:2181 >>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Exception >>>> closing session 0x0 to sun.nio.ch.selectionkeyi...@65efb4be >>>> java.net.ConnectException: Connection refused >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >>>> at >>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574) >>>> at >>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:885) >>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Ignoring >>>> exception during shutdown input >>>> java.nio.channels.ClosedChannelException >>>> at >>>> sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:638) >>>> at >> sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) >>>> at >>>> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:951) >>>> at >>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:922) >>>> 2010-11-19 23:19:44,201 WARN org.apache.zookeeper.ClientCnxn: Ignoring >>>> exception during shutdown output >>>> java.nio.channels.ClosedChannelException >>>> at >>>> sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:649) >>>> at >> sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368) >>>> at >>>> org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:956) >>>> at >>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:922) >>>> 2010-11-19 23:19:44,303 WARN >>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create >> /hbase >>>> -- check quorum servers, currently=localhost:2181 >>>> org.apache.zookeeper.KeeperException$ConnectionLossException: >>>> KeeperErrorCode = ConnectionLoss for /hbase >>>> Looks like it doesn't know where my hbase/conf/hbase-site.xml file is? >> Not >>>> sure how would this get passed to the HBaseStorage class? >>>> >>>> On Nov 19, 2010, at 5:09 PM, Jeff Zhang wrote: >>>> >>>>> Does the mapreduce job start ? Could you check the logs on hadoop >> side ? >>>>> >>>>> >>>>> On Sat, Nov 20, 2010 at 7:56 AM, Corbin Hoenes <cor...@tynt.com> >> wrote: >>>>>> We are trying to use the HBaseStorage LoadFunc in pig 0.8 and >> getting an >>>> exception. >>>>>> >>>>>> org.apache.pig.impl.logicalLayer.FrontendException: ERROR 1066: >> Unable >>>> to open iterator for alias raw >>>>>> at org.apache.pig.PigServer.openIterator(PigServer.java:754) >>>>>> at >>>> >> org.apache.pig.tools.grunt.GruntParser.processDump(GruntParser.java:612) >>>>>> at >>>> >> org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:303) >>>>>> at >>>> >> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165) >>>>>> at >>>> >> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:141) >>>>>> at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:76) >>>>>> at org.apache.pig.Main.run(Main.java:465) >>>>>> at org.apache.pig.Main.main(Main.java:107) >>>>>> Caused by: java.io.IOException: Couldn't retrieve job. >>>>>> at org.apache.pig.PigServer.store(PigServer.java:818) >>>>>> at org.apache.pig.PigServer.openIterator(PigServer.java:728) >>>>>> ... 7 more >>>>>> >>>>>> >>>>>> Other jobs seem to work. >>>>>> >>>>>> What are the requirements for getting hbase storage to work? >>>>>> >>>>>> This is what I am doing: >>>>>> 1 - added hbase config and hadoop config to my PIG_CLASSPATH >>>>>> 2 - pig this script: >>>>>> >>>>>> REGISTER ../lib/hbase-0.20.6.jar >>>>>> >>>>>> raw = LOAD 'hbase://piggytest' USING >>>> org.apache.pig.backend.hadoop.hbase.HBaseStorage('content:field1 >>>> anchor:field1a anchor:field2a') as (content_field1, anchor_field1a, >>>> anchor_field2a); >>>>>> >>>>>> dump raw; >>>>>> >>>>>> --- >>>>>> what else am I missing? >>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards >>>>> >>>>> Jeff Zhang >>>> >>>> >>