Hbase Bulk Load - Unknown Host Exception

2015-07-15 Thread Vineet Mishra
Hi All,

I am facing a strange issue, I am running a Hbase Bulk Load to load a Hfile
to my hbase table, while running the same I am landing into the same issue
over and over again.

java.io.IOException: BulkLoad encountered an unrecoverable problem
at
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:381)
at
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:310)
at
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.run(LoadIncrementalHFiles.java:896)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.kylin.job.hadoop.hbase.BulkLoadJob.run(BulkLoadJob.java:83)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at
org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed
after attempts=35, exceptions:
Tue Jul 14 23:18:48 PDT 2015,
org.apache.hadoop.hbase.client.RpcRetryingCaller@216b1af0,
java.net.UnknownHostException: unknown host: prod-hadoop-data02
Tue Jul 14 23:18:49 PDT 2015,
org.apache.hadoop.hbase.client.RpcRetryingCaller@216b1af0,
java.net.UnknownHostException: unknown host: prod-hadoop-data02

So there are multiple jobs which are initiating table load process out of
which one of the job is failing intermittently, let me clarify the failing
job is not the same everytime.
So for instance last time my job 1 got failed but today its job 2, but out
of all the exception remains the same.

I am having the respective host entry on the all my hosts hadoop-yarn 10
node cluster(10 data node).

Prominent suggestion are appreciated.

Thanks!


Re: Hbase not taking inserts from Remote Machine

2015-02-17 Thread Vineet Mishra
Thanks Serega!

Don't know how could I miss that, Its working good now!

On Tue, Feb 17, 2015 at 12:41 PM, Serega Sheypak 
wrote:

> You need to open region server ports. Client directly sends put to
> appropriate region server.
>
> вторник, 17 февраля 2015 г. пользователь Vineet Mishra написал:
>
> > -- Forwarded message ------
> > From: Vineet Mishra >
> > Date: Tue, Feb 17, 2015 at 12:32 PM
> > Subject: Hbase not taking inserts from Remote Machine
> > To: cdh-u...@cloudera.org 
> >
> >
> >
> > Hi All,
> >
> > I am trying to communicate and insert some data to my
> > Hbase(0.98.6-cdh5.3.0) running on Hadoop 2.5 using Hbase Java API.
> >
> > Although if I run the code within the cluster it connects fine but If I
> am
> > outside cluster, even though I have opened the port on external IPs for
> > Zookeeper and HMaster, I am stuck without any error on the logs, so the
> > code hangs up on Table.put() during insertion,
> >
> > Attached below is the Stack Trace for the Job,
> >
> > 15/02/17 12:05:08 WARN util.NativeCodeLoader: Unable to load
> native-hadoop
> > library for your platform... using builtin-java classes where applicable
> > 15/02/17 12:05:08 INFO zookeeper.RecoverableZooKeeper: Process
> > identifier=hconnection-0x77dacebf connecting to ZooKeeper ensemble=
> > namenode.com:2181,cloud-manager.com:2181
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:zookeeper.version=3.4.5-cdh5.3.0--1, built on 12/17/2014
> 02:45
> > GMT
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:host.name
> > =ip-20-0-0-75
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.version=1.7.0_75
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.vendor=Oracle Corporation
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.75.x86_64/jre
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.class.path=.:hbase-connect-1.0.0.jar
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> >
> >
> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.io.tmpdir=/tmp
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:java.compiler=
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:os.name
> > =Linux
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:os.arch=amd64
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:os.version=3.14.23-22.44.amzn1.x86_64
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:user.name
> > =tom
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:user.home=/home/tom
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
> > environment:user.dir=/home/tom/jobs
> > 15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Initiating client connection,
> > connectString=namenode.com:2181,cloud-manager.com:2181
> > sessionTimeout=9
> > watcher=hconnection-0x77dacebf, quorum=namenode.com:2181,
> > cloud-manager.com:2181, baseZNode=/hbase
> > 15/02/17 12:05:08 INFO zookeeper.ClientCnxn: Opening socket connection to
> > server namenode.com/54.172.21.54:2181. Will not attempt to authenticate
> > using SASL (unknown error)
> > 15/02/17 12:05:09 INFO zookeeper.ClientCnxn: Socket connection
> established
> > to namenode.com/54.172.21.54:2181, initiating session
> > 15/02/17 12:05:09 INFO zookeeper.ClientCnxn: Session establishment
> complete
> > on server namenode.com/54.172.21.54:2181, sessionid = 0x24b7c7ba3532b32,
> > negotiated timeout = 6
> > 15/02/17 12:05:11 INFO
> client.HConnectionManager$HConnectionImplementation:
> > Closing master protocol: MasterService
> > 15/02/17 12:05:11 INFO
> client.HConnectionManager$HConnectionImplementation:
> > Closing zookeeper sessionid=0x24b7c7ba3532b32
> > 15/02/17 12:05:11 INFO zookeeper.ZooKeeper: Session: 0x24b7c7ba3532b32
> > closed
> > 15/02/17 12:05:11 INFO zookeeper.ClientCnxn: EventThread shut down
> > Hbase Running
> > 15/02/17 12:05:11 INFO zookeeper.RecoverableZooKeeper: Process
> > identifier=hconnection-0x1003cac6 connecting to ZooKeeper ensemble=
> > namenode.com:2181,cloud-manager.com:2181
> > 15/02/17 12:05:11 INFO zookeeper.ZooKeeper: Initiating client connection,
> > connectString=namenode.com:2181,cloud-manager.com:2181
> &

Fwd: Hbase not taking inserts from Remote Machine

2015-02-16 Thread Vineet Mishra
-- Forwarded message --
From: Vineet Mishra 
Date: Tue, Feb 17, 2015 at 12:32 PM
Subject: Hbase not taking inserts from Remote Machine
To: cdh-u...@cloudera.org



Hi All,

I am trying to communicate and insert some data to my
Hbase(0.98.6-cdh5.3.0) running on Hadoop 2.5 using Hbase Java API.

Although if I run the code within the cluster it connects fine but If I am
outside cluster, even though I have opened the port on external IPs for
Zookeeper and HMaster, I am stuck without any error on the logs, so the
code hangs up on Table.put() during insertion,

Attached below is the Stack Trace for the Job,

15/02/17 12:05:08 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
15/02/17 12:05:08 INFO zookeeper.RecoverableZooKeeper: Process
identifier=hconnection-0x77dacebf connecting to ZooKeeper ensemble=
namenode.com:2181,cloud-manager.com:2181
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:zookeeper.version=3.4.5-cdh5.3.0--1, built on 12/17/2014 02:45
GMT
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:host.name
=ip-20-0-0-75
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.version=1.7.0_75
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.vendor=Oracle Corporation
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.75.x86_64/jre
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.class.path=.:hbase-connect-1.0.0.jar
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.io.tmpdir=/tmp
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:java.compiler=
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd64
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:os.version=3.14.23-22.44.amzn1.x86_64
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client environment:user.name=tom
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:user.home=/home/tom
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Client
environment:user.dir=/home/tom/jobs
15/02/17 12:05:08 INFO zookeeper.ZooKeeper: Initiating client connection,
connectString=namenode.com:2181,cloud-manager.com:2181 sessionTimeout=9
watcher=hconnection-0x77dacebf, quorum=namenode.com:2181,
cloud-manager.com:2181, baseZNode=/hbase
15/02/17 12:05:08 INFO zookeeper.ClientCnxn: Opening socket connection to
server namenode.com/54.172.21.54:2181. Will not attempt to authenticate
using SASL (unknown error)
15/02/17 12:05:09 INFO zookeeper.ClientCnxn: Socket connection established
to namenode.com/54.172.21.54:2181, initiating session
15/02/17 12:05:09 INFO zookeeper.ClientCnxn: Session establishment complete
on server namenode.com/54.172.21.54:2181, sessionid = 0x24b7c7ba3532b32,
negotiated timeout = 6
15/02/17 12:05:11 INFO client.HConnectionManager$HConnectionImplementation:
Closing master protocol: MasterService
15/02/17 12:05:11 INFO client.HConnectionManager$HConnectionImplementation:
Closing zookeeper sessionid=0x24b7c7ba3532b32
15/02/17 12:05:11 INFO zookeeper.ZooKeeper: Session: 0x24b7c7ba3532b32
closed
15/02/17 12:05:11 INFO zookeeper.ClientCnxn: EventThread shut down
Hbase Running
15/02/17 12:05:11 INFO zookeeper.RecoverableZooKeeper: Process
identifier=hconnection-0x1003cac6 connecting to ZooKeeper ensemble=
namenode.com:2181,cloud-manager.com:2181
15/02/17 12:05:11 INFO zookeeper.ZooKeeper: Initiating client connection,
connectString=namenode.com:2181,cloud-manager.com:2181 sessionTimeout=9
watcher=hconnection-0x1003cac6, quorum=namenode.com:2181,
cloud-manager.com:2181, baseZNode=/hbase
15/02/17 12:05:11 INFO zookeeper.ClientCnxn: Opening socket connection to
server namenode.com/54.172.21.54:2181. Will not attempt to authenticate
using SASL (unknown error)
15/02/17 12:05:12 INFO zookeeper.ClientCnxn: Socket connection established
to namenode.com/54.172.21.54:2181, initiating session
15/02/17 12:05:12 INFO zookeeper.ClientCnxn: Session establishment complete
on server namenode.com/54.172.21.54:2181, sessionid = 0x24b7c7ba3532b34,
negotiated timeout = 6

Looking at the revert urgently!

Thanks!


Re: copy from one cluster to another of different version

2014-11-28 Thread Vineet Mishra
You can set the block replication in your hdfs-site.xml configuration file
by specifying the property dfs.replication with replication value.

Moreover once the remote table dump is completed, you can run the
replication for your individual or group of table by,

hadoop fs –setrep –w [int REPLICATION_FACTOR] -R
/path/to/your/directory/or/file

Cheers!


On Fri, Nov 28, 2014 at 1:44 PM, Li Li  wrote:

> thank you. How can I specify replication factor with this command?
> will hbase  -Ddfs.replication=1
> org.apache.hadoop.hbase.mapreduce.CopyTable work?
>
> On Fri, Nov 28, 2014 at 3:32 PM, Vineet Mishra 
> wrote:
> > Hi Li Li,
> >
> > You can copy the Hbase Tables Remotely to another machine with the
> > following commands,
> >
> > # create new tableOrig on destination cluster
> > dstCluster$ echo "create 'tableOrig', 'cf1', 'cf2'" | hbase shell
> > # on source cluster run copy table with destination ZK quorum specified
> > using --peer.adr
> > # WARNING: In older versions, you are not alerted about any typo in these
> > arguments!
> > srcCluster$ hbase org.apache.hadoop.hbase.mapreduce.CopyTable --peer.adr=
> > dstClusterZK:2181:/hbase tableOrig
> >
> > More information given in the section Remote HBase instance backup at
> >
> >
> http://blog.cloudera.com/blog/2012/06/online-hbase-backups-with-copytable-2/
> >
> > Hope that works for you pretty straightforwardly.
> >
> > Cheers!
> >
> >
> > On Fri, Nov 28, 2014 at 12:44 PM, Li Li  wrote:
> >
> >> I have a hbase cluster of version 0.98.5 with hadoop-1.2.1(no mapreduce)
> >> I want to copy all the tables to another cluster whose version is
> >> 0.98.1-cdh5.1.0 with 2.3.0-cdh5.1.0.
> >> And also I want specify the hdfs replication factor of the files in
> >> new cluster. is it possible?
> >>
>


Re: Unable to run map reduce in HBase

2014-11-27 Thread Vineet Mishra
Hi,

It seems your job is not getting enough memory as required, can you try
adding these properties to your configuration and run schedule the job once
again.

mapred.cluster.map.memory.mb=2048
mapred.cluster.reduce.memory.mb=2048

Hope that solve your problem.

Cheers!

On Fri, Nov 28, 2014 at 11:33 AM,  wrote:

> Hi,
>
> I am importing a csv file into HBase using the command bin/hbase
> org.apache.hadoop.hbase.mapreduce.ImportTsv
>
> When I execute the this map reduce program I am getting the following
> error. I am using Hadoop 2.4.1 and HBase 0.98.8-hadoop2
>
> I have set export JAVA_OPTS="-Xms1024m -Xmx10240m" in .bashrc, the server
> has 32 GB of RAM.
>
> 2014-11-28 18:56:44,029 INFO [IPC Server listener on 56283]
> org.apache.hadoop.ipc.Server: IPC Server listener on 56283: starting
> 2014-11-28 18:56:44,031 FATAL [main]
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting MRAppMaster
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:693)
> at org.apache.hadoop.ipc.Server.start(Server.java:2392)
> at
>
> org.apache.hadoop.mapred.TaskAttemptListenerImpl.startRpcServer(TaskAttemptListenerImpl.java:137)
> at
>
> org.apache.hadoop.mapred.TaskAttemptListenerImpl.serviceStart(TaskAttemptListenerImpl.java:107)
> at
> org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
> at
>
> org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
> at
>
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.serviceStart(MRAppMaster.java:1071)
> at
> org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
> at
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster$1.run(MRAppMaster.java:1456)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
>
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1556)
> at
>
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.initAndStartAppMaster(MRAppMaster.java:1452)
> at
> org.apache.hadoop.mapreduce.v2.app.MRAppMaster.main(MRAppMaster.java:1385)
> 2014-11-28 18:56:44,034 INFO [main] org.apache.hadoop.util.ExitUtil:
> Exiting with status 1
>
> Any help is very much appreciated.
>
> Thanks & Regards
> Dhamodharan Ramalingam
> =-=-=
> Notice: The information contained in this e-mail
> message and/or attachments to it may contain
> confidential or privileged information. If you are
> not the intended recipient, any dissemination, use,
> review, distribution, printing or copying of the
> information contained in this e-mail message
> and/or attachments to it are strictly prohibited. If
> you have received this communication in error,
> please notify us by reply e-mail or telephone and
> immediately and permanently delete the message
> and any attachments. Thank you
>
>
>


Re: copy from one cluster to another of different version

2014-11-27 Thread Vineet Mishra
Hi Li Li,

You can copy the Hbase Tables Remotely to another machine with the
following commands,

# create new tableOrig on destination cluster
dstCluster$ echo "create 'tableOrig', 'cf1', 'cf2'" | hbase shell
# on source cluster run copy table with destination ZK quorum specified
using --peer.adr
# WARNING: In older versions, you are not alerted about any typo in these
arguments!
srcCluster$ hbase org.apache.hadoop.hbase.mapreduce.CopyTable --peer.adr=
dstClusterZK:2181:/hbase tableOrig

More information given in the section Remote HBase instance backup at

http://blog.cloudera.com/blog/2012/06/online-hbase-backups-with-copytable-2/

Hope that works for you pretty straightforwardly.

Cheers!


On Fri, Nov 28, 2014 at 12:44 PM, Li Li  wrote:

> I have a hbase cluster of version 0.98.5 with hadoop-1.2.1(no mapreduce)
> I want to copy all the tables to another cluster whose version is
> 0.98.1-cdh5.1.0 with 2.3.0-cdh5.1.0.
> And also I want specify the hdfs replication factor of the files in
> new cluster. is it possible?
>


Re: Merging Hbase Region for a Table

2013-12-02 Thread Vineet Mishra
Ok! So can you tell me in the Offline merge, how to put the cluster down?
Should we stop the HDFS or Region Server, and if so how to achieve that?


On Mon, Dec 2, 2013 at 7:34 PM, Jean-Marc Spaggiari  wrote:

> Hi Vineet.
>
> For 0.94 you can only offline-merge.
>
> http://hbase.apache.org/book/ops.regionmgt.html#ops.regionmgt.merge
>
> JM
>
>
> 2013/12/2 Vineet Mishra 
>
> > Hi
> >
> > I have some 2000+ of Auto-Region created which I want to make down to a
> > less number.
> > I am using Hbase 0.94, is there a way I can merge the Region without
> > loosing or dirtying up the data.
> >
> > Thanks!
> >
>


Re: Hbase Region Size

2013-12-02 Thread Vineet Mishra
Actually I am looking for the Size of the Region, and not for the whole
table. Although the Hbase internally do the Max file size check to split
the Region in a autonomous manner, hence there should be some way to get it.


On Mon, Dec 2, 2013 at 7:51 PM, Jean-Marc Spaggiari  wrote:

> Same for a single region. If it's compressed, you might want to look into
> HDFS directly...
>
>
> 2013/12/2 Mike Axiak 
>
> > Are you looking to get the MAX_FILESIZE paramter? If so, there's nothing
> in
> > the client, but HBaseAdmin has what you need [1].
> >
> >HTableDescriptor myDescriptor =
> > hbaseAdmin.getDescriptor(Bytes.toBytes("my-table"));
> >System.out.println("my-table has a max region size of " +
> > myDescriptor.getMaxFileSize());
> >
> >
> > 1:
> >
> >
> http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/HBaseAdmin.html
> >
> >
> > On Mon, Dec 2, 2013 at 9:05 AM, Jean-Marc Spaggiari <
> > jean-m...@spaggiari.org
> > > wrote:
> >
> > > Hi Vineet,
> > >
> > > If you want the entire table size I don't think there is any API for
> > that.
> > > If you want the size of the table on the disk (compressed) they you are
> > > better to use HDFS API.
> > >
> > > JM
> > >
> > >
> > > 2013/12/2 Vineet Mishra 
> > >
> > > > Hi
> > > >
> > > > Can Anyone tell me the Java API for getting the Region Size of a
> table!
> > > >
> > > > Thanks!
> > > >
> > >
> >
>


Merging Hbase Region for a Table

2013-12-02 Thread Vineet Mishra
Hi

I have some 2000+ of Auto-Region created which I want to make down to a
less number.
I am using Hbase 0.94, is there a way I can merge the Region without
loosing or dirtying up the data.

Thanks!


Hbase Region Size

2013-12-02 Thread Vineet Mishra
Hi

Can Anyone tell me the Java API for getting the Region Size of a table!

Thanks!


Re: Hbase Filter for RDBMS IN and AND query operation

2013-11-26 Thread Vineet Mishra
Yes,

Applying Filter List with Pass ALL containing two filters for name and flag
and iteratively adding it(previous filter list) to another filter list with
(Pass One).



On Tue, Nov 26, 2013 at 7:27 PM, Jean-Marc Spaggiari <
jean-m...@spaggiari.org> wrote:

> Hi vineet,
>
> So how have you solved that finally? Using filters?
>
> JM
>
>
> 2013/11/26 Vineet Mishra 
>
> > Thanks Marc,
> >
> > I got it working!
> >
> >
> > On Tue, Nov 26, 2013 at 6:49 PM, Jean-Marc Spaggiari <
> > jean-m...@spaggiari.org> wrote:
> >
> > > Hum.
> > >
> > > Phoenix guys will most probably reply that Phoenix is the solution for
> > that
> > > ;) And I will probably agree with them.
> > >
> > > I don't see any filter which can do that within a single call.
> > >
> > > One option I will apply in your case is to do a scan+filter on the
> flag=1
> > > value then do a get name based on the return keys. But that might be to
> > > many gets if you have a lot of values returned.
> > >
> > > Based on your usecase and the size of the name column, another one is
> to
> > > build a MR job. In the map method you keep track of the name and flag
> > > columns, and if you get the 2 before going to the new row, you emit it
> to
> > > another table.
> > >
> > > HTH.
> > >
> > > JM
> > >
> > >
> > > 2013/11/26 Vineet Mishra 
> > >
> > > > Hi All,
> > > >
> > > > I want to fetch some rows from a table where the value for the
> > > > Qualifier(say name) is any of ("joseph" ,"sandy", "lar" ) and another
> > > > Qualifier(say flag) should always be 1.
> > > >
> > > > I am making multi SingleColumnValueFilter scan on a Table with
> Compare
> > > > Operator to
> > > > Equal and those list of values for each Filter.
> > > > Applying multi SingleColumnValueFilter is going good, only till I am
> > not
> > > > including flag, and without flag my result are incomplete.
> > > >
> > > > Can you tell me which filter or approach to use to get my expected
> > > result.
> > > >
> > > > Thanks!
> > > >
> > >
> >
>


Re: Hbase Filter for RDBMS IN and AND query operation

2013-11-26 Thread Vineet Mishra
Thanks Marc,

I got it working!


On Tue, Nov 26, 2013 at 6:49 PM, Jean-Marc Spaggiari <
jean-m...@spaggiari.org> wrote:

> Hum.
>
> Phoenix guys will most probably reply that Phoenix is the solution for that
> ;) And I will probably agree with them.
>
> I don't see any filter which can do that within a single call.
>
> One option I will apply in your case is to do a scan+filter on the flag=1
> value then do a get name based on the return keys. But that might be to
> many gets if you have a lot of values returned.
>
> Based on your usecase and the size of the name column, another one is to
> build a MR job. In the map method you keep track of the name and flag
> columns, and if you get the 2 before going to the new row, you emit it to
> another table.
>
> HTH.
>
> JM
>
>
> 2013/11/26 Vineet Mishra 
>
> > Hi All,
> >
> > I want to fetch some rows from a table where the value for the
> > Qualifier(say name) is any of ("joseph" ,"sandy", "lar" ) and another
> > Qualifier(say flag) should always be 1.
> >
> > I am making multi SingleColumnValueFilter scan on a Table with Compare
> > Operator to
> > Equal and those list of values for each Filter.
> > Applying multi SingleColumnValueFilter is going good, only till I am not
> > including flag, and without flag my result are incomplete.
> >
> > Can you tell me which filter or approach to use to get my expected
> result.
> >
> > Thanks!
> >
>


Hbase Filter for RDBMS IN and AND query operation

2013-11-26 Thread Vineet Mishra
Hi All,

I want to fetch some rows from a table where the value for the
Qualifier(say name) is any of ("joseph" ,"sandy", "lar" ) and another
Qualifier(say flag) should always be 1.

I am making multi SingleColumnValueFilter scan on a Table with Compare
Operator to
Equal and those list of values for each Filter.
Applying multi SingleColumnValueFilter is going good, only till I am not
including flag, and without flag my result are incomplete.

Can you tell me which filter or approach to use to get my expected result.

Thanks!