Any update for this issue HBASE-3529

2013-05-30 Thread dong.yajun
Hello list,

Any one can give me some follow up information about this issue HBASE-3529,
 I'm wondering it has more than 2 years no update.

Best,
-- 
*Rick Dong*


Re: Any update for this issue HBASE-3529

2013-05-30 Thread dong.yajun
hi Ted,

not yet right now about lily.

I would like to use hbase to store product reviews,  the system should
support secondary index, full text search and faceting(Lucene) which
support paging and sorting.


On Fri, May 31, 2013 at 10:24 AM, Ted Yu  wrote:

> Jason is no longer working on this issue.
>
> Can you tell us your use case ?
>
> Have you looked at http://www.lilyproject.org/lily/index.html ?
>
> Thanks
>
> On Thu, May 30, 2013 at 7:06 PM, dong.yajun  wrote:
>
> > Hello list,
> >
> > Any one can give me some follow up information about this issue
> HBASE-3529,
> >  I'm wondering it has more than 2 years no update.
> >
> > Best,
> > --
> > *Rick Dong*
> >
>



-- 
*Ric Dong*


(info) How can i load data from CDH4.3.0 to CDH5.4.0 in Hbase

2015-05-18 Thread dong.yajun
hello list,

is there a way to load the existing data(HFiles) from CDH4.3.0 to CDH5.4.0?

we use the complete bulkload utility which reference the link:
http://hbase.apache.org/0.94/book/ops_mgt.html#completebulkload

the command: hbase org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles
/tmp/IM_ItemPrice/096518a1aa5823c4aec9477d7b1b63cf/ IM_ItemPrice

the region of *096518a1aa5823c4aec9477d7b1b63cf *which contains several
family names: BaseInfo / of / ol / q4s etc.

but it seems not work, see the output after command typed below:

15/05/19 09:57:59 INFO mapreduce.LoadIncrementalHFiles: Trying to load
hfile=hdfs://nameservice1/tmp/IM_ItemPrice/096518a1aa5823c4aec9477d7b1b63cf/
*BaseInfo*/88fc1240fa8f4e31aa27469b7bd66750
first=e04116151155a5fae8dfa37281df89304c5e62219c31a761024cfac80f8e204c
last=baa9f23ae8fd718aa888f814665e19d04fcdee09de9926c8690db00af905
15/05/19 09:57:59 INFO mapreduce.LoadIncrementalHFiles: Trying to load
hfile=hdfs://nameservice1/tmp/IM_ItemPrice/096518a1aa5823c4aec9477d7b1b63cf/
*ol*/2a886066311343f98737ad2e4e804260
first=e045a94bb684ce8bbb5cf34b9e0dd939c03946bd445711204b30c17d72b55874
last=7f5766b39d218e772dc357d9588b7a7363c03b3b7a07be7bcbb41dc267b3
15/05/19 09:57:59 INFO mapreduce.LoadIncrementalHFiles: Trying to load
hfile=hdfs://nameservice1/tmp/IM_ItemPrice/096518a1aa5823c4aec9477d7b1b63cf/
*of*/c01e895d483b4c86beb4eeae503e8fa9
first=e046037883152a81168507bf9faefc8ba716f15a6a028b81cbf83e2894896ec3
last=fff93051701920c2848f091369e484b685f42ed7a1378ffcb4e9dddf8bcd7ef7

15/05/19 10:09:17 INFO client.RpcRetryingCaller: Call exception, tries=10,
retries=35, started=677650 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
15/05/19 10:10:32 INFO client.RpcRetryingCaller: Call exception, tries=11,
retries=35, started=752931 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
15/05/19 10:11:48 INFO client.RpcRetryingCaller: Call exception, tries=12,
retries=35, started=828151 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
15/05/19 10:13:03 INFO client.RpcRetryingCaller: Call exception, tries=13,
retries=35, started=903409 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
15/05/19 10:14:18 INFO client.RpcRetryingCaller: Call exception, tries=14,
retries=35, started=978634 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
15/05/19 10:15:33 INFO client.RpcRetryingCaller: Call exception, tries=15,
retries=35, started=1054003 ms ago, cancelled=false, msg=row '' on table
'IM_ItemPrice' at
region=IM_ItemPrice,,1432000677144.36c13c3160de2e67e4fdb1d77c3c9ade.,
hostname=ssspark03,60020,1431768671791, seqNum=2
..


any suggestion?

-- 
*Ric Dong*


(req) unsubscribe

2013-03-10 Thread dong.yajun
list, I'd like to unsubscribe this mail list.

thanks.


-- 
*Ric Dong *


Re: Region autoSplit when not reach 'hbase.hregion.max.filesize' ?

2012-06-06 Thread dong.yajun
Hi NNever

If you find any issues, please let us known, thanks.


On Wed, Jun 6, 2012 at 5:09 PM, NNever  wrote:

> I'm sorry, the log4j now is WARN, not INFO
>
> 2012/6/6 NNever 
>
> > We currently run in INFO mode.
> > It actully did the split, but I cannot find any logs about this split.
> > I will change the log4j to DEBUG, if got any log valuable, I will paste
> > here...
> >
> > Thanks Ram,
> > NN
> >
> > 2012/6/6 Ramkrishna.S.Vasudevan 
> >
> > You have any logs corresponding to this?
> >>
> >> Regards
> >> Ram
> >>
> >> > -Original Message-
> >> > From: NNever [mailto:nnever...@gmail.com]
> >> > Sent: Wednesday, June 06, 2012 2:12 PM
> >> > To: user@hbase.apache.org
> >> > Subject: Region autoSplit when not reach 'hbase.hregion.max.filesize'
> ?
> >> >
> >> > The 'hbase.hregion.max.filesize' are set to 100G (The recommed value
> to
> >> > act
> >> > as auto-split turn off). And there is a table, we keep put datas into
> >> > it.
> >> > When the storefileUncompressedSizeMB reached about 1Gb, the region
> auto
> >> > splite to 2.
> >> > I don't know how it happened? 1G is far more less than max.filesize-
> >> > 100G.
> >> > So if there is any possible scenery that will
> >> > ignore hbase.hregion.max.filesize and do split?
> >> >
> >> > How can I totally shutdown the autoSplit?
> >> >
> >> >
> >> > -
> >> > Best regards,
> >> > NN
> >>
> >>
> >
>



-- 
*Ric Dong *
Newegg Ecommerce, MIS department


Re: Full text search on HBase

2012-06-07 Thread dong.yajun
Hi Jack

you can use solr over hbase, solr stores the index data, and hbase stores
the actual data .


Thanks
Rick

On Wed, Jun 6, 2012 at 9:56 PM, Otis Gospodnetic  wrote:

> https://issues.apache.org/jira/browse/HBASE-3529
>
>
> Otis
> 
> Performance Monitoring for Solr / ElasticSearch / HBase -
> http://sematext.com/spm
>
>
>
> >
> > From: Jacques 
> >To: user@hbase.apache.org
> >Sent: Tuesday, June 5, 2012 6:27 PM
> >Subject: Re: Full text search on HBase
> >
> >You should check out the solbase deck by photo bucket from hbasecon.  Also
> >the work that jason ruthelgren did... Don't have jira offhand
> >On Jun 5, 2012 2:40 PM, "jack chrispoo"  wrote:
> >
> >> Hi everyone,
> >>
> >> Does anyone have experience with full text search on HBase? I was
> reading
> >> about hbasene, but last update was 2 years ago. I also read about lily,
> I
> >> was planning to try out lily. But before I try, it would be really nice
> if
> >> anyone can tell me what other options there are to do full text search
> on
> >> HBase, and if anyone has experience with lily please let me know how it
> >> performs.
> >>
> >> Thanks
> >> jack
> >>
> >
> >
> >
>



-- 
*Ric Dong *
Newegg Ecommerce, MIS department


Which is the fastest way to dump the content of Hbase table?

2015-08-19 Thread dong.yajun
Hello,

Which is the fastest way to dump  the content of Hbase table to Hdfs?  is
it possible to use the hbase snapshot + Spark to do this?

now we have already use the hbase snapshot + mapreduce-v2(does not via the
Htable) to convert the HFiles to OrcFile, but we found the 'spilling map
output' occupied most of whole time.  so the spark can decrease the cost?

map task: read the hfile, and convert it to KeyValues

reduce task: merge the keyvalues of same rowkey

thanks.

-- 
*Ric Dong*