Have you asked this question on MapR mailing list ?

Cheers

On Jun 18, 2014, at 12:14 AM, Chen Wang <chen.apache.s...@gmail.com> wrote:

> I actually tried that already, but it didn't work..I added
> 
> <dependency>
> 
> <groupId>org.apache.hbase</groupId>
> 
> <artifactId>hbase</artifactId>
> 
> <version>0.94.9-mapr-1308</version>
> 
> </dependency>
> 
> and removed the original hbase dependency..
> 
> 
> On Wed, Jun 18, 2014 at 12:05 AM, Rabbit's Foot <rabbitsf...@is-land.com.tw>
> wrote:
> 
>> Maybe you can refer the Maven Repository and Artifacts for MapR
>> <http://doc.mapr.com/display/MapR/Maven+Repository+and+Artifacts+for+MapR>
>> to
>> set pom
>> 
>> 
>> 2014-06-18 13:33 GMT+08:00 Chen Wang <chen.apache.s...@gmail.com>:
>> 
>>> Is this error indicating that I basically need a hbase mapr client?
>>> currently my pom looks like this;
>>> 
>>> <dependency>
>>> 
>>> <groupId>org.apache.hadoop</groupId>
>>> 
>>> <artifactId>hadoop-client</artifactId>
>>> 
>>> <version>1.0.3</version>
>>> 
>>> </dependency>
>>> 
>>> <dependency>
>>> 
>>> <groupId>org.apache.hadoop</groupId>
>>> 
>>> <artifactId>hadoop-core</artifactId>
>>> 
>>> <version>1.2.1</version>
>>> 
>>> </dependency>
>>> 
>>> <dependency>
>>> 
>>> <groupId>org.apache.httpcomponents</groupId>
>>> 
>>> <artifactId>httpclient</artifactId>
>>> 
>>> <version>4.1.1</version>
>>> 
>>> </dependency>
>>> 
>>> <dependency>
>>> 
>>> <groupId>com.google.code.gson</groupId>
>>> 
>>> <artifactId>gson</artifactId>
>>> 
>>> <version>2.2.4</version>
>>> 
>>> </dependency>
>>> 
>>> 
>>> <dependency>
>>> 
>>> <groupId>org.apache.hbase</groupId>
>>> 
>>> <artifactId>hbase</artifactId>
>>> 
>>> <version>0.94.6.1</version>
>>> 
>>> </dependency>
>>> 
>>> 
>>> On Tue, Jun 17, 2014 at 10:04 PM, Chen Wang <chen.apache.s...@gmail.com>
>>> wrote:
>>> 
>>>> Yes, the hadoop cluster is using maprfs, so the hdfs files are are in
>>>> maprfs:/ format:
>>>> 
>>>> 
>>>> 2014-06-17 21:48:58 WARN:
>>>> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles - Skipping
>>>> non-directory maprfs:/user/chen/hbase/_SUCCESS
>>>> 2014-06-17 21:48:58 INFO: org.apache.hadoop.hbase.io.hfile.CacheConfig
>> -
>>>> Allocating LruBlockCache with maximum size 239.6m
>>>> 2014-06-17 21:48:58 INFO: org.apache.hadoop.hbase.util.ChecksumType -
>>>> Checksum using org.apache.hadoop.util.PureJavaCrc32
>>>> 2014-06-17 21:48:58 INFO:
>>>> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles - Trying to
>> load
>>>> hfile=maprfs:/user/chen/hbase/m/cdd83ff3007b4955869d69c82a9f5b91
>>> first=row1
>>>> last=row9
>>>> 
>>>> Chen
>>>> 
>>>> On Tue, Jun 17, 2014 at 9:59 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>> 
>>>>> The scheme says maprfs.
>>>>> Do you happen to use MapR product ?
>>>>> 
>>>>> Cheers
>>>>> 
>>>>> On Jun 17, 2014, at 9:53 PM, Chen Wang <chen.apache.s...@gmail.com>
>>>>> wrote:
>>>>> 
>>>>>> Folk,
>>>>>> I am trying to bulk load the hdfs file into hbase with
>>>>>> 
>>>>>> LoadIncrementalHFiles loader = new LoadIncrementalHFiles(conf);
>>>>>> 
>>>>>> loader.doBulkLoad(new Path(args[1]), hTable);
>>>>>> 
>>>>>> 
>>>>>> However, i receive exception of java.io.IOException:
>>>>> java.io.IOException:
>>>>>> No FileSystem for scheme: maprfs
>>>>>> 
>>>>>> Exception in thread "main" java.io.IOException: BulkLoad encountered
>>> an
>>>>>> unrecoverable problem
>>>>>> 
>>>>>> at
>> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:331)
>>>>>> 
>>>>>> at
>> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:261)
>>>>>> 
>>>>>> at com.walmartlabs.targeting.mapred.Driver.main(Driver.java:81)
>>>>>> 
>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>> 
>>>>>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>>>>> 
>>>>>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>>>> 
>>>>>> at java.lang.reflect.Method.invoke(Method.java:597)
>>>>>> 
>>>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:197)
>>>>>> 
>>>>>> Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException:
>>>>> Failed
>>>>>> after attempts=10, exceptions:
>>>>>> 
>>>>>> Tue Jun 17 21:48:58 PDT 2014,
>>>>>> org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$3@482d59a3,
>>>>>> java.io.IOException: java.io.IOException: No FileSystem for scheme:
>>>>> maprfs
>>>>>> 
>>>>>> 
>>>>>> What is the reason for this exception? I did some googling, and
>> tried
>>> to
>>>>>> add some config to Hbase configuration:
>>>>>> 
>>>>>> hbaseConf.set("fs.hdfs.impl",
>>>>>> 
>>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.class.getName());
>>>>>> 
>>>>>> hbaseConf.set("fs.file.impl",
>>>>>> 
>>>>>> org.apache.hadoop.fs.LocalFileSystem.class.getName());
>>>>>> 
>>>>>> 
>>>>>> But it does not have any effect.
>>>>>> 
>>>>>> Any idea?
>>>>>> 
>>>>>> Thanks advance.
>>>>>> 
>>>>>> Chen
>> 

Reply via email to