p.s. Please use a HadoopFileSystemFactory in secondary file system config,
as described there
https://apacheignite-fs.readme.io/docs/installing-on-hortonworks-hdp ,
constructor org.apache.ignite.hadoop.fs.IgniteHadoopIgfsSecondaryFileSystem(2)
is deprecated.

On Mon, Apr 24, 2017 at 7:44 PM, Ivan V. <iveselovs...@gridgain.com> wrote:

> Hi, Aloha,
> First of all, Hadoop Accelerator consists of 2 parts that are independent
> and can be used one without the other: (1) IGFS and (2) map-reduce
> execution engine.
>
> IGFS is not used in your case because default file system in your cluster
> is still hdfs://.... (specified by global property "fs.default.name").
> The 2 properties you set (*.igfs.impl=..) define the IGFS implementation
> classes, but they come into play only when igfs:// schema encounters.
> Idea to set fs.default.name=igfs://myhost:10500/ is not so good as it may
> appear, because HDFS daemons (namenode, datanode) cannot run with such
> property value, while you probably need HDFS as the underlying (secondary)
> file system.
>
> So, to use IGFS you should either use explicit URI with igfs:// schema as
> you do in your example above "hadoop fs -ls igfs:///user/hive", or try to
> instruct Hive to use igfs as default property, like this:
> hive-1.2/bin/beeline \
> --hiveconf fs.default.name=igfs://myhost:10500/ \
> --hiveconf hive.rpc.query.plan=true \
> --hiveconf mapreduce.framework.name=ignite \
> --hiveconf mapreduce.jobtracker.address=myhost:11211 -u jdbc:hive2://
> 127.0.0.1:10000
>
> Also , in order to use Ignite Map-Reduce engine with Hive,  in HDP 2.4+
> the Hive execution engine (property "hive.execution.engine") should
> explicitly be set to "mr", because the default value is different.
>
> On Mon, Apr 24, 2017 at 3:09 PM, <al...@74.ru> wrote:
>
>> Hi,
>>
>> I have a cluster HDP 2.6 (High Available, 8 nodes) and like to try using
>> Hive+Orc+Tez with Ignite. I guess I should use IFGS as cache layer for HDFS.
>> I installed Hadoop Accelerator  1.9 on all cluster nodes and run one
>> ignite-node on every cluster node.
>>
>> I added these settings using Ambari  and then restarted HDFS, MapReduce,
>> Yarn, Hive.
>> HDFS, add 2 new properties to Custom core-site
>> fs.igfs.impl=org.apache.ignite.hadoop.fs.v1.IgniteHadoopFileSystem
>> fs.AbstractFileSystem.igfs.impl=org.apache.ignite.hadoop.fs.
>> v2.IgniteHadoopFileSystem
>>
>> Mapred, Custom mapred-site
>> mapreduce.framework.name=ignite
>> mapreduce.jobtracker.address=dev-nn1:11211
>>
>> Hive, Custom hive-site
>> hive.rpc.query.plan=true
>>
>> Now I can get access to HDFS through IGFS
>> hadoop fs -ls igfs:///user/hive
>> Found 3 items
>> drwx------  - hive hdfs          0 2017-04-19 21:00
>> igfs:///user/hive/.Trash
>> drwxr-xr-x  - hive hdfs          0 2017-04-19 10:07
>> igfs:///user/hive/.hiveJars
>> drwx------  - hive hdfs          0 2017-04-22 14:27
>> igfs:///user/hive/.staging
>>
>> I thought that Hive read data from HDFS first time and then read the same
>> data from IFGS.
>> But when I run HIVE (cli or beeline) it still reads data from HDFS (I
>> tried a few times), in igniteVisor "Avg. free heap" remains the same
>> before/during/after running query (about 80%).
>> What is wrong? May be I should load data to IFGS manually for every query?
>
>
>

Reply via email to