[ 
https://issues.apache.org/jira/browse/KYLIN-1515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15252244#comment-15252244
 ] 

Rachel Asher Silver commented on KYLIN-1515:
--------------------------------------------

[~cal...@gmail.com] Yes, this workaround gets me past that error although I am 
never able to build anything. I'd love to talk to you about your experience if 
you could spare a moment: rsil...@mapr.com

[~Shaofengshi] Pulling the kylin_hive_conf.xml file doesn't allow you to 
progress to the build as you just get an "Oops, fail to locate 
kylin_hive_conf.xml" message from the GUI. How can we resolve this without the 
workaround that Richard discovered?  

> Cube Build - java.io.IOException: Cannot initialize Cluster. Please check 
> your configuration for mapreduce.framework.name and the correspond server 
> addresses
> -------------------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: KYLIN-1515
>                 URL: https://issues.apache.org/jira/browse/KYLIN-1515
>             Project: Kylin
>          Issue Type: Bug
>          Components: Job Engine
>    Affects Versions: v1.5.0
>         Environment: MapR - Hadoop 2.5.1
>            Reporter: Richard Calaba
>            Assignee: Shaofeng SHI
>
> Knowing that MapR is not officially supported we were able to use Kylin 1.2 
> in our MapR distro successfully. 
> After upgrade to Kylin 1.5.0 we are facing issue with the Cube Build process 
> - the one which worked on 1.2 without issues. The Cube is created from 
> scratch (no Kylin metadata migration) on clean install of Kylinn1.5.0 (HDFS 
> directory /kytlin and HBase tables KYLIN* and kylin* deleted prior upgrade 
> from 1.2 to 1.5.0).
> The build process is Failing in Step 1 complaining about property value 
> "mapreduce.framework.name". According to this post 
> https://stackoverflow.com/questions/19642862/cannot-initialize-cluster-exception-while-running-job-on-hadoop-2
>  - the solution should be to ensure the respective property is correctly set 
> in the file mapred-site.xml.
> Originally in our MapR distro the property was commented (and having value 
> yarn-tez) - even after adding the "yarn" value -> the Build process still 
> fails with same exception - I am not sure what is wrong with our cluster 
> configuration.  Anyone has an idea ???
> Below is our mapred-site.xml content:
> ==============================
> cat /opt/mapr/hadoop/hadoop-2.5.1/etc/hadoop/mapred-site.xml
> <!-- Put site-specific property overrides in this file. -->
> <configuration>
>   <property>
>     <name>mapreduce.jobhistory.address</name>
>     <value>node1:10020</value>
>   </property>
>   <property>
>     <name>mapreduce.jobhistory.webapp.address</name>
>     <value>node1:19888</value>
>   </property>
>   <!--
>   <property>
>     <name>mapreduce.framework.name</name>
>     <value>yarn-tez</value>
>   </property>
>   -->
>   <property>
>     <name>mapreduce.framework.name</name>
>     <value>yarn</value>
>   </property>
> </configuration>
> Known workaround:
> ================
> Know workaround to make this error to disappear is to delete from 
> conf/kylin_hive_conf.xml this property section:
> <property>
> <name>dfs.block.size</name>
> <value>32000000</value>
> <description>Want more mappers for in-mem cubing, thus smaller the DFS block 
> size</description>
> </property>
> The full log output of Cube Build Step 1 - attached below: 
> ==============================================
> OS command error exit with 1 -- hive -e "USE default;
> DROP TABLE IF EXISTS 
> kylin_intermediate_TestCube_clone2_19700101000000_2922789940817071255;
> CREATE EXTERNAL TABLE IF NOT EXISTS 
> kylin_intermediate_TestCube_clone2_19700101000000_2922789940817071255
> (
> DEFAULT_BATTING_PLAYER_ID string
> ,DEFAULT_BATTING_YEAR int
> ,DEFAULT_BATTING_RUNS int
> )
> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\177'
> STORED AS SEQUENCEFILE
> LOCATION 
> '/kylin/kylin_metadata/kylin-3eb4b652-a2a4-4659-8b6a-dc822e1341fb/kylin_intermediate_TestCube_clone2_19700101000000_2922789940817071255';
> SET dfs.replication=2;
> SET dfs.block.size=32000000;
> SET hive.exec.compress.output=true;
> SET hive.auto.convert.join.noconditionaltask=true;
> SET hive.auto.convert.join.noconditionaltask.size=300000000;
> SET 
> mapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.SnappyCodec;
> SET 
> mapreduce.output.fileoutputformat.compress.codec=org.apache.hadoop.io.compress.SnappyCodec;
> SET hive.merge.mapfiles=true;
> SET hive.merge.mapredfiles=true;
> SET mapred.output.compression.type=BLOCK;
> SET hive.merge.size.per.task=256000000;
> SET hive.support.concurrency=false;
> SET mapreduce.job.split.metainfo.maxsize=-1;
> INSERT OVERWRITE TABLE 
> kylin_intermediate_TestCube_clone2_19700101000000_2922789940817071255 SELECT
> BATTING.PLAYER_ID
> ,BATTING.YEAR
> ,BATTING.RUNS
> FROM DEFAULT.BATTING as BATTING 
> LEFT JOIN DEFAULT.TEMP_BATTING as TEMP_BATTING
> ON BATTING.PLAYER_ID = TEMP_BATTING.COL_VALUE
> ;
> "
> Logging initialized using configuration in 
> jar:file:/opt/mapr/hive/hive-1.0/lib/hive-common-1.0.0-mapr-1510.jar!/hive-log4j.properties
> OK
> Time taken: 0.611 seconds
> OK
> Time taken: 0.83 seconds
> OK
> Time taken: 0.474 seconds
> Query ID = mapr_20160321201212_610078b4-5805-43eb-8fd1-87304530a84e
> Total jobs = 3
> 2016-03-21 08:12:32   Starting to launch local task to process map join;      
> maximum memory = 477102080
> 2016-03-21 08:12:32   Dump the side-table for tag: 1 with group count: 95196 
> into file: 
> file:/tmp/mapr/b35c5ac2-3231-4ef1-9e6b-216c0a1bd9ef/hive_2016-03-21_20-12-31_085_8296009472449837835-1/-local-10003/HashTable-Stage-9/MapJoin-mapfile01--.hashtable
> 2016-03-21 08:12:32   Uploaded 1 File to: 
> file:/tmp/mapr/b35c5ac2-3231-4ef1-9e6b-216c0a1bd9ef/hive_2016-03-21_20-12-31_085_8296009472449837835-1/-local-10003/HashTable-Stage-9/MapJoin-mapfile01--.hashtable
>  (7961069 bytes)
> 2016-03-21 08:12:32   End of local task; Time Taken: 0.853 sec.
> Launching Job 1 out of 3
> Number of reduce tasks is set to 0 since there's no reduce operator
> java.io.IOException: Cannot initialize Cluster. Please check your 
> configuration for mapreduce.framework.name and the correspond server 
> addresses.
>       at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:121)
>       at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:83)
>       at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:76)
>       at org.apache.hadoop.mapred.JobClient.init(JobClient.java:470)
>       at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:449)
>       at 
> org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:399)
>       at 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:137)
>       at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
>       at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
>       at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1619)
>       at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1379)
>       at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1192)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1019)
>       at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1009)
>       at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
>       at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
>       at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:299)
>       at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:662)
>       at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
>       at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
> Job Submission failed with exception 'java.io.IOException(Cannot initialize 
> Cluster. Please check your configuration for mapreduce.framework.name and the 
> correspond server addresses.)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to