Can someone give me a link to sample hive-site.xml which is working with
kylin?
I feel I need to set some mandatory parameter which hive shell can get but
need to be specified if being accessed from external client? Do I need to
set megastore db path?
On Fri, 20 Feb 2015 at 9:12 pm, Santosh Akhilesh <[email protected]>
wrote:

> You replied to wrong thread. Didn't want to hijack that thread.
> My hive Conf is under hive install. Following is the hive dependency I
> export in startkylin.sh
>
> export
> hive_dependency=/home/santosh/work/frameworks/apache-hive-1.0.0/conf/:/home/santosh/work/frameworks/apache-hive-1.0.0/lib/*:/home/santosh/work/frameworks/apache-hive-1.0.0/hcatalog/share/hcatalog/*
> On Fri, 20 Feb 2015 at 8:43 pm, Santosh Akhilesh <
> [email protected]> wrote:
>
>> Hi ,
>>    I have set the conf path in start_kylin.sh as below. I have removed
>> the sourcing of find-hive-dependency.sh as its not working in my env.
>>
>> In the hive-site.xml , I have set following; still i get same error , is
>> there anything else I need to do ?
>>
>> <configuration>
>>   <property>
>>     <name>hive.exec.scratchdir</name>
>>     <value>/tmp/hive</value>
>>     <description>Scratch space for Hive jobs</description>
>>   </property>
>> </configuration>
>>
>>
>> #source ${dir}/find-hive-dependency.sh
>> export hive_dependency=/home/santosh/work/frameworks/apache-hive-1.
>> 0.0/conf/:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> lib/*:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> hcatalog/share/hcatalog/*
>> export HBASE_CLASSPATH_PREFIX=${tomcat_root}/bin/bootstrap.
>> jar:${tomcat_root}/bin/tomcat-juli.jar:${tomcat_root}/lib/*:
>> $HBASE_CLASSPATH_PREFIX
>> export HBASE_CLASSPATH=$hive_dependency:${HBASE_CLASSPATH}
>>
>> hbase -Djava.util.logging.config.file=${tomcat_root}/conf/logging.properties
>> \
>> -Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager \
>> -Dorg.apache.tomcat.util.buf.UDecoder.ALLOW_ENCODED_SLASH=true \
>> -Dorg.apache.catalina.connector.CoyoteAdapter.ALLOW_BACKSLASH=true \
>> -Djava.endorsed.dirs=${tomcat_root}/endorsed  \
>> -Dcatalina.base=${tomcat_root} \
>> -Dcatalina.home=${tomcat_root} \
>> -Djava.io.tmpdir=${tomcat_root}/temp  \
>> -Dkylin.hive.dependency=${hive_dependency} \
>> -Dspring.profiles.active=${spring_profile} \
>> org.apache.hadoop.util.RunJar ${tomcat_root}/bin/bootstrap.jar
>>  org.apache.catalina.startup.Bootstrap start >
>> ${tomcat_root}/logs/kylin.log 2>&1 &
>> echo "A new Kylin instance is started by $USER, stop it using
>> \"stop-kylin.sh\""
>> if [ "$useSandbox" = "true" ]
>>     then echo "Please visit http://<your_sandbox_ip>:7070/kylin to play
>> with the cubes! (Useranme: ADMIN, Password: KYLIN)"
>> else
>>     echo "Please visit http://<ip>:7070/kylin"
>> fi
>> echo "You can check the log at ${tomcat_root}/logs/kylin.log"
>>
>>
>> On Fri, Feb 20, 2015 at 8:10 PM, Shi, Shaofeng <[email protected]> wrote:
>>
>>> Hi,
>>>
>>> Kylin from 7.0.x and later needs hive-site.xml on the classpth, to allow
>>> Kylin connect with the hive metadata service via API, to get the table
>>> info; As you can execute hive commands on your machine, the hive client
>>> should have been properly set, you can search the hive-site.xml from / to
>>> get its path, and then add it to the script as Qianhao mentioned.
>>>
>>> On 2/20/15, 10:28 PM, "Santosh Akhilesh" <[email protected]>
>>> wrote:
>>>
>>> >Hi Qianhao,
>>> >        I really wonder what kind of config I need to set for making
>>> sure
>>> >kylin can load the hive tables.
>>> >        I am not using the ready made sandbox , I have downloaded hbase
>>> ,
>>> >hive , hadoop from apcahe site to make the env. As far as I can see
>>> >currently I can query hive tables in hive shell. Mine is single node
>>> >pseudo
>>> >distributed cluster. I have not set any changes to hive-site.xml , the
>>> >metastoredb is created under $HIVE_INSTALL/bin
>>> >data is on hdfs under /user/hive/warehouse.
>>> >
>>> >I am really out of clue why the exception , I suspect its some setting
>>> >which kylin expects to make hive query run
>>> >
>>> >Please help me to get over this issue , thanks.
>>> >2015-02-20 19:38:10,435 INFO  [http-bio-7070-exec-3]
>>> HiveMetaStore.audit:
>>> >ugi=santosh ip=unknown-ip-addr cmd=get_table : db=DEFAULT tbl=books
>>> >NoSuchObjectException(message:DEFAULT.books table not found)
>>> >at
>>> >org.apache.hadoop.hive.metastore.HiveMetaStore$
>>> HMSHandler.get_table_core(H
>>> >iveMetaStore.java:1729)
>>> >at
>>> >org.apache.hadoop.hive.metastore.HiveMetaStore$
>>> HMSHandler.get_table(HiveMe
>>> >taStore.java:1699)
>>> >at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> >at
>>> >sun.reflect.NativeMethodAccessorImpl.invoke(
>>> NativeMethodAccessorImpl.java:
>>> >57)
>>> >at
>>> >sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>> DelegatingMethodAccessorIm
>>> >pl.java:43)
>>> >at java.lang.reflect.Method.invoke(Method.java:606)
>>> >at
>>> >org.apache.hadoop.hive.metastore.RetryingHMSHandler.
>>> invoke(RetryingHMSHand
>>> >ler.java:102)
>>> >at com.sun.proxy.$Proxy37.get_table(Unknown Source)
>>> >at
>>> >org.apache.hadoop.hive.metastore.HiveMetaStoreClient.
>>> getTable(HiveMetaStor
>>> >eClient.java:1097)
>>> >at
>>> >org.apache.kylin.common.util.HiveClient.getHiveTable(
>>> HiveClient.java:115)
>>> >at
>>> >org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>>> extractHiveTables(Hiv
>>> >eSourceTableLoader.java:86)
>>> >
>>> >On Fri, Feb 20, 2015 at 4:53 PM, Santosh Akhilesh
>>> ><[email protected]
>>> >> wrote:
>>> >
>>> >> Hi Qianhao,
>>> >>     *I have installed hive from Apache hive site. I dont have
>>> following
>>> >> folder. *
>>> >> /usr/hdp/current/hive-client/conf/ in this script.
>>> >> *I am using Hive 1.0.0 version.*
>>> >> *Where can I get this ? and which conf files I need to set under conf
>>> >> folder ?*
>>> >> *Regards,*
>>> >> *Santosh Akhilesh*
>>> >>
>>> >>
>>> >>
>>> >>
>>> >> On Fri, Feb 20, 2015 at 10:51 AM, 周千昊 <[email protected]> wrote:
>>> >>
>>> >>> Hi, Santosh
>>> >>>      Thanks for reporting the issue, it looks to me that it is still
>>> >>>the
>>> >>> problem of find-hive-dependecny.sh.
>>> >>>      As a workaround, you can just manually add
>>> >>> /usr/hdp/current/hive-client/conf/ in this script.
>>> >>>
>>> >>> On Thu Feb 19 2015 at 8:01:39 PM Santoshakhilesh <
>>> >>> [email protected]> wrote:
>>> >>>
>>> >>> > Hi ,
>>> >>> >     I analyzed the Kylin.log and found that find-hive-dependency.sh
>>> >>>was
>>> >>> > not able to locate hive installation path correctly , so I modified
>>> >>>the
>>> >>> > script to set the correct hive lib path.
>>> >>> >     Now I dont get no classs found error so it means that class
>>> path
>>> >>>is
>>> >>> > set right now to resolve hive dependencies.
>>> >>> >
>>> >>> > The next exception I get is following ,
>>> >>> > NoSuchObjectException(message:DEFAULT.books table not found)
>>> >>> >
>>> >>> > I could run following query on hive console;
>>> >>> > select * from default.books;
>>> >>> >
>>> >>> > So again problem seems to be some settings related to hive , can
>>> >>>someone
>>> >>> > help me locate this issue ?
>>> >>> > Also I want to know do I need to include following lib; I dont have
>>> >>>it
>>> >>> > currently in my path
>>> >>> > /usr/hdp/current/hive-client/conf/
>>> >>> >
>>> >>> > Exception:
>>> >>> > 2015-02-20 01:15:18,163 INFO  [http-bio-7070-exec-10]
>>> >>> > DataNucleus.Datastore: The class
>>> >>> "org.apache.hadoop.hive.metastore.model.MFieldSchema"
>>> >>> > is tagged as "embedded-only" so does not have its own datastore
>>> >>>table.
>>> >>> > 2015-02-20 01:15:18,163 INFO  [http-bio-7070-exec-10]
>>> >>> > DataNucleus.Datastore: The class
>>> >>> "org.apache.hadoop.hive.metastore.model.MOrder"
>>> >>> > is tagged as "embedded-only" so does not have its own datastore
>>> >>>table.
>>> >>> > 2015-02-20 01:15:18,532 INFO  [http-bio-7070-exec-10]
>>> >>> > DataNucleus.Datastore: The class
>>> >>> "org.apache.hadoop.hive.metastore.model.MFieldSchema"
>>> >>> > is tagged as "embedded-only" so does not have its own datastore
>>> >>>table.
>>> >>> > 2015-02-20 01:15:18,532 INFO  [http-bio-7070-exec-10]
>>> >>> > DataNucleus.Datastore: The class
>>> >>> "org.apache.hadoop.hive.metastore.model.MOrder"
>>> >>> > is tagged as "embedded-only" so does not have its own datastore
>>> >>>table.
>>> >>> > 2015-02-20 01:15:18,710 INFO  [http-bio-7070-exec-10]
>>> >>>DataNucleus.Query:
>>> >>> > Reading in results for query
>>> >>> "org.datanucleus.store.rdbms.query.SQLQuery@0"
>>> >>> > since the connection used is closing
>>> >>> > 2015-02-20 01:15:18,713 INFO  [http-bio-7070-exec-10]
>>> >>> > metastore.ObjectStore: Initialized ObjectStore
>>> >>> > 2015-02-20 01:15:19,002 INFO  [http-bio-7070-exec-10]
>>> >>> > metastore.HiveMetaStore: Added admin role in metastore
>>> >>> > 2015-02-20 01:15:19,004 INFO  [http-bio-7070-exec-10]
>>> >>> > metastore.HiveMetaStore: Added public role in metastore
>>> >>> > 2015-02-20 01:15:19,090 INFO  [http-bio-7070-exec-10]
>>> >>> > metastore.HiveMetaStore: No user is added in admin role, since
>>> >>>config is
>>> >>> > empty
>>> >>> > 2015-02-20 01:15:19,213 INFO  [http-bio-7070-exec-10]
>>> >>> > metastore.HiveMetaStore: 0: get_table : db=DEFAULT tbl=books
>>> >>> > 2015-02-20 01:15:19,213 INFO  [http-bio-7070-exec-10]
>>> >>> HiveMetaStore.audit:
>>> >>> > ugi=root ip=unknown-ip-addr cmd=get_table : db=DEFAULT tbl=books
>>> >>> > NoSuchObjectException(message:DEFAULT.books table not found)
>>> >>> >  at org.apache.hadoop.hive.metastore.HiveMetaStore$
>>> >>> > HMSHandler.get_table_core(HiveMetaStore.java:1729)
>>> >>> >  at
>>> >>>org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_table(
>>> >>> > HiveMetaStore.java:1699)
>>> >>> >  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> >>> >  at sun.reflect.NativeMethodAccessorImpl.invoke(
>>> >>> > NativeMethodAccessorImpl.java:57)
>>> >>> >  at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>> >>> > DelegatingMethodAccessorImpl.java:43)
>>> >>> >  at java.lang.reflect.Method.invoke(Method.java:606)
>>> >>> >  at org.apache.hadoop.hive.metastore.RetryingHMSHandler.
>>> >>> > invoke(RetryingHMSHandler.java:102)
>>> >>> >  at com.sun.proxy.$Proxy37.get_table(Unknown Source)
>>> >>> >  at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.
>>> >>> > getTable(HiveMetaStoreClient.java:1091)
>>> >>> >  at org.apache.kylin.common.util.HiveClient.getHiveTable(
>>> >>> > HiveClient.java:115)
>>> >>> >  at org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>>> >>> > extractHiveTables(HiveSourceTableLoader.java:86)
>>> >>> >  at
>>> >>> org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>>> reloadHiveTables(
>>> >>> > HiveSourceTableLoader.java:70)
>>> >>> >  at org.apache.kylin.rest.service.CubeService.reloadHiveTable(
>>> >>> > CubeService.java:537)
>>> >>> >  at org.apache.kylin.rest.service.CubeService$$FastClassByCGLIB$
>>> >>> > $17a07c0e.invoke(<generated>)
>>> >>> >
>>> >>> >
>>> >>> > Regards,
>>> >>> > Santosh Akhilesh
>>> >>> > Bangalore R&D
>>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> >
>>> >>> > www.huawei.com
>>> >>> > ------------------------------------------------------------
>>> >>> >
>>> >>>
>>> >>>---------------------------------------------------------
>>> ---------------
>>> >>>-
>>> >>> > This e-mail and its attachments contain confidential information
>>> from
>>> >>> > HUAWEI, which
>>> >>> > is intended only for the person or entity whose address is listed
>>> >>>above.
>>> >>> > Any use of the
>>> >>> > information contained herein in any way (including, but not limited
>>> >>>to,
>>> >>> > total or partial
>>> >>> > disclosure, reproduction, or dissemination) by persons other than
>>> the
>>> >>> > intended
>>> >>> > recipient(s) is prohibited. If you receive this e-mail in error,
>>> >>>please
>>> >>> > notify the sender by
>>> >>> > phone or email immediately and delete it!
>>> >>> >
>>> >>> > ________________________________________
>>> >>> > From: Santoshakhilesh [[email protected]]
>>> >>> > Sent: Thursday, February 19, 2015 3:37 PM
>>> >>> > To: [email protected]
>>> >>> > Subject: RE: Hive Table is not loaded in Kylin//RE: About building
>>> >>>kylin
>>> >>> >
>>> >>> > Looks like this is the issue , I find one exception in
>>> >>> > localhost.2015-02-19.log
>>> >>> >
>>> >>> >
>>> >>> > Does this mean there is some issue in hive version I use , I am
>>> using
>>> >>> > 0.14.0?
>>> >>> >
>>> >>> >
>>> >>> > SEVERE: Servlet.service() for servlet [kylin] in context with path
>>> >>> > [/kylin] threw exception [Handler processing failed; nested
>>> >>>exception is
>>> >>> > java.lang.NoClassDefFoundError:
>>> >>> org/apache/hadoop/hive/ql/session/SessionState]
>>> >>> > with root cause
>>> >>> > java.lang.ClassNotFoundException: org.apache.hadoop.hive.ql.
>>> >>> > session.SessionState
>>> >>> >  at org.apache.catalina.loader.WebappClassLoader.loadClass(
>>> >>> > WebappClassLoader.java:1720)
>>> >>> >  at org.apache.catalina.loader.WebappClassLoader.loadClass(
>>> >>> > WebappClassLoader.java:1571)
>>> >>> >  at org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>>> >>> > extractHiveTables(HiveSourceTableLoader.java:83)
>>> >>> >  at
>>> >>> org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>>> reloadHiveTables(
>>> >>> > HiveSourceTableLoader.java:70)
>>> >>> >  at org.apache.kylin.rest.service.CubeService.reloadHiveTable(
>>> >>> > CubeService.java:537)
>>> >>> >  at org.apache.kylin.rest.service.CubeService$$FastClassByCGLIB$
>>> >>> > $17a07c0e.invoke(<generated>)
>>> >>> >  at net.sf.cglib.proxy.MethodProxy.invoke(MethodProxy.java:204)
>>> >>> >
>>> >>> > Regards,
>>> >>> > Santosh Akhilesh
>>> >>> > Bangalore R&D
>>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> >
>>> >>> > www.huawei.com
>>> >>> > ------------------------------------------------------------
>>> >>> >
>>> >>>
>>> >>>---------------------------------------------------------
>>> ---------------
>>> >>>-
>>> >>> > This e-mail and its attachments contain confidential information
>>> from
>>> >>> > HUAWEI, which
>>> >>> > is intended only for the person or entity whose address is listed
>>> >>>above.
>>> >>> > Any use of the
>>> >>> > information contained herein in any way (including, but not limited
>>> >>>to,
>>> >>> > total or partial
>>> >>> > disclosure, reproduction, or dissemination) by persons other than
>>> the
>>> >>> > intended
>>> >>> > recipient(s) is prohibited. If you receive this e-mail in error,
>>> >>>please
>>> >>> > notify the sender by
>>> >>> > phone or email immediately and delete it!
>>> >>> >
>>> >>> > ________________________________________
>>> >>> > From: Santoshakhilesh [[email protected]]
>>> >>> > Sent: Thursday, February 19, 2015 1:27 PM
>>> >>> > To: [email protected]
>>> >>> > Cc: Kulbhushan Rana
>>> >>> > Subject: Hive Table is not loaded in Kylin//RE: About building
>>> kylin
>>> >>> >
>>> >>> > Hi ,
>>> >>> >
>>> >>> >    I have resolved this issue. I have hivesingle column table and I
>>> >>>can
>>> >>> > access from Hive console and do queries on it sucecssfully using
>>> hive
>>> >>> > console.
>>> >>> >
>>> >>> >    When I try to synch hive table using Kylin web specifiying the
>>> >>>hive
>>> >>> > table name and synch.
>>> >>> >
>>> >>> >    I get an error dialog "Oops Failed to take action" .
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > Kylin.log
>>> >>> >
>>> >>> >
>>> >>> > [pool-3-thread-1]:[2015-02-19 21:09:54,152][INFO][org.
>>> >>> >
>>> >>>
>>> >>>apache.kylin.job.impl.threadpool.DefaultScheduler$
>>> FetcherRunner.run(Defa
>>> >>>ultScheduler.java:117)]
>>> >>> > - Job Fetcher: 0 running, 0 actual running, 0 ready, 0 others
>>> >>> > [http-bio-7070-exec-4]:[2015-02-19 21:09:59,556][INFO][org.
>>> >>> >
>>> >>>
>>> >>>apache.kylin.rest.controller.TableController.
>>> getHiveTables(TableControll
>>> >>>er.java:74)]
>>> >>> > - Return all table metadata in 0 seconds
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > localhost_access_log.2015-02-19.txt
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > 10.19.36.67 - - [19/Feb/2015:21:21:58 +0800] "POST
>>> >>> > /kylin/api/tables/books/Sample HTTP/1.1" 500 1972
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > Here books is my table (I have used default database) and sample is
>>> >>>name
>>> >>> > of my project
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > What could be reason of not loading the hive tables ? Someone whi
>>> is
>>> >>> able
>>> >>> > to load the hive table successfully can help me to debug this
>>> issue ?
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > Regards,
>>> >>> > Santosh Akhilesh
>>> >>> > Bangalore R&D
>>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> >
>>> >>> > www.huawei.com
>>> >>> > ------------------------------------------------------------
>>> >>> >
>>> >>>
>>> >>>---------------------------------------------------------
>>> ---------------
>>> >>>-
>>> >>> > This e-mail and its attachments contain confidential information
>>> from
>>> >>> > HUAWEI, which
>>> >>> > is intended only for the person or entity whose address is listed
>>> >>>above.
>>> >>> > Any use of the
>>> >>> > information contained herein in any way (including, but not limited
>>> >>>to,
>>> >>> > total or partial
>>> >>> > disclosure, reproduction, or dissemination) by persons other than
>>> the
>>> >>> > intended
>>> >>> > recipient(s) is prohibited. If you receive this e-mail in error,
>>> >>>please
>>> >>> > notify the sender by
>>> >>> > phone or email immediately and delete it!
>>> >>> > ________________________________
>>> >>> > From: Santosh Akhilesh [[email protected]]
>>> >>> > Sent: Tuesday, February 17, 2015 8:13 PM
>>> >>> > To: [email protected]
>>> >>> > Subject: Re: About building kylin
>>> >>> >
>>> >>> > Hi,
>>> >>> >      These days I spent time to learn Hive and create tables
>>> required
>>> >>> for
>>> >>> > my test.
>>> >>> >       Unfortunately when I am all set with Hive tables , Kylin
>>> can't
>>> >>> start
>>> >>> > normally. Before this I could start and login to web ui. the logs
>>> are
>>> >>> like
>>> >>> > below. What has gone wring any idea ?
>>> >>> >
>>> >>> > SLF4J: Class path contains multiple SLF4J bindings.
>>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>>> >>> > frameworks/hbase-0.98.10-hadoop2/lib/slf4j-log4j12-1.6.
>>> >>> > 4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>>> >>> > frameworks/hadoop-1.2.1/lib/slf4j-log4j12-1.4.3.jar!/org/
>>> >>> > slf4j/impl/StaticLoggerBinder.class]
>>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>>> >>> > frameworks/kylin/tomcat/webapps/kylin/WEB-INF/lib/
>>> >>> > slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> >>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
>>> an
>>> >>> > explanation.
>>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,307][WARN][org.
>>> >>> >
>>> >>>
>>> >>>apache.kylin.common.KylinConfig.getKylinProperties(
>>> KylinConfig.java:527)
>>> >>>]
>>> >>> > - KYLIN_CONF_HOME has not been set
>>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,419][INFO][org.
>>> >>> >
>>> >>>springframework.core.io.support.PropertiesLoaderSupport.
>>> loadProperties(
>>> >>> > PropertiesLoaderSupport.java:177)] - Loading properties file from
>>> >>> > resource loaded through InputStream
>>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,815][WARN][org.
>>> >>> >
>>> >>>
>>> >>>apache.kylin.common.KylinConfig.getKylinProperties(
>>> KylinConfig.java:527)
>>> >>>]
>>> >>> > - KYLIN_CONF_HOME has not been set
>>> >>> > SLF4J: Class path contains multiple SLF4J bindings.
>>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>>> >>> > frameworks/hbase-0.98.10-hadoop2/lib/slf4j-log4j12-1.6.
>>> >>> > 4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>>> >>> > frameworks/hadoop-1.2.1/lib/slf4j-log4j12-1.4.3.jar!/org/
>>> >>> > slf4j/impl/StaticLoggerBinder.class]
>>> >>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
>>> an
>>> >>> > explanation.
>>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:43,051][ERROR][org.
>>> >>> >
>>> >>>
>>> >>>springframework.web.context.ContextLoader.
>>> initWebApplicationContext(Cont
>>> >>>extLoader.java:307)]
>>> >>> > - Context initialization failed
>>> >>> > org.springframework.beans.factory.BeanCreationException: Error
>>> >>>creating
>>> >>> > bean with name 'com.ryantenney.metrics.spring.
>>> >>> > MetricsBeanPostProcessorFactory#0': BeanPostProcessor before
>>> >>> > instantiation of bean failed; nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name
>>> >>> 'org.springframework.cache.config.internalCacheAdvisor':
>>> >>> > Cannot resolve reference to bean
>>> >>>'org.springframework.cache.annotation.
>>> >>> > AnnotationCacheOperationSource#0' while setting bean property
>>> >>> > 'cacheOperationSource'; nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name 'org.springframework.cache.
>>> annotation.
>>> >>> > AnnotationCacheOperationSource#0': BeanPostProcessor before
>>> >>> instantiation
>>> >>> > of bean failed; nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name 'org.springframework.security.
>>> >>> > methodSecurityMetadataSourceAdvisor': Cannot resolve reference to
>>> >>>bean
>>> >>> >
>>> >>>
>>> >>>'org.springframework.security.access.method.
>>> DelegatingMethodSecurityMeta
>>> >>>dataSource#0'
>>> >>> > while setting constructor argument; nested exception is
>>> >>> > org.springframework.beans.factory.BeanCreationException: Error
>>> >>>creating
>>> >>> > bean with name 'org.springframework.security.access.method.
>>> >>> > DelegatingMethodSecurityMetadataSource#0': Cannot create inner
>>> bean
>>> >>> > '(inner bean)' of type [org.springframework.security.
>>> access.prepost.
>>> >>> > PrePostAnnotationSecurityMetadataSource] while setting constructor
>>> >>> > argument with key [0]; nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name '(inner bean)': Cannot create inner
>>> >>>bean
>>> >>> > '(inner bean)' of type [org.springframework.security.
>>> >>> > access.expression.method.ExpressionBasedAnnotationAttri
>>> buteFactory]
>>> >>> while
>>> >>> > setting constructor argument; nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name '(inner bean)': Cannot resolve
>>> >>>reference
>>> >>> to
>>> >>> > bean 'expressionHandler' while setting constructor argument; nested
>>> >>> > exception is org.springframework.beans.
>>> factory.BeanCreationException:
>>> >>> > Error creating bean with name 'expressionHandler' defined in class
>>> >>>path
>>> >>> > resource [kylinSecurity.xml]: Cannot resolve reference to bean
>>> >>> > 'permissionEvaluator' while setting bean property
>>> >>>'permissionEvaluator';
>>> >>> > nested exception is
>>> >>> org.springframework.beans.factory.BeanCreationException:
>>> >>> > Error creating bean with name 'permissionEvaluator' defined in
>>> class
>>> >>> path
>>> >>> > resource [kylinSecurity.xml]: Cannot resolve reference to bean
>>> >>> 'aclService'
>>> >>> > while setting constructor argument; nested exception is
>>> >>> > org.springframework.beans.factory.BeanCreationException: Error
>>> >>>creating
>>> >>> > bean with name 'aclService' defined in file [/home/santosh/work/
>>> >>> > frameworks/kylin/tomcat/webapps/kylin/WEB-INF/classes/
>>> >>> > org/apache/kylin/rest/service/AclService.class]: Instantiation of
>>> >>>bean
>>> >>> > failed; nested exception is
>>> >>> org.springframework.beans.BeanInstantiationException:
>>> >>> > Could not instantiate bean class
>>> >>> [org.apache.kylin.rest.service.AclService]:
>>> >>> > Constructor threw exception; nested exception is
>>> >>> org.apache.kylin.common.persistence.StorageException:
>>> >>> > Error when open connection hbase
>>> >>> > at org.springframework.beans.factory.support.
>>> >>> > AbstractAutowireCapableBeanFactory.createBean(
>>> >>> > AbstractAutowireCapableBeanFactory.java:452)
>>> >>> > at org.springframework.beans.factory.support.
>>> AbstractBeanFactory$1.
>>> >>> > getObject(AbstractBeanFactory.java:294)
>>> >>> > at
>>> >>> org.springframework.beans.factory.support.
>>> DefaultSingletonBeanRegistry.
>>> >>> > getSingleton(DefaultSingletonBeanRegistry.java:225)
>>> >>> > at org.springframework.beans.factory.support.
>>> >>> > AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:291)
>>> >>> > at
>>> >>> org.springframework.beans.factory.support.
>>> AbstractBeanFactory.getBean(
>>> >>> > AbstractBeanFactory.java:197)
>>> >>> > at org.springframework.context.support.AbstractApplicationContext.
>>> >>> > registerBeanPostProcessors(AbstractApplicationContext.java:745)
>>> >>> > at
>>> >>> org.springframework.context.support.AbstractApplicationContext.
>>> refresh(
>>> >>> > AbstractApplicationContext.java:454)
>>> >>> > at org.springframework.web.context.ContextLoader.
>>> >>> > configureAndRefreshWebApplicationContext(ContextLoader.java:383)
>>> >>> > at org.springframework.web.context.ContextLoader.
>>> >>> > initWebApplicationContext(ContextLoader.java:283)
>>> >>> > at org.springframework.web.context.ContextLoaderListener.
>>> >>> > contextInitialized(ContextLoaderListener.java:111)
>>> >>> > at org.apache.catalina.core.StandardContext.listenerStart(
>>> >>> > StandardContext.java:5016)
>>> >>> > at org.apache.catalina.core.StandardContext.startInternal(
>>> >>> > StandardContext.java:5524)
>>> >>> > at
>>> >>>org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
>>> >>> > at org.apache.catalina.core.ContainerBase.addChildInternal(
>>> >>> > ContainerBase.java:901)
>>> >>> > at
>>> >>> org.apache.catalina.core.ContainerBase.addChild(
>>> ContainerBase.java:877)
>>> >>> > at
>>> >>>org.apache.catalina.core.StandardHost.addChild(StandardHost.java:649)
>>> >>> > at
>>> >>> org.apache.catalina.startup.HostConfig.deployWAR(
>>> HostConfig.java:1081)
>>> >>> > at org.apache.catalina.startup.HostConfig$DeployWar.run(
>>> >>> > HostConfig.java:1877)
>>> >>> > at
>>> >>> java.util.concurrent.Executors$RunnableAdapter.
>>> call(Executors.java:471)
>>> >>> > at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>> >>> > at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>> >>> > ThreadPoolExecutor.java:1145)
>>> >>> > at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>> >>> > ThreadPoolExecutor.java:615)
>>> >>> > at java.lang.Thread.run(Thread.java:745)
>>> >>> > Caused by: org.springframework.beans.
>>> factory.BeanCreationException:
>>> >>> Error
>>> >>> > creating bean with name
>>> >>> 'org.springframework.cache.config.internalCacheAdvisor':
>>> >>> > Cannot resolve reference to bean
>>> >>> >
>>> >>> > On Fri Feb 13 2015 at 8:32:48 PM Zhou, Qianhao <[email protected]
>>> >>> <mailto:
>>> >>> > [email protected]>> wrote:
>>> >>> > You should create your own table in hive.
>>> >>> >
>>> >>> > 2015年2月13日 下午10:47于 Santosh Akhilesh
>>> >>><[email protected]<mailto:
>>> >>> > [email protected]>>写道:
>>> >>> > Oh thanks. You mean I should create my own table schema and insert
>>> >>>data
>>> >>> and
>>> >>> > then need to import?
>>> >>> > Does it mean I should create tables in hbase ? Any specific
>>> location
>>> >>>?
>>> >>> I am
>>> >>> > quite new to hbase and hive so may be I have many silly questions.
>>> >>> > On Fri, 13 Feb 2015 at 8:05 pm, 周千昊 <[email protected]<mailto:z.
>>> >>> > [email protected]>> wrote:
>>> >>> >
>>> >>> > > Hi Santosh,
>>> >>> > >      As you have setup a new hadoop environment, so there is no
>>> >>>data
>>> >>> in
>>> >>> > > hive. And we have not provided standalone data pack for this
>>> >>> situation.
>>> >>> > > That means you have to manually import the data into hive table.
>>> >>> > >
>>> >>> > > On Fri Feb 13 2015 at 10:26:57 PM Santosh Akhilesh <
>>> >>> > > [email protected]<mailto:[email protected]>>
>>> wrote:
>>> >>> > >
>>> >>> > > > I have seen this link but my question is how to get EDW data
>>> >>>source
>>> >>> as
>>> >>> > > > mentioned in tutorial
>>> >>> > > > On Fri, 13 Feb 2015 at 7:41 pm, 周千昊
>>> >>><[email protected]<mailto:z.
>>> >>> > [email protected]>> wrote:
>>> >>> > > >
>>> >>> > > > > Hi, Santosh
>>> >>> > > > >      There is a tutorial about how to create a cube:
>>> >>> > > > >      https://github.com/KylinOLAP/Kylin/wiki/Kylin-Cube-
>>> >>> > > > Creation-Tutorial
>>> >>> > > > >
>>> >>> > > > > On Fri Feb 13 2015 at 10:05:55 PM Santosh Akhilesh <
>>> >>> > > > > [email protected]<mailto:[email protected]>>
>>> >>> wrote:
>>> >>> > > > >
>>> >>> > > > > > Hi Zhou,
>>> >>> > > > > > Yes I understand. I was wondering how do I use the create
>>> >>>cube
>>> >>> > > > tutorial.
>>> >>> > > > > If
>>> >>> > > > > > there are some sample tables exist for me to try out the
>>> >>>sample
>>> >>> > cube
>>> >>> > > > > > described in tutorial. I am quite new to these stuffs so if
>>> >>>you
>>> >>> can
>>> >>> > > > give
>>> >>> > > > > me
>>> >>> > > > > > pointers about how do I create sample hive tables and
>>> import
>>> >>>it
>>> >>> > would
>>> >>> > > > be
>>> >>> > > > > > great. Thanks for your quick responses. I know in China you
>>> >>>have
>>> >>> > > spring
>>> >>> > > > > > holidays but you guys still working :)
>>> >>> > > > > > Regards
>>> >>> > > > > > Santosh
>>> >>> > > > > > On Fri, 13 Feb 2015 at 7:13 pm, Zhou, Qianhao <
>>> >>> [email protected]
>>> >>> > <mailto:[email protected]>>
>>> >>> > > > wrote:
>>> >>> > > > > >
>>> >>> > > > > > > Hi Santosh
>>> >>> > > > > > >     Kylin is not responsible for source of the data,
>>> >>>however
>>> >>> you
>>> >>> > > can
>>> >>> > > > > > > import table from your hive table, there is a tab in the
>>> >>>web
>>> >>> ui
>>> >>> > > > called
>>> >>> > > > > > > table, you can import the table there
>>> >>> > > > > > >
>>> >>> > > > > > > 2015年2月13日 下午9:33于 Santoshakhilesh <
>>> >>> [email protected]<
>>> >>> > mailto:[email protected]>>
>>> >>> > > 写道:
>>> >>> > > > > > > I saw the following link about creating the cube.
>>> >>> > > > > > >
>>> >>> > > > > > > https://github.com/KylinOLAP/Kylin/wiki/Kylin-Cube-
>>> >>> > > Creation-Tutorial
>>> >>> > > > > > >
>>> >>> > > > > > > In this there is already source table EDW exist , But in
>>> my
>>> >>> > > > environment
>>> >>> > > > > > > there is no such table.
>>> >>> > > > > > > How can I create this table manually or I need to export
>>> >>>from
>>> >>> > > > > somewhere ?
>>> >>> > > > > > >
>>> >>> > > > > > > Also , how do I insert tables , data in table when I
>>> start
>>> >>> > creating
>>> >>> > > > my
>>> >>> > > > > > own
>>> >>> > > > > > > cubes and tables ?
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > > Regards,
>>> >>> > > > > > > Santosh Akhilesh
>>> >>> > > > > > > Bangalore R&D
>>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > -------------
>>> >>> > > > > > > This e-mail and its attachments contain confidential
>>> >>> information
>>> >>> > > from
>>> >>> > > > > > > HUAWEI, which
>>> >>> > > > > > > is intended only for the person or entity whose address
>>> is
>>> >>> listed
>>> >>> > > > > above.
>>> >>> > > > > > > Any use of the
>>> >>> > > > > > > information contained herein in any way (including, but
>>> not
>>> >>> > limited
>>> >>> > > > to,
>>> >>> > > > > > > total or partial
>>> >>> > > > > > > disclosure, reproduction, or dissemination) by persons
>>> >>>other
>>> >>> than
>>> >>> > > the
>>> >>> > > > > > > intended
>>> >>> > > > > > > recipient(s) is prohibited. If you receive this e-mail in
>>> >>> error,
>>> >>> > > > please
>>> >>> > > > > > > notify the sender by
>>> >>> > > > > > > phone or email immediately and delete it!
>>> >>> > > > > > >
>>> >>> > > > > > > ________________________________________
>>> >>> > > > > > > From: Santoshakhilesh [[email protected]<
>>> mailto:
>>> >>> > [email protected]>]
>>> >>> > > > > > > Sent: Friday, February 13, 2015 5:40 PM
>>> >>> > > > > > > To:
>>> >>>[email protected]<mailto:[email protected]
>>> >>> .
>>> >>> > apache.org>
>>> >>> > > > > > > Subject: RE: About building kylin
>>> >>> > > > > > >
>>> >>> > > > > > > Hi Zhou,
>>> >>> > > > > > >    Thanks a million, it was hbase issue , now I am able
>>> to
>>> >>> login
>>> >>> > to
>>> >>> > > > > main
>>> >>> > > > > > > interface page.
>>> >>> > > > > > >    In the documents I see a section of building cubes.
>>> are
>>> >>> there
>>> >>> > > any
>>> >>> > > > > > > sample cubles I can try with ?
>>> >>> > > > > > >    Can you please tell me where can I get them and steps
>>> to
>>> >>> > import
>>> >>> > > > and
>>> >>> > > > > > > query them.
>>> >>> > > > > > >
>>> >>> > > > > > > Regards,
>>> >>> > > > > > > Santosh Akhilesh
>>> >>> > > > > > > Bangalore R&D
>>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > -------------
>>> >>> > > > > > > This e-mail and its attachments contain confidential
>>> >>> information
>>> >>> > > from
>>> >>> > > > > > > HUAWEI, which
>>> >>> > > > > > > is intended only for the person or entity whose address
>>> is
>>> >>> listed
>>> >>> > > > > above.
>>> >>> > > > > > > Any use of the
>>> >>> > > > > > > information contained herein in any way (including, but
>>> not
>>> >>> > limited
>>> >>> > > > to,
>>> >>> > > > > > > total or partial
>>> >>> > > > > > > disclosure, reproduction, or dissemination) by persons
>>> >>>other
>>> >>> than
>>> >>> > > the
>>> >>> > > > > > > intended
>>> >>> > > > > > > recipient(s) is prohibited. If you receive this e-mail in
>>> >>> error,
>>> >>> > > > please
>>> >>> > > > > > > notify the sender by
>>> >>> > > > > > > phone or email immediately and delete it!
>>> >>> > > > > > >
>>> >>> > > > > > > ________________________________________
>>> >>> > > > > > > From: Santoshakhilesh [[email protected]<
>>> mailto:
>>> >>> > [email protected]>]
>>> >>> > > > > > > Sent: Friday, February 13, 2015 2:38 PM
>>> >>> > > > > > > To:
>>> >>>[email protected]<mailto:[email protected]
>>> >>> .
>>> >>> > apache.org>
>>> >>> > > > > > > Cc: Kulbhushan Rana
>>> >>> > > > > > > Subject: RE: About building kylin
>>> >>> > > > > > >
>>> >>> > > > > > > Hi Zhou,
>>> >>> > > > > > >     The exceptions I see are below , So it seems Kylin
>>> >>> instance
>>> >>> > is
>>> >>> > > > not
>>> >>> > > > > > > deployed correctly;
>>> >>> > > > > > >
>>> >>> > > > > > > WARNING: Failed to scan
>>> >>> [file:/contrib/capacity-scheduler/*.jar]
>>> >>> > > > from
>>> >>> > > > > > > classloader hierarchy
>>> >>> > > > > > > java.io.FileNotFoundException:
>>> >>> /contrib/capacity-scheduler/*.jar
>>> >>> > > (No
>>> >>> > > > > > such
>>> >>> > > > > > > file or directory)
>>> >>> > > > > > >  at java.util.zip.ZipFile.open(Native Method)
>>> >>> > > > > > >  at java.util.zip.ZipFile.<init>(ZipFile.java:215)
>>> >>> > > > > > >  at java.util.zip.ZipFile.<init>(ZipFile.java:145)
>>> >>> > > > > > >  at java.util.jar.JarFile.<init>(JarFile.java:154)
>>> >>> > > > > > >  at java.util.jar.JarFile.<init>(JarFile.java:91)
>>> >>> > > > > > >
>>> >>> > > > > > > Feb 13, 2015 10:27:32 PM org.apache.catalina.startup.
>>> >>> > ContextConfig
>>> >>> > > > > > > processResourceJARs
>>> >>> > > > > > > SEVERE: Failed to processes JAR found at URL
>>> >>> > > > > [jar:file:/contrib/capacity-
>>> >>> > > > > > scheduler/*.jar!/]
>>> >>> > > > > > > for static resources to be included in context with name
>>> >>> > > > > > > [jar:file:/contrib/capacity-scheduler/*.jar!/]
>>> >>> > > > > > >
>>> >>> > > > > > > resource loaded through InputStream
>>> >>> > > > > > > [localhost-startStop-1]:[2015-02-13
>>> >>>22:27:34,475][WARN][org.
>>> >>> > > > > > > apache.kylin.common.KylinConfig.getKylinProperties(
>>> >>> > > > > > KylinConfig.java:527)]
>>> >>> > > > > > > - KYLIN_CONF_HOME has not been set
>>> >>> > > > > > > 2015-02-13 22:27:34,919 INFO  [localhost-startStop-1]
>>> >>> zookeeper.
>>> >>> > > > > > RecoverableZooKeeper:
>>> >>> > > > > > > Process identifier=hconnection-0x2b5a652e connecting to
>>> >>> > ZooKeeper
>>> >>> > > > > > > ensemble=localhost:2181
>>> >>> > > > > > >
>>> >>> > > > > > > 2015-02-13 22:27:34,958 INFO  [localhost-startStop-1-
>>> >>> > > > > > SendThread(localhost:2181)]
>>> >>> > > > > > > zookeeper.ClientCnxn: Opening socket connection to server
>>> >>> > > > > > > localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to
>>> >>> authenticate
>>> >>> > > > using
>>> >>> > > > > > > SASL (unknown error)
>>> >>> > > > > > > 2015-02-13 22:27:34,965 WARN  [localhost-startStop-1-
>>> >>> > > > > > SendThread(localhost:2181)]
>>> >>> > > > > > > zookeeper.ClientCnxn: Session 0x0 for server null,
>>> >>>unexpected
>>> >>> > > error,
>>> >>> > > > > > > closing socket connection and attempting reconnect
>>> >>> > > > > > > java.net.ConnectException: Connection refused
>>> >>> > > > > > >  at sun.nio.ch.SocketChannelImpl.checkConnect(Native
>>> >>>Method)
>>> >>> > > > > > >  at sun.nio.ch.SocketChannelImpl.finishConnect(
>>> >>> > > > > > SocketChannelImpl.java:739)
>>> >>> > > > > > >  at org.apache.zookeeper.ClientCnxnSocketNIO.
>>> doTransport(
>>> >>> > > > > > > ClientCnxnSocketNIO.java:361)
>>> >>> > > > > > >  at org.apache.zookeeper.ClientCnxn$SendThread.run(
>>> >>> > > > > ClientCnxn.java:1081)
>>> >>> > > > > > > 2015-02-13 22:27:35,070 INFO  [localhost-startStop-1-
>>> >>> > > > > > SendThread(localhost:2181)]
>>> >>> > > > > > > zookeeper.ClientCnxn: Opening socket connection to server
>>> >>> > > localhost/
>>> >>> > > > > > > 127.0.0.1:2181<http://127.0.0.1:2181>. Will not attempt
>>> to
>>> >>> > authenticate using SASL
>>> >>> > > (unknown
>>> >>> > > > > > > error)
>>> >>> > > > > > >
>>> >>> > > > > > > 2015-02-13 22:27:51,920 WARN  [localhost-startStop-1]
>>> >>> zookeeper.
>>> >>> > > > > > RecoverableZooKeeper:
>>> >>> > > > > > > Possibly transient ZooKeeper, quorum=localhost:2181,
>>> >>> > > > > > exception=org.apache.
>>> >>> > > > > > > zookeeper.KeeperException$ConnectionLossException:
>>> >>> > > KeeperErrorCode =
>>> >>> > > > > > > ConnectionLoss for /hbase/hbaseid
>>> >>> > > > > > > 2015-02-13 22:27:51,921 ERROR [localhost-startStop-1]
>>> >>> zookeeper.
>>> >>> > > > > > RecoverableZooKeeper:
>>> >>> > > > > > > ZooKeeper exists failed after 4 attempts
>>> >>> > > > > > > 2015-02-13 22:27:51,921 WARN  [localhost-startStop-1]
>>> >>> > > > zookeeper.ZKUtil:
>>> >>> > > > > > > hconnection-0x2b5a652e, quorum=localhost:2181,
>>> >>> baseZNode=/hbase
>>> >>> > > > Unable
>>> >>> > > > > to
>>> >>> > > > > > > set watcher on znode (/hbase/hbaseid)
>>> >>> > > > > > >
>>> >>>org.apache.zookeeper.KeeperException$ConnectionLossException:
>>> >>> > > > > > > KeeperErrorCode = ConnectionLoss for /hbase/hbaseid
>>> >>> > > > > > >  at org.apache.zookeeper.KeeperException.create(
>>> >>> > > > > KeeperException.java:99)
>>> >>> > > > > > >  at org.apache.zookeeper.KeeperException.create(
>>> >>> > > > > KeeperException.java:51)
>>> >>> > > > > > >  at
>>> >>>org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1045)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.zookeeper.
>>> >>> > RecoverableZooKeeper.exists(
>>> >>> > > > > > > RecoverableZooKeeper.java:222)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.
>>> zookeeper.ZKUtil.checkExists(
>>> >>> > > > > > ZKUtil.java:479)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.zookeeper.ZKClusterId.
>>> >>> > > > readClusterIdZNode(
>>> >>> > > > > > > ZKClusterId.java:65)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.client.ZooKeeperRegistry.
>>> >>> > getClusterId(
>>> >>> > > > > > > ZooKeeperRegistry.java:83)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.client.HConnectionManager$
>>> >>> > > > > > > HConnectionImplementation.retrieveClusterId(
>>> >>> > > > > HConnectionManager.java:897)
>>> >>> > > > > > >  at org.apache.hadoop.hbase.client.HConnectionManager$
>>> >>> > > > > > >
>>> >>>HConnectionImplementation.<init>(HConnectionManager.java:694)
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > > Regards,
>>> >>> > > > > > > Santosh Akhilesh
>>> >>> > > > > > > Bangalore R&D
>>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > >
>>> >>>------------------------------------------------------------
>>> >>> > > > > > -------------
>>> >>> > > > > > > This e-mail and its attachments contain confidential
>>> >>> information
>>> >>> > > from
>>> >>> > > > > > > HUAWEI, which
>>> >>> > > > > > > is intended only for the person or entity whose address
>>> is
>>> >>> listed
>>> >>> > > > > above.
>>> >>> > > > > > > Any use of the
>>> >>> > > > > > > information contained herein in any way (including, but
>>> not
>>> >>> > limited
>>> >>> > > > to,
>>> >>> > > > > > > total or partial
>>> >>> > > > > > > disclosure, reproduction, or dissemination) by persons
>>> >>>other
>>> >>> than
>>> >>> > > the
>>> >>> > > > > > > intended
>>> >>> > > > > > > recipient(s) is prohibited. If you receive this e-mail in
>>> >>> error,
>>> >>> > > > please
>>> >>> > > > > > > notify the sender by
>>> >>> > > > > > > phone or email immediately and delete it!
>>> >>> > > > > > >
>>> >>> > > > > > > ________________________________________
>>> >>> > > > > > > From: Zhou, Qianhao [[email protected]<mailto:
>>> >>> [email protected]
>>> >>> > >]
>>> >>> > > > > > > Sent: Friday, February 13, 2015 2:07 PM
>>> >>> > > > > > > To:
>>> >>>[email protected]<mailto:[email protected]
>>> >>> .
>>> >>> > apache.org>
>>> >>> > > > > > > Subject: Re: About building kylin
>>> >>> > > > > > >
>>> >>> > > > > > > Hi, Santosh
>>> >>> > > > > > > Please check tomcat/logs/kylin.log to see if there is any
>>> >>> > exception
>>> >>> > > > > > >
>>> >>> > > > > > > Best Regard
>>> >>> > > > > > > Zhou QianHao
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > >
>>> >>> > > > > > > On 2/13/15, 4:24 PM, "Santoshakhilesh" <
>>> >>> > > [email protected]<mailto:[email protected]>>
>>> >>> > > > > > > wrote:
>>> >>> > > > > > >
>>> >>> > > > > > > >Hi Zhou,
>>> >>> > > > > > > >      I am able to access the
>>> http://127.0.0.17070/kylin
>>> >>> using
>>> >>> > > > > firefox
>>> >>> > > > > > > >      All I get is a blank page , its not 404 , it
>>> means I
>>> >>> could
>>> >>> > > > > connect
>>> >>> > > > > > > >to web server successfully but there is some issue in
>>> >>> > rendering. I
>>> >>> > > > was
>>> >>> > > > > > > >noty asked for user name and password.
>>> >>> > > > > > > >       When I run start kylin shell file , following
>>> >>>prompt
>>> >>> > really
>>> >>> > > > > means
>>> >>> > > > > > > >everuthing went ok , or I should check some log files to
>>> >>>know
>>> >>> > the
>>> >>> > > > real
>>> >>> > > > > > > >status ?
>>> >>> > > > > > > >A new Kylin instance is started by root, stop it using
>>> >>> > > > "stop-kylin.sh"
>>> >>> > > > > > > >Please visit http://<your_sandbox_ip>:7070/kylin to
>>> play
>>> >>> with
>>> >>> > the
>>> >>> > > > > > cubes!
>>> >>> > > > > > > >
>>> >>> > > > > > > >Regards,
>>> >>> > > > > > > >Santosh Akhilesh
>>> >>> > > > > > > >Bangalore R&D
>>> >>> > > > > > > >HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> > > > > > > >
>>> >>> > > > > > >
>>> >>>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>>> >>> > > > > > >
>>> >>>>-----------------------------------------------------------
>>> >>> > > > > > > ---------------
>>> >>> > > > > > >
>>> >>>>-----------------------------------------------------------
>>> >>> > > > > > > >This e-mail and its attachments contain confidential
>>> >>> information
>>> >>> > > > from
>>> >>> > > > > > > >HUAWEI, which
>>> >>> > > > > > > >is intended only for the person or entity whose address
>>> is
>>> >>> > listed
>>> >>> > > > > above.
>>> >>> > > > > > > >Any use of the
>>> >>> > > > > > > >information contained herein in any way (including, but
>>> >>>not
>>> >>> > > limited
>>> >>> > > > > to,
>>> >>> > > > > > > >total or partial
>>> >>> > > > > > > >disclosure, reproduction, or dissemination) by persons
>>> >>>other
>>> >>> > than
>>> >>> > > > the
>>> >>> > > > > > > >intended
>>> >>> > > > > > > >recipient(s) is prohibited. If you receive this e-mail
>>> in
>>> >>> error,
>>> >>> > > > > please
>>> >>> > > > > > > >notify the sender by
>>> >>> > > > > > > >phone or email immediately and delete it!
>>> >>> > > > > > > >
>>> >>> > > > > > > >________________________________________
>>> >>> > > > > > > >From: Zhou, Qianhao [[email protected]<mailto:
>>> >>> [email protected]
>>> >>> > >]
>>> >>> > > > > > > >Sent: Friday, February 13, 2015 1:04 PM
>>> >>> > > > > > > >To: [email protected]<mailto:
>>> >>> [email protected].
>>> >>> > apache.org>
>>> >>> > > > > > > >Cc: Kulbhushan Rana
>>> >>> > > > > > > >Subject: Re: About building kylin
>>> >>> > > > > > > >
>>> >>> > > > > > > >Hi, Santosh
>>> >>> > > > > > > >    You will see a kylin.war inside
>>> >>> > ${KYLIN_HOME}/tomcat/webapps.
>>> >>> > > > That
>>> >>> > > > > > is
>>> >>> > > > > > > >the server war package. The tomcat version is 7.0.59.
>>> >>> > > > > > > >    If you want to bind 0.0.0.0:7070<http://0.0.0.0:
>>> 7070>,
>>> >>> > please refer to the tomcat
>>> >>> > > > > > > document,
>>> >>> > > > > > > >there should be some configurations to let you do that.
>>> >>> > > > > > > >
>>> >>> > > > > > > >Best Regard
>>> >>> > > > > > > >Zhou QianHao
>>> >>> > > > > > > >
>>> >>> > > > > > > >
>>> >>> > > > > > > >
>>> >>> > > > > > > >
>>> >>> > > > > > > >
>>> >>> > > > > > > >On 2/13/15, 1:46 PM, "Santoshakhilesh" <
>>> >>> > > [email protected]<mailto:[email protected]>
>>> >>> > > > >
>>> >>> > > > > > > >wrote:
>>> >>> > > > > > > >
>>> >>> > > > > > > >>Hi Zhou,
>>> >>> > > > > > > >>     Due to some practical constarints , I have to
>>> setup
>>> >>> > clusster
>>> >>> > > > > > > >>manually.
>>> >>> > > > > > > >>     I have downloaded all the dependency like hadoop ,
>>> >>> hbase ,
>>> >>> > > > hive
>>> >>> > > > > ,
>>> >>> > > > > > > >>tomcat and I have run the check-env.sh after it passes
>>> ,
>>> >>>I
>>> >>> have
>>> >>> > > run
>>> >>> > > > > the
>>> >>> > > > > > > >>start-kylin.sh
>>> >>> > > > > > > >>     Following is the log on console after running the
>>> >>> > > > > > ./start-kylin.sh.
>>> >>> > > > > > > >>     Does this mean everything is ok  or I should check
>>> >>>some
>>> >>> > > other
>>> >>> > > > > > logs ?
>>> >>> > > > > > > >>Currently I am unable to login to http://
>>> >>> > <your_sandbox_ip>:7070/
>>> >>> > > > > kylin
>>> >>> > > > > > > >>from outside the machine.
>>> >>> > > > > > > >>    Is it possible to bind it to 0.0.0.0 :7070 so that
>>> I
>>> >>>can
>>> >>> > > > connect
>>> >>> > > > > > from
>>> >>> > > > > > > >>outside the lab network ? Where can I configure this ?
>>> >>> > > > > > > >>
>>> >>> > > > > > > >> ./start-kylin.sh
>>> >>> > > > > > > >>Checking KYLIN_HOME...
>>> >>> > > > > > > >>KYLIN_HOME is set to /opt/kylinb
>>> >>> > > > > > > >>Checking hbase...
>>> >>> > > > > > > >>hbase check passed
>>> >>> > > > > > > >>Checking hive...
>>> >>> > > > > > > >>hive check passed
>>> >>> > > > > > > >>Checking hadoop...
>>> >>> > > > > > > >>hadoop check passed
>>> >>> > > > > > > >>Logging initialized using configuration in
>>> >>> > > > > > > >>jar:file:/opt/hive/apache-hive-0.14.0-bin/lib/hive-
>>> >>> > > > > > > common-0.14.0.jar!/hiv
>>> >>> > > > > > > >>e
>>> >>> > > > > > > >>-log4j.properties
>>> >>> > > > > > > >>SLF4J: Class path contains multiple SLF4J bindings.
>>> >>> > > > > > > >>SLF4J: Found binding in
>>> >>> > > > > > > >>[jar:file:/opt/hadoop/hadoop-2.6.0/share/hadoop/
>>> >>> > > > > > > common/lib/slf4j-log4j12-
>>> >>> > > > > > > >>1
>>> >>> > > > > > > >>.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> >>> > > > > > > >>SLF4J: Found binding in
>>> >>> > > > > > > >>[jar:file:/opt/hive/apache-
>>> hive-0.14.0-bin/lib/hive-jdbc-
>>> >>> > > > > > > 0.14.0-standalon
>>> >>> > > > > > > >>e
>>> >>> > > > > > > >>.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> >>> > > > > > > >>SLF4J: See
>>> >>> http://www.slf4j.org/codes.html#multiple_bindings
>>> >>> > for
>>> >>> > > > an
>>> >>> > > > > > > >>explanation.
>>> >>> > > > > > > >>SLF4J: Actual binding is of type [org.slf4j.impl.
>>> >>> > > > Log4jLoggerFactory]
>>> >>> > > > > > > >>hive dependency:
>>> >>> > > > > > > >>/usr/hdp/current/hive-client/conf/:/hive/lib/*:/
>>> >>> > > > > > > hive-hcatalog/share/hcata
>>> >>> > > > > > > >>l
>>> >>> > > > > > > >>og/*
>>> >>> > > > > > > >>A new Kylin instance is started by root, stop it using
>>> >>> > > > > "stop-kylin.sh"
>>> >>> > > > > > > >>Please visit http://<your_sandbox_ip>:7070/kylin to
>>> play
>>> >>> with
>>> >>> > > the
>>> >>> > > > > > cubes!
>>> >>> > > > > > > >>(Useranme: ADMIN, Password: KYLIN)
>>> >>> > > > > > > >>You can check the log at ./../tomcat/logs/kylin.log
>>> >>> > > > > > > >>
>>> >>> > > > > > > >>Regards,
>>> >>> > > > > > > >>Santosh Akhilesh
>>> >>> > > > > > > >>Bangalore R&D
>>> >>> > > > > > > >>HUAWEI TECHNOLOGIES CO.,LTD.
>>> >>> > > > > > > >>
>>> >>> > > > > > >
>>> >>>>>www.huawei.com<http://www.huawei.com><http://www.huawei.com
>>> >>> >
>>> >>> > > > > > >
>>> >>>>>----------------------------------------------------------
>>> >>> > > > > > > ---------------
>>> >>> > > > > > > >>-
>>> >>> > > > > > >
>>> >>>>>-----------------------------------------------------------
>>> >>> > > > > > > >>This e-mail and its attachments contain confidential
>>> >>> > information
>>> >>> > > > from
>>> >>> > > > > > > >>HUAWEI, which
>>> >>> > > > > > > >>is intended only for the person or entity whose address
>>> >>>is
>>> >>> > listed
>>> >>> > > > > > above.
>>> >>> > > > > > > >>Any use of the
>>> >>> > > > > > > >>information contained herein in any way (including, but
>>> >>>not
>>> >>> > > limited
>>> >>> > > > > to,
>>> >>> > > > > > > >>total or partial
>>> >>> > > > > > > >>disclosure, reproduction, or dissemination) by persons
>>> >>>other
>>> >>> > than
>>> >>> > > > the
>>> >>> > > > > > > >>intended
>>> >>> > > > > > > >>recipient(s) is prohibited. If you receive this e-mail
>>> in
>>> >>> > error,
>>> >>> > > > > please
>>> >>> > > > > > > >>notify the sender by
>>> >>> > > > > > > >>phone or email immediately and delete it!
>>> >>> > > > > > > >>
>>> >>> > > > > > > >>________________________________________
>>> >>> > > > > > > >>From: Zhou, Qianhao [[email protected]<mailto:qian
>>> >>> > [email protected]>]
>>> >>> > > > > > > >>Sent: Thursday, February 12, 2015 2:59 PM
>>> >>> > > > > > > >>To: [email protected]<mailto:
>>> >>> [email protected].
>>> >>> > apache.org>
>>> >>> > > > > > > >>Subject: Re: About building kylin
>>> >>> > > > > > > >>
>>> >>> > > > > > > >>Hi Santosh
>>> >>> > > > > > > >>   Surely you can run kylin on single node.
>>> >>> > > > > > > >>   Currently kylin is depend on
>>> >>> > > > > > > >>   1. Hive (where source data come from)
>>> >>> > > > > > > >>   2. HBase (where metadata and pre-computed data
>>> stored)
>>> >>> > > > > > > >>   However once you have the hadoop environment ready,
>>> it
>>> >>> does
>>> >>> > > not
>>> >>> > > > > > matter
>>> >>> > > > > > > >>how many vm running in the cluster.
>>> >>> > > > > > > >>   As for your circumstances, a sandbox is strongly
>>> >>> recommended
>>> >>> > > > which
>>> >>> > > > > > > >>will
>>> >>> > > > > > > >>save you a lot of time of setting up the environment.
>>> >>> > Hortonworks
>>> >>> > > > or
>>> >>> > > > > > > >>Cloudera will be good choice.
>>> >>> > > > > > > >>
>>> >>> > > > > > > >>Best Regard
>>> >>> > > > > > > >>Zhou QianHao
>>> >>> > > > > > >
>>> >>> > > > > >
>>> >>> > > > >
>>> >>> > > >
>>> >>> > >
>>> >>>
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> Regards,
>>> >> Santosh Akhilesh
>>> >> +91-0-9845482201
>>> >>
>>> >
>>> >
>>> >
>>> >--
>>> >Regards,
>>> >Santosh Akhilesh
>>> >+91-0-9845482201
>>>
>>>
>>
>>
>> --
>> Regards,
>> Santosh Akhilesh
>> +91-0-9845482201
>>
>

Reply via email to