HI ,
I tried to set the hive-site.xml using the example you gave. But after
applying this , hive shell itself can't load.
There are some configs which are specific to hortonworks sandbox , I had
removed them.
If someone can tell me whats the process of connecting to hive database ,
is it a jdbc connection , depend on sql or through thirft which can help me
to find which params really necessary. As per exception there is exception
while calling getHivetable so some params are not correct in HiveConf.
Regards,
Santosh

On Fri, Feb 20, 2015 at 10:19 PM, Santosh Akhilesh <
[email protected]> wrote:

> Oops thats a big list of configure items :). Most of them will take time
> for me to understand and apply. Thanks any way will get back after I apply
> in my env. To Chinese friends Xin Niam Quai la. Gong xi fa cai:)
>
> On Fri, 20 Feb 2015 at 9:56 pm, 周千昊 <[email protected]> wrote:
>
>> Here is the link:
>> https://github.com/KylinOLAP/Kylin/blob/master/examples/test
>> _case_data/sandbox/hive-site.xml
>>
>> On Sat Feb 21 2015 at 12:21:59 AM Santosh Akhilesh <
>> [email protected]> wrote:
>>
>> > Can someone give me a link to sample hive-site.xml which is working with
>> > kylin?
>> > I feel I need to set some mandatory parameter which hive shell can get
>> but
>> > need to be specified if being accessed from external client? Do I need
>> to
>> > set megastore db path?
>> > On Fri, 20 Feb 2015 at 9:12 pm, Santosh Akhilesh <
>> > [email protected]>
>> > wrote:
>> >
>> > > You replied to wrong thread. Didn't want to hijack that thread.
>> > > My hive Conf is under hive install. Following is the hive dependency I
>> > > export in startkylin.sh
>> > >
>> > > export
>> > > hive_dependency=/home/santosh/work/frameworks/apache-hive-1.
>> > 0.0/conf/:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> > lib/*:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> > hcatalog/share/hcatalog/*
>> > > On Fri, 20 Feb 2015 at 8:43 pm, Santosh Akhilesh <
>> > > [email protected]> wrote:
>> > >
>> > >> Hi ,
>> > >>    I have set the conf path in start_kylin.sh as below. I have
>> removed
>> > >> the sourcing of find-hive-dependency.sh as its not working in my env.
>> > >>
>> > >> In the hive-site.xml , I have set following; still i get same error
>> , is
>> > >> there anything else I need to do ?
>> > >>
>> > >> <configuration>
>> > >>   <property>
>> > >>     <name>hive.exec.scratchdir</name>
>> > >>     <value>/tmp/hive</value>
>> > >>     <description>Scratch space for Hive jobs</description>
>> > >>   </property>
>> > >> </configuration>
>> > >>
>> > >>
>> > >> #source ${dir}/find-hive-dependency.sh
>> > >> export hive_dependency=/home/santosh/work/frameworks/apache-hive-1.
>> > >> 0.0/conf/:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> > >> lib/*:/home/santosh/work/frameworks/apache-hive-1.0.0/
>> > >> hcatalog/share/hcatalog/*
>> > >> export HBASE_CLASSPATH_PREFIX=${tomcat_root}/bin/bootstrap.
>> > >> jar:${tomcat_root}/bin/tomcat-juli.jar:${tomcat_root}/lib/*:
>> > >> $HBASE_CLASSPATH_PREFIX
>> > >> export HBASE_CLASSPATH=$hive_dependency:${HBASE_CLASSPATH}
>> > >>
>> > >> hbase -Djava.util.logging.config.file=${tomcat_root}/conf/
>> > logging.properties
>> > >> \
>> > >> -Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager \
>> > >> -Dorg.apache.tomcat.util.buf.UDecoder.ALLOW_ENCODED_SLASH=true \
>> > >> -Dorg.apache.catalina.connector.CoyoteAdapter.ALLOW_BACKSLASH=true \
>> > >> -Djava.endorsed.dirs=${tomcat_root}/endorsed  \
>> > >> -Dcatalina.base=${tomcat_root} \
>> > >> -Dcatalina.home=${tomcat_root} \
>> > >> -Djava.io.tmpdir=${tomcat_root}/temp  \
>> > >> -Dkylin.hive.dependency=${hive_dependency} \
>> > >> -Dspring.profiles.active=${spring_profile} \
>> > >> org.apache.hadoop.util.RunJar ${tomcat_root}/bin/bootstrap.jar
>> > >>  org.apache.catalina.startup.Bootstrap start >
>> > >> ${tomcat_root}/logs/kylin.log 2>&1 &
>> > >> echo "A new Kylin instance is started by $USER, stop it using
>> > >> \"stop-kylin.sh\""
>> > >> if [ "$useSandbox" = "true" ]
>> > >>     then echo "Please visit http://<your_sandbox_ip>:7070/kylin to
>> play
>> > >> with the cubes! (Useranme: ADMIN, Password: KYLIN)"
>> > >> else
>> > >>     echo "Please visit http://<ip>:7070/kylin"
>> > >> fi
>> > >> echo "You can check the log at ${tomcat_root}/logs/kylin.log"
>> > >>
>> > >>
>> > >> On Fri, Feb 20, 2015 at 8:10 PM, Shi, Shaofeng <[email protected]>
>> > wrote:
>> > >>
>> > >>> Hi,
>> > >>>
>> > >>> Kylin from 7.0.x and later needs hive-site.xml on the classpth, to
>> > allow
>> > >>> Kylin connect with the hive metadata service via API, to get the
>> table
>> > >>> info; As you can execute hive commands on your machine, the hive
>> client
>> > >>> should have been properly set, you can search the hive-site.xml
>> from /
>> > to
>> > >>> get its path, and then add it to the script as Qianhao mentioned.
>> > >>>
>> > >>> On 2/20/15, 10:28 PM, "Santosh Akhilesh" <[email protected]
>> >
>> > >>> wrote:
>> > >>>
>> > >>> >Hi Qianhao,
>> > >>> >        I really wonder what kind of config I need to set for
>> making
>> > >>> sure
>> > >>> >kylin can load the hive tables.
>> > >>> >        I am not using the ready made sandbox , I have downloaded
>> > hbase
>> > >>> ,
>> > >>> >hive , hadoop from apcahe site to make the env. As far as I can see
>> > >>> >currently I can query hive tables in hive shell. Mine is single
>> node
>> > >>> >pseudo
>> > >>> >distributed cluster. I have not set any changes to hive-site.xml ,
>> the
>> > >>> >metastoredb is created under $HIVE_INSTALL/bin
>> > >>> >data is on hdfs under /user/hive/warehouse.
>> > >>> >
>> > >>> >I am really out of clue why the exception , I suspect its some
>> setting
>> > >>> >which kylin expects to make hive query run
>> > >>> >
>> > >>> >Please help me to get over this issue , thanks.
>> > >>> >2015-02-20 19:38:10,435 INFO  [http-bio-7070-exec-3]
>> > >>> HiveMetaStore.audit:
>> > >>> >ugi=santosh ip=unknown-ip-addr cmd=get_table : db=DEFAULT tbl=books
>> > >>> >NoSuchObjectException(message:DEFAULT.books table not found)
>> > >>> >at
>> > >>> >org.apache.hadoop.hive.metastore.HiveMetaStore$
>> > >>> HMSHandler.get_table_core(H
>> > >>> >iveMetaStore.java:1729)
>> > >>> >at
>> > >>> >org.apache.hadoop.hive.metastore.HiveMetaStore$
>> > >>> HMSHandler.get_table(HiveMe
>> > >>> >taStore.java:1699)
>> > >>> >at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> > >>> >at
>> > >>> >sun.reflect.NativeMethodAccessorImpl.invoke(
>> > >>> NativeMethodAccessorImpl.java:
>> > >>> >57)
>> > >>> >at
>> > >>> >sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > >>> DelegatingMethodAccessorIm
>> > >>> >pl.java:43)
>> > >>> >at java.lang.reflect.Method.invoke(Method.java:606)
>> > >>> >at
>> > >>> >org.apache.hadoop.hive.metastore.RetryingHMSHandler.
>> > >>> invoke(RetryingHMSHand
>> > >>> >ler.java:102)
>> > >>> >at com.sun.proxy.$Proxy37.get_table(Unknown Source)
>> > >>> >at
>> > >>> >org.apache.hadoop.hive.metastore.HiveMetaStoreClient.
>> > >>> getTable(HiveMetaStor
>> > >>> >eClient.java:1097)
>> > >>> >at
>> > >>> >org.apache.kylin.common.util.HiveClient.getHiveTable(
>> > >>> HiveClient.java:115)
>> > >>> >at
>> > >>> >org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>> > >>> extractHiveTables(Hiv
>> > >>> >eSourceTableLoader.java:86)
>> > >>> >
>> > >>> >On Fri, Feb 20, 2015 at 4:53 PM, Santosh Akhilesh
>> > >>> ><[email protected]
>> > >>> >> wrote:
>> > >>> >
>> > >>> >> Hi Qianhao,
>> > >>> >>     *I have installed hive from Apache hive site. I dont have
>> > >>> following
>> > >>> >> folder. *
>> > >>> >> /usr/hdp/current/hive-client/conf/ in this script.
>> > >>> >> *I am using Hive 1.0.0 version.*
>> > >>> >> *Where can I get this ? and which conf files I need to set under
>> > conf
>> > >>> >> folder ?*
>> > >>> >> *Regards,*
>> > >>> >> *Santosh Akhilesh*
>> > >>> >>
>> > >>> >>
>> > >>> >>
>> > >>> >>
>> > >>> >> On Fri, Feb 20, 2015 at 10:51 AM, 周千昊 <[email protected]>
>> wrote:
>> > >>> >>
>> > >>> >>> Hi, Santosh
>> > >>> >>>      Thanks for reporting the issue, it looks to me that it is
>> > still
>> > >>> >>>the
>> > >>> >>> problem of find-hive-dependecny.sh.
>> > >>> >>>      As a workaround, you can just manually add
>> > >>> >>> /usr/hdp/current/hive-client/conf/ in this script.
>> > >>> >>>
>> > >>> >>> On Thu Feb 19 2015 at 8:01:39 PM Santoshakhilesh <
>> > >>> >>> [email protected]> wrote:
>> > >>> >>>
>> > >>> >>> > Hi ,
>> > >>> >>> >     I analyzed the Kylin.log and found that
>> > find-hive-dependency.sh
>> > >>> >>>was
>> > >>> >>> > not able to locate hive installation path correctly , so I
>> > modified
>> > >>> >>>the
>> > >>> >>> > script to set the correct hive lib path.
>> > >>> >>> >     Now I dont get no classs found error so it means that
>> class
>> > >>> path
>> > >>> >>>is
>> > >>> >>> > set right now to resolve hive dependencies.
>> > >>> >>> >
>> > >>> >>> > The next exception I get is following ,
>> > >>> >>> > NoSuchObjectException(message:DEFAULT.books table not found)
>> > >>> >>> >
>> > >>> >>> > I could run following query on hive console;
>> > >>> >>> > select * from default.books;
>> > >>> >>> >
>> > >>> >>> > So again problem seems to be some settings related to hive ,
>> can
>> > >>> >>>someone
>> > >>> >>> > help me locate this issue ?
>> > >>> >>> > Also I want to know do I need to include following lib; I dont
>> > have
>> > >>> >>>it
>> > >>> >>> > currently in my path
>> > >>> >>> > /usr/hdp/current/hive-client/conf/
>> > >>> >>> >
>> > >>> >>> > Exception:
>> > >>> >>> > 2015-02-20 01:15:18,163 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > DataNucleus.Datastore: The class
>> > >>> >>> "org.apache.hadoop.hive.metastore.model.MFieldSchema"
>> > >>> >>> > is tagged as "embedded-only" so does not have its own
>> datastore
>> > >>> >>>table.
>> > >>> >>> > 2015-02-20 01:15:18,163 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > DataNucleus.Datastore: The class
>> > >>> >>> "org.apache.hadoop.hive.metastore.model.MOrder"
>> > >>> >>> > is tagged as "embedded-only" so does not have its own
>> datastore
>> > >>> >>>table.
>> > >>> >>> > 2015-02-20 01:15:18,532 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > DataNucleus.Datastore: The class
>> > >>> >>> "org.apache.hadoop.hive.metastore.model.MFieldSchema"
>> > >>> >>> > is tagged as "embedded-only" so does not have its own
>> datastore
>> > >>> >>>table.
>> > >>> >>> > 2015-02-20 01:15:18,532 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > DataNucleus.Datastore: The class
>> > >>> >>> "org.apache.hadoop.hive.metastore.model.MOrder"
>> > >>> >>> > is tagged as "embedded-only" so does not have its own
>> datastore
>> > >>> >>>table.
>> > >>> >>> > 2015-02-20 01:15:18,710 INFO  [http-bio-7070-exec-10]
>> > >>> >>>DataNucleus.Query:
>> > >>> >>> > Reading in results for query
>> > >>> >>> "org.datanucleus.store.rdbms.query.SQLQuery@0"
>> > >>> >>> > since the connection used is closing
>> > >>> >>> > 2015-02-20 01:15:18,713 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > metastore.ObjectStore: Initialized ObjectStore
>> > >>> >>> > 2015-02-20 01:15:19,002 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > metastore.HiveMetaStore: Added admin role in metastore
>> > >>> >>> > 2015-02-20 01:15:19,004 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > metastore.HiveMetaStore: Added public role in metastore
>> > >>> >>> > 2015-02-20 01:15:19,090 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > metastore.HiveMetaStore: No user is added in admin role, since
>> > >>> >>>config is
>> > >>> >>> > empty
>> > >>> >>> > 2015-02-20 01:15:19,213 INFO  [http-bio-7070-exec-10]
>> > >>> >>> > metastore.HiveMetaStore: 0: get_table : db=DEFAULT tbl=books
>> > >>> >>> > 2015-02-20 01:15:19,213 INFO  [http-bio-7070-exec-10]
>> > >>> >>> HiveMetaStore.audit:
>> > >>> >>> > ugi=root ip=unknown-ip-addr cmd=get_table : db=DEFAULT
>> tbl=books
>> > >>> >>> > NoSuchObjectException(message:DEFAULT.books table not found)
>> > >>> >>> >  at org.apache.hadoop.hive.metastore.HiveMetaStore$
>> > >>> >>> > HMSHandler.get_table_core(HiveMetaStore.java:1729)
>> > >>> >>> >  at
>> > >>> >>>org.apache.hadoop.hive.metastore.HiveMetaStore$
>> > HMSHandler.get_table(
>> > >>> >>> > HiveMetaStore.java:1699)
>> > >>> >>> >  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>> > >>> >>> >  at sun.reflect.NativeMethodAccessorImpl.invoke(
>> > >>> >>> > NativeMethodAccessorImpl.java:57)
>> > >>> >>> >  at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> > >>> >>> > DelegatingMethodAccessorImpl.java:43)
>> > >>> >>> >  at java.lang.reflect.Method.invoke(Method.java:606)
>> > >>> >>> >  at org.apache.hadoop.hive.metastore.RetryingHMSHandler.
>> > >>> >>> > invoke(RetryingHMSHandler.java:102)
>> > >>> >>> >  at com.sun.proxy.$Proxy37.get_table(Unknown Source)
>> > >>> >>> >  at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.
>> > >>> >>> > getTable(HiveMetaStoreClient.java:1091)
>> > >>> >>> >  at org.apache.kylin.common.util.HiveClient.getHiveTable(
>> > >>> >>> > HiveClient.java:115)
>> > >>> >>> >  at org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>> > >>> >>> > extractHiveTables(HiveSourceTableLoader.java:86)
>> > >>> >>> >  at
>> > >>> >>> org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>> > >>> reloadHiveTables(
>> > >>> >>> > HiveSourceTableLoader.java:70)
>> > >>> >>> >  at org.apache.kylin.rest.service.
>> CubeService.reloadHiveTable(
>> > >>> >>> > CubeService.java:537)
>> > >>> >>> >  at org.apache.kylin.rest.service.
>> CubeService$$FastClassByCGLIB$
>> > >>> >>> > $17a07c0e.invoke(<generated>)
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > Regards,
>> > >>> >>> > Santosh Akhilesh
>> > >>> >>> > Bangalore R&D
>> > >>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> >
>> > >>> >>> > www.huawei.com
>> > >>> >>> > ------------------------------------------------------------
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>---------------------------------------------------------
>> > >>> ---------------
>> > >>> >>>-
>> > >>> >>> > This e-mail and its attachments contain confidential
>> information
>> > >>> from
>> > >>> >>> > HUAWEI, which
>> > >>> >>> > is intended only for the person or entity whose address is
>> listed
>> > >>> >>>above.
>> > >>> >>> > Any use of the
>> > >>> >>> > information contained herein in any way (including, but not
>> > limited
>> > >>> >>>to,
>> > >>> >>> > total or partial
>> > >>> >>> > disclosure, reproduction, or dissemination) by persons other
>> than
>> > >>> the
>> > >>> >>> > intended
>> > >>> >>> > recipient(s) is prohibited. If you receive this e-mail in
>> error,
>> > >>> >>>please
>> > >>> >>> > notify the sender by
>> > >>> >>> > phone or email immediately and delete it!
>> > >>> >>> >
>> > >>> >>> > ________________________________________
>> > >>> >>> > From: Santoshakhilesh [[email protected]]
>> > >>> >>> > Sent: Thursday, February 19, 2015 3:37 PM
>> > >>> >>> > To: [email protected]
>> > >>> >>> > Subject: RE: Hive Table is not loaded in Kylin//RE: About
>> > building
>> > >>> >>>kylin
>> > >>> >>> >
>> > >>> >>> > Looks like this is the issue , I find one exception in
>> > >>> >>> > localhost.2015-02-19.log
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > Does this mean there is some issue in hive version I use , I
>> am
>> > >>> using
>> > >>> >>> > 0.14.0?
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > SEVERE: Servlet.service() for servlet [kylin] in context with
>> > path
>> > >>> >>> > [/kylin] threw exception [Handler processing failed; nested
>> > >>> >>>exception is
>> > >>> >>> > java.lang.NoClassDefFoundError:
>> > >>> >>> org/apache/hadoop/hive/ql/session/SessionState]
>> > >>> >>> > with root cause
>> > >>> >>> > java.lang.ClassNotFoundException: org.apache.hadoop.hive.ql.
>> > >>> >>> > session.SessionState
>> > >>> >>> >  at org.apache.catalina.loader.WebappClassLoader.loadClass(
>> > >>> >>> > WebappClassLoader.java:1720)
>> > >>> >>> >  at org.apache.catalina.loader.WebappClassLoader.loadClass(
>> > >>> >>> > WebappClassLoader.java:1571)
>> > >>> >>> >  at org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>> > >>> >>> > extractHiveTables(HiveSourceTableLoader.java:83)
>> > >>> >>> >  at
>> > >>> >>> org.apache.kylin.metadata.tool.HiveSourceTableLoader.
>> > >>> reloadHiveTables(
>> > >>> >>> > HiveSourceTableLoader.java:70)
>> > >>> >>> >  at org.apache.kylin.rest.service.
>> CubeService.reloadHiveTable(
>> > >>> >>> > CubeService.java:537)
>> > >>> >>> >  at org.apache.kylin.rest.service.
>> CubeService$$FastClassByCGLIB$
>> > >>> >>> > $17a07c0e.invoke(<generated>)
>> > >>> >>> >  at net.sf.cglib.proxy.MethodProxy.invoke(
>> MethodProxy.java:204)
>> > >>> >>> >
>> > >>> >>> > Regards,
>> > >>> >>> > Santosh Akhilesh
>> > >>> >>> > Bangalore R&D
>> > >>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> >
>> > >>> >>> > www.huawei.com
>> > >>> >>> > ------------------------------------------------------------
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>---------------------------------------------------------
>> > >>> ---------------
>> > >>> >>>-
>> > >>> >>> > This e-mail and its attachments contain confidential
>> information
>> > >>> from
>> > >>> >>> > HUAWEI, which
>> > >>> >>> > is intended only for the person or entity whose address is
>> listed
>> > >>> >>>above.
>> > >>> >>> > Any use of the
>> > >>> >>> > information contained herein in any way (including, but not
>> > limited
>> > >>> >>>to,
>> > >>> >>> > total or partial
>> > >>> >>> > disclosure, reproduction, or dissemination) by persons other
>> than
>> > >>> the
>> > >>> >>> > intended
>> > >>> >>> > recipient(s) is prohibited. If you receive this e-mail in
>> error,
>> > >>> >>>please
>> > >>> >>> > notify the sender by
>> > >>> >>> > phone or email immediately and delete it!
>> > >>> >>> >
>> > >>> >>> > ________________________________________
>> > >>> >>> > From: Santoshakhilesh [[email protected]]
>> > >>> >>> > Sent: Thursday, February 19, 2015 1:27 PM
>> > >>> >>> > To: [email protected]
>> > >>> >>> > Cc: Kulbhushan Rana
>> > >>> >>> > Subject: Hive Table is not loaded in Kylin//RE: About building
>> > >>> kylin
>> > >>> >>> >
>> > >>> >>> > Hi ,
>> > >>> >>> >
>> > >>> >>> >    I have resolved this issue. I have hivesingle column table
>> > and I
>> > >>> >>>can
>> > >>> >>> > access from Hive console and do queries on it sucecssfully
>> using
>> > >>> hive
>> > >>> >>> > console.
>> > >>> >>> >
>> > >>> >>> >    When I try to synch hive table using Kylin web specifiying
>> the
>> > >>> >>>hive
>> > >>> >>> > table name and synch.
>> > >>> >>> >
>> > >>> >>> >    I get an error dialog "Oops Failed to take action" .
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > Kylin.log
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > [pool-3-thread-1]:[2015-02-19 21:09:54,152][INFO][org.
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>apache.kylin.job.impl.threadpool.DefaultScheduler$
>> > >>> FetcherRunner.run(Defa
>> > >>> >>>ultScheduler.java:117)]
>> > >>> >>> > - Job Fetcher: 0 running, 0 actual running, 0 ready, 0 others
>> > >>> >>> > [http-bio-7070-exec-4]:[2015-02-19 21:09:59,556][INFO][org.
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>apache.kylin.rest.controller.TableController.
>> > >>> getHiveTables(TableControll
>> > >>> >>>er.java:74)]
>> > >>> >>> > - Return all table metadata in 0 seconds
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > localhost_access_log.2015-02-19.txt
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > 10.19.36.67 - - [19/Feb/2015:21:21:58 +0800] "POST
>> > >>> >>> > /kylin/api/tables/books/Sample HTTP/1.1" 500 1972
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > Here books is my table (I have used default database) and
>> sample
>> > is
>> > >>> >>>name
>> > >>> >>> > of my project
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > What could be reason of not loading the hive tables ? Someone
>> whi
>> > >>> is
>> > >>> >>> able
>> > >>> >>> > to load the hive table successfully can help me to debug this
>> > >>> issue ?
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> >
>> > >>> >>> > Regards,
>> > >>> >>> > Santosh Akhilesh
>> > >>> >>> > Bangalore R&D
>> > >>> >>> > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> >
>> > >>> >>> > www.huawei.com
>> > >>> >>> > ------------------------------------------------------------
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>---------------------------------------------------------
>> > >>> ---------------
>> > >>> >>>-
>> > >>> >>> > This e-mail and its attachments contain confidential
>> information
>> > >>> from
>> > >>> >>> > HUAWEI, which
>> > >>> >>> > is intended only for the person or entity whose address is
>> listed
>> > >>> >>>above.
>> > >>> >>> > Any use of the
>> > >>> >>> > information contained herein in any way (including, but not
>> > limited
>> > >>> >>>to,
>> > >>> >>> > total or partial
>> > >>> >>> > disclosure, reproduction, or dissemination) by persons other
>> than
>> > >>> the
>> > >>> >>> > intended
>> > >>> >>> > recipient(s) is prohibited. If you receive this e-mail in
>> error,
>> > >>> >>>please
>> > >>> >>> > notify the sender by
>> > >>> >>> > phone or email immediately and delete it!
>> > >>> >>> > ________________________________
>> > >>> >>> > From: Santosh Akhilesh [[email protected]]
>> > >>> >>> > Sent: Tuesday, February 17, 2015 8:13 PM
>> > >>> >>> > To: [email protected]
>> > >>> >>> > Subject: Re: About building kylin
>> > >>> >>> >
>> > >>> >>> > Hi,
>> > >>> >>> >      These days I spent time to learn Hive and create tables
>> > >>> required
>> > >>> >>> for
>> > >>> >>> > my test.
>> > >>> >>> >       Unfortunately when I am all set with Hive tables , Kylin
>> > >>> can't
>> > >>> >>> start
>> > >>> >>> > normally. Before this I could start and login to web ui. the
>> logs
>> > >>> are
>> > >>> >>> like
>> > >>> >>> > below. What has gone wring any idea ?
>> > >>> >>> >
>> > >>> >>> > SLF4J: Class path contains multiple SLF4J bindings.
>> > >>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>> > >>> >>> > frameworks/hbase-0.98.10-hadoop2/lib/slf4j-log4j12-1.6.
>> > >>> >>> > 4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > >>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>> > >>> >>> > frameworks/hadoop-1.2.1/lib/slf4j-log4j12-1.4.3.jar!/org/
>> > >>> >>> > slf4j/impl/StaticLoggerBinder.class]
>> > >>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>> > >>> >>> > frameworks/kylin/tomcat/webapps/kylin/WEB-INF/lib/
>> > >>> >>> > slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.
>> > class]
>> > >>> >>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings
>> for
>> > >>> an
>> > >>> >>> > explanation.
>> > >>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,307][WARN][org.
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>apache.kylin.common.KylinConfig.getKylinProperties(
>> > >>> KylinConfig.java:527)
>> > >>> >>>]
>> > >>> >>> > - KYLIN_CONF_HOME has not been set
>> > >>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,419][INFO][org.
>> > >>> >>> >
>> > >>> >>>springframework.core.io.support.PropertiesLoaderSupport.
>> > >>> loadProperties(
>> > >>> >>> > PropertiesLoaderSupport.java:177)] - Loading properties file
>> > from
>> > >>> >>> > resource loaded through InputStream
>> > >>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:41,815][WARN][org.
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>apache.kylin.common.KylinConfig.getKylinProperties(
>> > >>> KylinConfig.java:527)
>> > >>> >>>]
>> > >>> >>> > - KYLIN_CONF_HOME has not been set
>> > >>> >>> > SLF4J: Class path contains multiple SLF4J bindings.
>> > >>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>> > >>> >>> > frameworks/hbase-0.98.10-hadoop2/lib/slf4j-log4j12-1.6.
>> > >>> >>> > 4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > >>> >>> > SLF4J: Found binding in [jar:file:/home/santosh/work/
>> > >>> >>> > frameworks/hadoop-1.2.1/lib/slf4j-log4j12-1.4.3.jar!/org/
>> > >>> >>> > slf4j/impl/StaticLoggerBinder.class]
>> > >>> >>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings
>> for
>> > >>> an
>> > >>> >>> > explanation.
>> > >>> >>> > [localhost-startStop-1]:[2015-02-17 19:57:43,051][ERROR][org.
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>springframework.web.context.ContextLoader.
>> > >>> initWebApplicationContext(Cont
>> > >>> >>>extLoader.java:307)]
>> > >>> >>> > - Context initialization failed
>> > >>> >>> > org.springframework.beans.factory.BeanCreationException:
>> Error
>> > >>> >>>creating
>> > >>> >>> > bean with name 'com.ryantenney.metrics.spring.
>> > >>> >>> > MetricsBeanPostProcessorFactory#0': BeanPostProcessor before
>> > >>> >>> > instantiation of bean failed; nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name
>> > >>> >>> 'org.springframework.cache.config.internalCacheAdvisor':
>> > >>> >>> > Cannot resolve reference to bean
>> > >>> >>>'org.springframework.cache.annotation.
>> > >>> >>> > AnnotationCacheOperationSource#0' while setting bean property
>> > >>> >>> > 'cacheOperationSource'; nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name 'org.springframework.cache.
>> > >>> annotation.
>> > >>> >>> > AnnotationCacheOperationSource#0': BeanPostProcessor before
>> > >>> >>> instantiation
>> > >>> >>> > of bean failed; nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name 'org.springframework.security.
>> > >>> >>> > methodSecurityMetadataSourceAdvisor': Cannot resolve
>> reference
>> > to
>> > >>> >>>bean
>> > >>> >>> >
>> > >>> >>>
>> > >>> >>>'org.springframework.security.access.method.
>> > >>> DelegatingMethodSecurityMeta
>> > >>> >>>dataSource#0'
>> > >>> >>> > while setting constructor argument; nested exception is
>> > >>> >>> > org.springframework.beans.factory.BeanCreationException:
>> Error
>> > >>> >>>creating
>> > >>> >>> > bean with name 'org.springframework.security.access.method.
>> > >>> >>> > DelegatingMethodSecurityMetadataSource#0': Cannot create
>> inner
>> > >>> bean
>> > >>> >>> > '(inner bean)' of type [org.springframework.security.
>> > >>> access.prepost.
>> > >>> >>> > PrePostAnnotationSecurityMetadataSource] while setting
>> > constructor
>> > >>> >>> > argument with key [0]; nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name '(inner bean)': Cannot create
>> inner
>> > >>> >>>bean
>> > >>> >>> > '(inner bean)' of type [org.springframework.security.
>> > >>> >>> > access.expression.method.ExpressionBasedAnnotationAttri
>> > >>> buteFactory]
>> > >>> >>> while
>> > >>> >>> > setting constructor argument; nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name '(inner bean)': Cannot resolve
>> > >>> >>>reference
>> > >>> >>> to
>> > >>> >>> > bean 'expressionHandler' while setting constructor argument;
>> > nested
>> > >>> >>> > exception is org.springframework.beans.
>> > >>> factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name 'expressionHandler' defined in
>> > class
>> > >>> >>>path
>> > >>> >>> > resource [kylinSecurity.xml]: Cannot resolve reference to bean
>> > >>> >>> > 'permissionEvaluator' while setting bean property
>> > >>> >>>'permissionEvaluator';
>> > >>> >>> > nested exception is
>> > >>> >>> org.springframework.beans.factory.BeanCreationException:
>> > >>> >>> > Error creating bean with name 'permissionEvaluator' defined in
>> > >>> class
>> > >>> >>> path
>> > >>> >>> > resource [kylinSecurity.xml]: Cannot resolve reference to bean
>> > >>> >>> 'aclService'
>> > >>> >>> > while setting constructor argument; nested exception is
>> > >>> >>> > org.springframework.beans.factory.BeanCreationException:
>> Error
>> > >>> >>>creating
>> > >>> >>> > bean with name 'aclService' defined in file
>> [/home/santosh/work/
>> > >>> >>> > frameworks/kylin/tomcat/webapps/kylin/WEB-INF/classes/
>> > >>> >>> > org/apache/kylin/rest/service/AclService.class]:
>> Instantiation
>> > of
>> > >>> >>>bean
>> > >>> >>> > failed; nested exception is
>> > >>> >>> org.springframework.beans.BeanInstantiationException:
>> > >>> >>> > Could not instantiate bean class
>> > >>> >>> [org.apache.kylin.rest.service.AclService]:
>> > >>> >>> > Constructor threw exception; nested exception is
>> > >>> >>> org.apache.kylin.common.persistence.StorageException:
>> > >>> >>> > Error when open connection hbase
>> > >>> >>> > at org.springframework.beans.factory.support.
>> > >>> >>> > AbstractAutowireCapableBeanFactory.createBean(
>> > >>> >>> > AbstractAutowireCapableBeanFactory.java:452)
>> > >>> >>> > at org.springframework.beans.factory.support.
>> > >>> AbstractBeanFactory$1.
>> > >>> >>> > getObject(AbstractBeanFactory.java:294)
>> > >>> >>> > at
>> > >>> >>> org.springframework.beans.factory.support.
>> > >>> DefaultSingletonBeanRegistry.
>> > >>> >>> > getSingleton(DefaultSingletonBeanRegistry.java:225)
>> > >>> >>> > at org.springframework.beans.factory.support.
>> > >>> >>> > AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:291)
>> > >>> >>> > at
>> > >>> >>> org.springframework.beans.factory.support.
>> > >>> AbstractBeanFactory.getBean(
>> > >>> >>> > AbstractBeanFactory.java:197)
>> > >>> >>> > at org.springframework.context.support.
>> > AbstractApplicationContext.
>> > >>> >>> > registerBeanPostProcessors(AbstractApplicationContext.java:
>> 745)
>> > >>> >>> > at
>> > >>> >>> org.springframework.context.support.AbstractApplicationContext.
>> > >>> refresh(
>> > >>> >>> > AbstractApplicationContext.java:454)
>> > >>> >>> > at org.springframework.web.context.ContextLoader.
>> > >>> >>> > configureAndRefreshWebApplicationContext(ContextLoader.java:
>> 383)
>> > >>> >>> > at org.springframework.web.context.ContextLoader.
>> > >>> >>> > initWebApplicationContext(ContextLoader.java:283)
>> > >>> >>> > at org.springframework.web.context.ContextLoaderListener.
>> > >>> >>> > contextInitialized(ContextLoaderListener.java:111)
>> > >>> >>> > at org.apache.catalina.core.StandardContext.listenerStart(
>> > >>> >>> > StandardContext.java:5016)
>> > >>> >>> > at org.apache.catalina.core.StandardContext.startInternal(
>> > >>> >>> > StandardContext.java:5524)
>> > >>> >>> > at
>> > >>> >>>org.apache.catalina.util.LifecycleBase.start(
>> > LifecycleBase.java:150)
>> > >>> >>> > at org.apache.catalina.core.ContainerBase.addChildInternal(
>> > >>> >>> > ContainerBase.java:901)
>> > >>> >>> > at
>> > >>> >>> org.apache.catalina.core.ContainerBase.addChild(
>> > >>> ContainerBase.java:877)
>> > >>> >>> > at
>> > >>> >>>org.apache.catalina.core.StandardHost.addChild(
>> > StandardHost.java:649)
>> > >>> >>> > at
>> > >>> >>> org.apache.catalina.startup.HostConfig.deployWAR(
>> > >>> HostConfig.java:1081)
>> > >>> >>> > at org.apache.catalina.startup.HostConfig$DeployWar.run(
>> > >>> >>> > HostConfig.java:1877)
>> > >>> >>> > at
>> > >>> >>> java.util.concurrent.Executors$RunnableAdapter.
>> > >>> call(Executors.java:471)
>> > >>> >>> > at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>> > >>> >>> > at java.util.concurrent.ThreadPoolExecutor.runWorker(
>> > >>> >>> > ThreadPoolExecutor.java:1145)
>> > >>> >>> > at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>> > >>> >>> > ThreadPoolExecutor.java:615)
>> > >>> >>> > at java.lang.Thread.run(Thread.java:745)
>> > >>> >>> > Caused by: org.springframework.beans.
>> > >>> factory.BeanCreationException:
>> > >>> >>> Error
>> > >>> >>> > creating bean with name
>> > >>> >>> 'org.springframework.cache.config.internalCacheAdvisor':
>> > >>> >>> > Cannot resolve reference to bean
>> > >>> >>> >
>> > >>> >>> > On Fri Feb 13 2015 at 8:32:48 PM Zhou, Qianhao <
>> > [email protected]
>> > >>> >>> <mailto:
>> > >>> >>> > [email protected]>> wrote:
>> > >>> >>> > You should create your own table in hive.
>> > >>> >>> >
>> > >>> >>> > 2015年2月13日 下午10:47于 Santosh Akhilesh
>> > >>> >>><[email protected]<mailto:
>> > >>> >>> > [email protected]>>写道:
>> > >>> >>> > Oh thanks. You mean I should create my own table schema and
>> > insert
>> > >>> >>>data
>> > >>> >>> and
>> > >>> >>> > then need to import?
>> > >>> >>> > Does it mean I should create tables in hbase ? Any specific
>> > >>> location
>> > >>> >>>?
>> > >>> >>> I am
>> > >>> >>> > quite new to hbase and hive so may be I have many silly
>> > questions.
>> > >>> >>> > On Fri, 13 Feb 2015 at 8:05 pm, 周千昊 <[email protected]
>> <mailto:
>> > z.
>> > >>> >>> > [email protected]>> wrote:
>> > >>> >>> >
>> > >>> >>> > > Hi Santosh,
>> > >>> >>> > >      As you have setup a new hadoop environment, so there
>> is no
>> > >>> >>>data
>> > >>> >>> in
>> > >>> >>> > > hive. And we have not provided standalone data pack for this
>> > >>> >>> situation.
>> > >>> >>> > > That means you have to manually import the data into hive
>> > table.
>> > >>> >>> > >
>> > >>> >>> > > On Fri Feb 13 2015 at 10:26:57 PM Santosh Akhilesh <
>> > >>> >>> > > [email protected]<mailto:[email protected]
>> >>
>> > >>> wrote:
>> > >>> >>> > >
>> > >>> >>> > > > I have seen this link but my question is how to get EDW
>> data
>> > >>> >>>source
>> > >>> >>> as
>> > >>> >>> > > > mentioned in tutorial
>> > >>> >>> > > > On Fri, 13 Feb 2015 at 7:41 pm, 周千昊
>> > >>> >>><[email protected]<mailto:z.
>> > >>> >>> > [email protected]>> wrote:
>> > >>> >>> > > >
>> > >>> >>> > > > > Hi, Santosh
>> > >>> >>> > > > >      There is a tutorial about how to create a cube:
>> > >>> >>> > > > >      https://github.com/KylinOLAP/K
>> ylin/wiki/Kylin-Cube-
>> > >>> >>> > > > Creation-Tutorial
>> > >>> >>> > > > >
>> > >>> >>> > > > > On Fri Feb 13 2015 at 10:05:55 PM Santosh Akhilesh <
>> > >>> >>> > > > > [email protected]<mailto:santoshakhilesh@gmail.
>> com
>> > >>
>> > >>> >>> wrote:
>> > >>> >>> > > > >
>> > >>> >>> > > > > > Hi Zhou,
>> > >>> >>> > > > > > Yes I understand. I was wondering how do I use the
>> create
>> > >>> >>>cube
>> > >>> >>> > > > tutorial.
>> > >>> >>> > > > > If
>> > >>> >>> > > > > > there are some sample tables exist for me to try out
>> the
>> > >>> >>>sample
>> > >>> >>> > cube
>> > >>> >>> > > > > > described in tutorial. I am quite new to these stuffs
>> so
>> > if
>> > >>> >>>you
>> > >>> >>> can
>> > >>> >>> > > > give
>> > >>> >>> > > > > me
>> > >>> >>> > > > > > pointers about how do I create sample hive tables and
>> > >>> import
>> > >>> >>>it
>> > >>> >>> > would
>> > >>> >>> > > > be
>> > >>> >>> > > > > > great. Thanks for your quick responses. I know in
>> China
>> > you
>> > >>> >>>have
>> > >>> >>> > > spring
>> > >>> >>> > > > > > holidays but you guys still working :)
>> > >>> >>> > > > > > Regards
>> > >>> >>> > > > > > Santosh
>> > >>> >>> > > > > > On Fri, 13 Feb 2015 at 7:13 pm, Zhou, Qianhao <
>> > >>> >>> [email protected]
>> > >>> >>> > <mailto:[email protected]>>
>> > >>> >>> > > > wrote:
>> > >>> >>> > > > > >
>> > >>> >>> > > > > > > Hi Santosh
>> > >>> >>> > > > > > >     Kylin is not responsible for source of the data,
>> > >>> >>>however
>> > >>> >>> you
>> > >>> >>> > > can
>> > >>> >>> > > > > > > import table from your hive table, there is a tab in
>> > the
>> > >>> >>>web
>> > >>> >>> ui
>> > >>> >>> > > > called
>> > >>> >>> > > > > > > table, you can import the table there
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > 2015年2月13日 下午9:33于 Santoshakhilesh <
>> > >>> >>> [email protected]<
>> > >>> >>> > mailto:[email protected]>>
>> > >>> >>> > > 写道:
>> > >>> >>> > > > > > > I saw the following link about creating the cube.
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > https://github.com/KylinOLAP/Kylin/wiki/Kylin-Cube-
>> > >>> >>> > > Creation-Tutorial
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > In this there is already source table EDW exist ,
>> But
>> > in
>> > >>> my
>> > >>> >>> > > > environment
>> > >>> >>> > > > > > > there is no such table.
>> > >>> >>> > > > > > > How can I create this table manually or I need to
>> > export
>> > >>> >>>from
>> > >>> >>> > > > > somewhere ?
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Also , how do I insert tables , data in table when I
>> > >>> start
>> > >>> >>> > creating
>> > >>> >>> > > > my
>> > >>> >>> > > > > > own
>> > >>> >>> > > > > > > cubes and tables ?
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Regards,
>> > >>> >>> > > > > > > Santosh Akhilesh
>> > >>> >>> > > > > > > Bangalore R&D
>> > >>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > -------------
>> > >>> >>> > > > > > > This e-mail and its attachments contain confidential
>> > >>> >>> information
>> > >>> >>> > > from
>> > >>> >>> > > > > > > HUAWEI, which
>> > >>> >>> > > > > > > is intended only for the person or entity whose
>> address
>> > >>> is
>> > >>> >>> listed
>> > >>> >>> > > > > above.
>> > >>> >>> > > > > > > Any use of the
>> > >>> >>> > > > > > > information contained herein in any way (including,
>> but
>> > >>> not
>> > >>> >>> > limited
>> > >>> >>> > > > to,
>> > >>> >>> > > > > > > total or partial
>> > >>> >>> > > > > > > disclosure, reproduction, or dissemination) by
>> persons
>> > >>> >>>other
>> > >>> >>> than
>> > >>> >>> > > the
>> > >>> >>> > > > > > > intended
>> > >>> >>> > > > > > > recipient(s) is prohibited. If you receive this
>> e-mail
>> > in
>> > >>> >>> error,
>> > >>> >>> > > > please
>> > >>> >>> > > > > > > notify the sender by
>> > >>> >>> > > > > > > phone or email immediately and delete it!
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > ________________________________________
>> > >>> >>> > > > > > > From: Santoshakhilesh [[email protected]<
>> > >>> mailto:
>> > >>> >>> > [email protected]>]
>> > >>> >>> > > > > > > Sent: Friday, February 13, 2015 5:40 PM
>> > >>> >>> > > > > > > To:
>> > >>> >>>[email protected]<mailto:[email protected]
>> > >>> >>> .
>> > >>> >>> > apache.org>
>> > >>> >>> > > > > > > Subject: RE: About building kylin
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Hi Zhou,
>> > >>> >>> > > > > > >    Thanks a million, it was hbase issue , now I am
>> able
>> > >>> to
>> > >>> >>> login
>> > >>> >>> > to
>> > >>> >>> > > > > main
>> > >>> >>> > > > > > > interface page.
>> > >>> >>> > > > > > >    In the documents I see a section of building
>> cubes.
>> > >>> are
>> > >>> >>> there
>> > >>> >>> > > any
>> > >>> >>> > > > > > > sample cubles I can try with ?
>> > >>> >>> > > > > > >    Can you please tell me where can I get them and
>> > steps
>> > >>> to
>> > >>> >>> > import
>> > >>> >>> > > > and
>> > >>> >>> > > > > > > query them.
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Regards,
>> > >>> >>> > > > > > > Santosh Akhilesh
>> > >>> >>> > > > > > > Bangalore R&D
>> > >>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > -------------
>> > >>> >>> > > > > > > This e-mail and its attachments contain confidential
>> > >>> >>> information
>> > >>> >>> > > from
>> > >>> >>> > > > > > > HUAWEI, which
>> > >>> >>> > > > > > > is intended only for the person or entity whose
>> address
>> > >>> is
>> > >>> >>> listed
>> > >>> >>> > > > > above.
>> > >>> >>> > > > > > > Any use of the
>> > >>> >>> > > > > > > information contained herein in any way (including,
>> but
>> > >>> not
>> > >>> >>> > limited
>> > >>> >>> > > > to,
>> > >>> >>> > > > > > > total or partial
>> > >>> >>> > > > > > > disclosure, reproduction, or dissemination) by
>> persons
>> > >>> >>>other
>> > >>> >>> than
>> > >>> >>> > > the
>> > >>> >>> > > > > > > intended
>> > >>> >>> > > > > > > recipient(s) is prohibited. If you receive this
>> e-mail
>> > in
>> > >>> >>> error,
>> > >>> >>> > > > please
>> > >>> >>> > > > > > > notify the sender by
>> > >>> >>> > > > > > > phone or email immediately and delete it!
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > ________________________________________
>> > >>> >>> > > > > > > From: Santoshakhilesh [[email protected]<
>> > >>> mailto:
>> > >>> >>> > [email protected]>]
>> > >>> >>> > > > > > > Sent: Friday, February 13, 2015 2:38 PM
>> > >>> >>> > > > > > > To:
>> > >>> >>>[email protected]<mailto:[email protected]
>> > >>> >>> .
>> > >>> >>> > apache.org>
>> > >>> >>> > > > > > > Cc: Kulbhushan Rana
>> > >>> >>> > > > > > > Subject: RE: About building kylin
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Hi Zhou,
>> > >>> >>> > > > > > >     The exceptions I see are below , So it seems
>> Kylin
>> > >>> >>> instance
>> > >>> >>> > is
>> > >>> >>> > > > not
>> > >>> >>> > > > > > > deployed correctly;
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > WARNING: Failed to scan
>> > >>> >>> [file:/contrib/capacity-scheduler/*.jar]
>> > >>> >>> > > > from
>> > >>> >>> > > > > > > classloader hierarchy
>> > >>> >>> > > > > > > java.io.FileNotFoundException:
>> > >>> >>> /contrib/capacity-scheduler/*.jar
>> > >>> >>> > > (No
>> > >>> >>> > > > > > such
>> > >>> >>> > > > > > > file or directory)
>> > >>> >>> > > > > > >  at java.util.zip.ZipFile.open(Native Method)
>> > >>> >>> > > > > > >  at java.util.zip.ZipFile.<init>(ZipFile.java:215)
>> > >>> >>> > > > > > >  at java.util.zip.ZipFile.<init>(ZipFile.java:145)
>> > >>> >>> > > > > > >  at java.util.jar.JarFile.<init>(JarFile.java:154)
>> > >>> >>> > > > > > >  at java.util.jar.JarFile.<init>(JarFile.java:91)
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Feb 13, 2015 10:27:32 PM
>> org.apache.catalina.startup.
>> > >>> >>> > ContextConfig
>> > >>> >>> > > > > > > processResourceJARs
>> > >>> >>> > > > > > > SEVERE: Failed to processes JAR found at URL
>> > >>> >>> > > > > [jar:file:/contrib/capacity-
>> > >>> >>> > > > > > scheduler/*.jar!/]
>> > >>> >>> > > > > > > for static resources to be included in context with
>> > name
>> > >>> >>> > > > > > > [jar:file:/contrib/capacity-scheduler/*.jar!/]
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > resource loaded through InputStream
>> > >>> >>> > > > > > > [localhost-startStop-1]:[2015-02-13
>> > >>> >>>22:27:34,475][WARN][org.
>> > >>> >>> > > > > > > apache.kylin.common.KylinConfig.getKylinProperties(
>> > >>> >>> > > > > > KylinConfig.java:527)]
>> > >>> >>> > > > > > > - KYLIN_CONF_HOME has not been set
>> > >>> >>> > > > > > > 2015-02-13 22:27:34,919 INFO
>> [localhost-startStop-1]
>> > >>> >>> zookeeper.
>> > >>> >>> > > > > > RecoverableZooKeeper:
>> > >>> >>> > > > > > > Process identifier=hconnection-0x2b5a652e
>> connecting
>> > to
>> > >>> >>> > ZooKeeper
>> > >>> >>> > > > > > > ensemble=localhost:2181
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > 2015-02-13 22:27:34,958 INFO
>> [localhost-startStop-1-
>> > >>> >>> > > > > > SendThread(localhost:2181)]
>> > >>> >>> > > > > > > zookeeper.ClientCnxn: Opening socket connection to
>> > server
>> > >>> >>> > > > > > > localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to
>> > >>> >>> authenticate
>> > >>> >>> > > > using
>> > >>> >>> > > > > > > SASL (unknown error)
>> > >>> >>> > > > > > > 2015-02-13 22:27:34,965 WARN
>> [localhost-startStop-1-
>> > >>> >>> > > > > > SendThread(localhost:2181)]
>> > >>> >>> > > > > > > zookeeper.ClientCnxn: Session 0x0 for server null,
>> > >>> >>>unexpected
>> > >>> >>> > > error,
>> > >>> >>> > > > > > > closing socket connection and attempting reconnect
>> > >>> >>> > > > > > > java.net.ConnectException: Connection refused
>> > >>> >>> > > > > > >  at sun.nio.ch.SocketChannelImpl.c
>> heckConnect(Native
>> > >>> >>>Method)
>> > >>> >>> > > > > > >  at sun.nio.ch.SocketChannelImpl.finishConnect(
>> > >>> >>> > > > > > SocketChannelImpl.java:739)
>> > >>> >>> > > > > > >  at org.apache.zookeeper.ClientCnxnSocketNIO.
>> > >>> doTransport(
>> > >>> >>> > > > > > > ClientCnxnSocketNIO.java:361)
>> > >>> >>> > > > > > >  at org.apache.zookeeper.ClientCnxn$SendThread.run(
>> > >>> >>> > > > > ClientCnxn.java:1081)
>> > >>> >>> > > > > > > 2015-02-13 22:27:35,070 INFO
>> [localhost-startStop-1-
>> > >>> >>> > > > > > SendThread(localhost:2181)]
>> > >>> >>> > > > > > > zookeeper.ClientCnxn: Opening socket connection to
>> > server
>> > >>> >>> > > localhost/
>> > >>> >>> > > > > > > 127.0.0.1:2181<http://127.0.0.1:2181>. Will not
>> > attempt
>> > >>> to
>> > >>> >>> > authenticate using SASL
>> > >>> >>> > > (unknown
>> > >>> >>> > > > > > > error)
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > 2015-02-13 22:27:51,920 WARN
>> [localhost-startStop-1]
>> > >>> >>> zookeeper.
>> > >>> >>> > > > > > RecoverableZooKeeper:
>> > >>> >>> > > > > > > Possibly transient ZooKeeper, quorum=localhost:2181,
>> > >>> >>> > > > > > exception=org.apache.
>> > >>> >>> > > > > > > zookeeper.KeeperException$ConnectionLossException:
>> > >>> >>> > > KeeperErrorCode =
>> > >>> >>> > > > > > > ConnectionLoss for /hbase/hbaseid
>> > >>> >>> > > > > > > 2015-02-13 22:27:51,921 ERROR
>> [localhost-startStop-1]
>> > >>> >>> zookeeper.
>> > >>> >>> > > > > > RecoverableZooKeeper:
>> > >>> >>> > > > > > > ZooKeeper exists failed after 4 attempts
>> > >>> >>> > > > > > > 2015-02-13 22:27:51,921 WARN
>> [localhost-startStop-1]
>> > >>> >>> > > > zookeeper.ZKUtil:
>> > >>> >>> > > > > > > hconnection-0x2b5a652e, quorum=localhost:2181,
>> > >>> >>> baseZNode=/hbase
>> > >>> >>> > > > Unable
>> > >>> >>> > > > > to
>> > >>> >>> > > > > > > set watcher on znode (/hbase/hbaseid)
>> > >>> >>> > > > > > >
>> > >>> >>>org.apache.zookeeper.KeeperException$ConnectionLossException:
>> > >>> >>> > > > > > > KeeperErrorCode = ConnectionLoss for /hbase/hbaseid
>> > >>> >>> > > > > > >  at org.apache.zookeeper.KeeperException.create(
>> > >>> >>> > > > > KeeperException.java:99)
>> > >>> >>> > > > > > >  at org.apache.zookeeper.KeeperException.create(
>> > >>> >>> > > > > KeeperException.java:51)
>> > >>> >>> > > > > > >  at
>> > >>> >>>org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1045)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.zookeeper.
>> > >>> >>> > RecoverableZooKeeper.exists(
>> > >>> >>> > > > > > > RecoverableZooKeeper.java:222)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.
>> > >>> zookeeper.ZKUtil.checkExists(
>> > >>> >>> > > > > > ZKUtil.java:479)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.zookeeper.ZKClusterId.
>> > >>> >>> > > > readClusterIdZNode(
>> > >>> >>> > > > > > > ZKClusterId.java:65)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.client
>> .ZooKeeperRegistry.
>> > >>> >>> > getClusterId(
>> > >>> >>> > > > > > > ZooKeeperRegistry.java:83)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.client
>> .HConnectionManager$
>> > >>> >>> > > > > > > HConnectionImplementation.retrieveClusterId(
>> > >>> >>> > > > > HConnectionManager.java:897)
>> > >>> >>> > > > > > >  at org.apache.hadoop.hbase.client
>> .HConnectionManager$
>> > >>> >>> > > > > > >
>> > >>> >>>HConnectionImplementation.<init>(HConnectionManager.java:694)
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Regards,
>> > >>> >>> > > > > > > Santosh Akhilesh
>> > >>> >>> > > > > > > Bangalore R&D
>> > >>> >>> > > > > > > HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > >
>> > >>> >>>------------------------------------------------------------
>> > >>> >>> > > > > > -------------
>> > >>> >>> > > > > > > This e-mail and its attachments contain confidential
>> > >>> >>> information
>> > >>> >>> > > from
>> > >>> >>> > > > > > > HUAWEI, which
>> > >>> >>> > > > > > > is intended only for the person or entity whose
>> address
>> > >>> is
>> > >>> >>> listed
>> > >>> >>> > > > > above.
>> > >>> >>> > > > > > > Any use of the
>> > >>> >>> > > > > > > information contained herein in any way (including,
>> but
>> > >>> not
>> > >>> >>> > limited
>> > >>> >>> > > > to,
>> > >>> >>> > > > > > > total or partial
>> > >>> >>> > > > > > > disclosure, reproduction, or dissemination) by
>> persons
>> > >>> >>>other
>> > >>> >>> than
>> > >>> >>> > > the
>> > >>> >>> > > > > > > intended
>> > >>> >>> > > > > > > recipient(s) is prohibited. If you receive this
>> e-mail
>> > in
>> > >>> >>> error,
>> > >>> >>> > > > please
>> > >>> >>> > > > > > > notify the sender by
>> > >>> >>> > > > > > > phone or email immediately and delete it!
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > ________________________________________
>> > >>> >>> > > > > > > From: Zhou, Qianhao [[email protected]<mailto:
>> > >>> >>> [email protected]
>> > >>> >>> > >]
>> > >>> >>> > > > > > > Sent: Friday, February 13, 2015 2:07 PM
>> > >>> >>> > > > > > > To:
>> > >>> >>>[email protected]<mailto:[email protected]
>> > >>> >>> .
>> > >>> >>> > apache.org>
>> > >>> >>> > > > > > > Subject: Re: About building kylin
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Hi, Santosh
>> > >>> >>> > > > > > > Please check tomcat/logs/kylin.log to see if there
>> is
>> > any
>> > >>> >>> > exception
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > Best Regard
>> > >>> >>> > > > > > > Zhou QianHao
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > On 2/13/15, 4:24 PM, "Santoshakhilesh" <
>> > >>> >>> > > [email protected]<mailto:santosh.akhilesh@huawei.
>> com
>> > >>
>> > >>> >>> > > > > > > wrote:
>> > >>> >>> > > > > > >
>> > >>> >>> > > > > > > >Hi Zhou,
>> > >>> >>> > > > > > > >      I am able to access the
>> > >>> http://127.0.0.17070/kylin
>> > >>> >>> using
>> > >>> >>> > > > > firefox
>> > >>> >>> > > > > > > >      All I get is a blank page , its not 404 , it
>> > >>> means I
>> > >>> >>> could
>> > >>> >>> > > > > connect
>> > >>> >>> > > > > > > >to web server successfully but there is some issue
>> in
>> > >>> >>> > rendering. I
>> > >>> >>> > > > was
>> > >>> >>> > > > > > > >noty asked for user name and password.
>> > >>> >>> > > > > > > >       When I run start kylin shell file ,
>> following
>> > >>> >>>prompt
>> > >>> >>> > really
>> > >>> >>> > > > > means
>> > >>> >>> > > > > > > >everuthing went ok , or I should check some log
>> files
>> > to
>> > >>> >>>know
>> > >>> >>> > the
>> > >>> >>> > > > real
>> > >>> >>> > > > > > > >status ?
>> > >>> >>> > > > > > > >A new Kylin instance is started by root, stop it
>> using
>> > >>> >>> > > > "stop-kylin.sh"
>> > >>> >>> > > > > > > >Please visit http://<your_sandbox_ip>:7070/kylin
>> to
>> > >>> play
>> > >>> >>> with
>> > >>> >>> > the
>> > >>> >>> > > > > > cubes!
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >Regards,
>> > >>> >>> > > > > > > >Santosh Akhilesh
>> > >>> >>> > > > > > > >Bangalore R&D
>> > >>> >>> > > > > > > >HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > >
>> > >>> >>>>www.huawei.com<http://www.huawei.com><http://www.huawei.com>
>> > >>> >>> > > > > > >
>> > >>> >>>>-----------------------------------------------------------
>> > >>> >>> > > > > > > ---------------
>> > >>> >>> > > > > > >
>> > >>> >>>>-----------------------------------------------------------
>> > >>> >>> > > > > > > >This e-mail and its attachments contain
>> confidential
>> > >>> >>> information
>> > >>> >>> > > > from
>> > >>> >>> > > > > > > >HUAWEI, which
>> > >>> >>> > > > > > > >is intended only for the person or entity whose
>> > address
>> > >>> is
>> > >>> >>> > listed
>> > >>> >>> > > > > above.
>> > >>> >>> > > > > > > >Any use of the
>> > >>> >>> > > > > > > >information contained herein in any way (including,
>> > but
>> > >>> >>>not
>> > >>> >>> > > limited
>> > >>> >>> > > > > to,
>> > >>> >>> > > > > > > >total or partial
>> > >>> >>> > > > > > > >disclosure, reproduction, or dissemination) by
>> persons
>> > >>> >>>other
>> > >>> >>> > than
>> > >>> >>> > > > the
>> > >>> >>> > > > > > > >intended
>> > >>> >>> > > > > > > >recipient(s) is prohibited. If you receive this
>> e-mail
>> > >>> in
>> > >>> >>> error,
>> > >>> >>> > > > > please
>> > >>> >>> > > > > > > >notify the sender by
>> > >>> >>> > > > > > > >phone or email immediately and delete it!
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >________________________________________
>> > >>> >>> > > > > > > >From: Zhou, Qianhao [[email protected]<mailto:
>> > >>> >>> [email protected]
>> > >>> >>> > >]
>> > >>> >>> > > > > > > >Sent: Friday, February 13, 2015 1:04 PM
>> > >>> >>> > > > > > > >To: [email protected]<mailto:
>> > >>> >>> [email protected].
>> > >>> >>> > apache.org>
>> > >>> >>> > > > > > > >Cc: Kulbhushan Rana
>> > >>> >>> > > > > > > >Subject: Re: About building kylin
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >Hi, Santosh
>> > >>> >>> > > > > > > >    You will see a kylin.war inside
>> > >>> >>> > ${KYLIN_HOME}/tomcat/webapps.
>> > >>> >>> > > > That
>> > >>> >>> > > > > > is
>> > >>> >>> > > > > > > >the server war package. The tomcat version is
>> 7.0.59.
>> > >>> >>> > > > > > > >    If you want to bind 0.0.0.0:7070<
>> http://0.0.0.0:
>> > >>> 7070>,
>> > >>> >>> > please refer to the tomcat
>> > >>> >>> > > > > > > document,
>> > >>> >>> > > > > > > >there should be some configurations to let you do
>> > that.
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >Best Regard
>> > >>> >>> > > > > > > >Zhou QianHao
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >On 2/13/15, 1:46 PM, "Santoshakhilesh" <
>> > >>> >>> > > [email protected]<mailto:santosh.akhilesh@huawei.
>> com
>> > >
>> > >>> >>> > > > >
>> > >>> >>> > > > > > > >wrote:
>> > >>> >>> > > > > > > >
>> > >>> >>> > > > > > > >>Hi Zhou,
>> > >>> >>> > > > > > > >>     Due to some practical constarints , I have to
>> > >>> setup
>> > >>> >>> > clusster
>> > >>> >>> > > > > > > >>manually.
>> > >>> >>> > > > > > > >>     I have downloaded all the dependency like
>> > hadoop ,
>> > >>> >>> hbase ,
>> > >>> >>> > > > hive
>> > >>> >>> > > > > ,
>> > >>> >>> > > > > > > >>tomcat and I have run the check-env.sh after it
>> > passes
>> > >>> ,
>> > >>> >>>I
>> > >>> >>> have
>> > >>> >>> > > run
>> > >>> >>> > > > > the
>> > >>> >>> > > > > > > >>start-kylin.sh
>> > >>> >>> > > > > > > >>     Following is the log on console after running
>> > the
>> > >>> >>> > > > > > ./start-kylin.sh.
>> > >>> >>> > > > > > > >>     Does this mean everything is ok  or I should
>> > check
>> > >>> >>>some
>> > >>> >>> > > other
>> > >>> >>> > > > > > logs ?
>> > >>> >>> > > > > > > >>Currently I am unable to login to http://
>> > >>> >>> > <your_sandbox_ip>:7070/
>> > >>> >>> > > > > kylin
>> > >>> >>> > > > > > > >>from outside the machine.
>> > >>> >>> > > > > > > >>    Is it possible to bind it to 0.0.0.0 :7070 so
>> > that
>> > >>> I
>> > >>> >>>can
>> > >>> >>> > > > connect
>> > >>> >>> > > > > > from
>> > >>> >>> > > > > > > >>outside the lab network ? Where can I configure
>> this
>> > ?
>> > >>> >>> > > > > > > >>
>> > >>> >>> > > > > > > >> ./start-kylin.sh
>> > >>> >>> > > > > > > >>Checking KYLIN_HOME...
>> > >>> >>> > > > > > > >>KYLIN_HOME is set to /opt/kylinb
>> > >>> >>> > > > > > > >>Checking hbase...
>> > >>> >>> > > > > > > >>hbase check passed
>> > >>> >>> > > > > > > >>Checking hive...
>> > >>> >>> > > > > > > >>hive check passed
>> > >>> >>> > > > > > > >>Checking hadoop...
>> > >>> >>> > > > > > > >>hadoop check passed
>> > >>> >>> > > > > > > >>Logging initialized using configuration in
>> > >>> >>> > > > > > > >>jar:file:/opt/hive/apache-hi
>> ve-0.14.0-bin/lib/hive-
>> > >>> >>> > > > > > > common-0.14.0.jar!/hiv
>> > >>> >>> > > > > > > >>e
>> > >>> >>> > > > > > > >>-log4j.properties
>> > >>> >>> > > > > > > >>SLF4J: Class path contains multiple SLF4J
>> bindings.
>> > >>> >>> > > > > > > >>SLF4J: Found binding in
>> > >>> >>> > > > > > > >>[jar:file:/opt/hadoop/hadoop-2.6.0/share/hadoop/
>> > >>> >>> > > > > > > common/lib/slf4j-log4j12-
>> > >>> >>> > > > > > > >>1
>> > >>> >>> > > > > > > >>.7.5.jar!/org/slf4j/impl/Sta
>> ticLoggerBinder.class]
>> > >>> >>> > > > > > > >>SLF4J: Found binding in
>> > >>> >>> > > > > > > >>[jar:file:/opt/hive/apache-
>> > >>> hive-0.14.0-bin/lib/hive-jdbc-
>> > >>> >>> > > > > > > 0.14.0-standalon
>> > >>> >>> > > > > > > >>e
>> > >>> >>> > > > > > > >>.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> > >>> >>> > > > > > > >>SLF4J: See
>> > >>> >>> http://www.slf4j.org/codes.html#multiple_bindings
>> > >>> >>> > for
>> > >>> >>> > > > an
>> > >>> >>> > > > > > > >>explanation.
>> > >>> >>> > > > > > > >>SLF4J: Actual binding is of type [org.slf4j.impl.
>> > >>> >>> > > > Log4jLoggerFactory]
>> > >>> >>> > > > > > > >>hive dependency:
>> > >>> >>> > > > > > > >>/usr/hdp/current/hive-client/conf/:/hive/lib/*:/
>> > >>> >>> > > > > > > hive-hcatalog/share/hcata
>> > >>> >>> > > > > > > >>l
>> > >>> >>> > > > > > > >>og/*
>> > >>> >>> > > > > > > >>A new Kylin instance is started by root, stop it
>> > using
>> > >>> >>> > > > > "stop-kylin.sh"
>> > >>> >>> > > > > > > >>Please visit http://<your_sandbox_ip>:7070/kylin
>> to
>> > >>> play
>> > >>> >>> with
>> > >>> >>> > > the
>> > >>> >>> > > > > > cubes!
>> > >>> >>> > > > > > > >>(Useranme: ADMIN, Password: KYLIN)
>> > >>> >>> > > > > > > >>You can check the log at
>> ./../tomcat/logs/kylin.log
>> > >>> >>> > > > > > > >>
>> > >>> >>> > > > > > > >>Regards,
>> > >>> >>> > > > > > > >>Santosh Akhilesh
>> > >>> >>> > > > > > > >>Bangalore R&D
>> > >>> >>> > > > > > > >>HUAWEI TECHNOLOGIES CO.,LTD.
>> > >>> >>> > > > > > > >>
>> > >>> >>> > > > > > >
>> > >>> >>>>> <http://www.huawei.com>
>
>


-- 
Regards,
Santosh Akhilesh
+91-0-9845482201

Reply via email to