Hi Shaofeng ,

   I am using chrome , When I click on button to add measures , following is 
error on chrome console. When I try to save the cube there is no error in 
console. I just get a error dialog saying failed to take action and gives me 
the JSON cube schema.

Error on chrome debug console is as below;

 ReferenceError: CubeDescModel is not defined
    at h.$scope.addNewMeasure (scripts.min.0.js:15984)
    at scripts.min.0.js:180
    at scripts.min.0.js:197
    at h.$eval (scripts.min.0.js:119)
    at h.$apply (scripts.min.0.js:119)
    at HTMLButtonElement.<anonymous> (scripts.min.0.js:197)
    at HTMLButtonElement.m.event.dispatch (scripts.min.0.js:3)
    at HTMLButtonElement.r.handle (scripts.min.0.js:3)scripts.min.0.js:100 
(anonymous function)scripts.min.0.js:77 (anonymous 
function)scripts.min.0.js:119 h.$applyscripts.min.0.js:197 (anonymous 
function)scripts.min.0.js:3 m.event.dispatchscripts.min.0.js:3 r.handle

   About the hive table import , I got pass the run shell command exception but 
it still fails the hadoop log is;
2015-02-26 20:46:48,332 INFO [main] org.apache.hadoop.mapred.YarnChild: 
mapreduce.cluster.local.dir for child: 
/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/application_1424953091340_0002
2015-02-26 20:46:48,776 INFO [main] 
org.apache.hadoop.conf.Configuration.deprecation: session.id is deprecated. 
Instead, use dfs.metrics.session-id
2015-02-26 20:46:49,310 INFO [main] org.apache.hadoop.mapred.Task:  Using 
ResourceCalculatorProcessTree : [ ]
2015-02-26 20:46:49,386 FATAL [main] org.apache.hadoop.mapred.YarnChild: Error 
running child : java.lang.NoClassDefFoundError: 
org/apache/kylin/common/mr/KylinMapper
 at java.lang.ClassLoader.defineClass1(Native Method)
 at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
 at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
 at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
 at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
 at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
 at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
 at java.security.AccessController.doPrivileged(Native Method)
 at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
 at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
 at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)

tomcat logs:
usage: HiveColumnCardinalityJob
 -output <path>        Output path
 -table <table name>   The hive table name
[pool-4-thread-2]:[2015-02-26 
20:47:49,936][ERROR][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:64)]
 - error execute 
HadoopShellExecutable{id=d4730d26-7fe6-412e-9841-3288ab362c5b-00, name=null, 
state=RUNNING}
java.lang.IllegalArgumentException: No enum constant 
org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_MAPS
 at java.lang.Enum.valueOf(Enum.java:236)
 at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
 at 
org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
 at 
org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
 at org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
 at 
org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
 at org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
 at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
 at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
 at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
 at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
 at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
 at 
org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:134)
 at 
org.apache.kylin.job.hadoop.cardinality.HiveColumnCardinalityJob.run(HiveColumnCardinalityJob.java:114)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
 at 
org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
 at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:99)
 at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
 at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:99)
 at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)

Regards,
Santosh Akhilesh
Bangalore R&D
HUAWEI TECHNOLOGIES CO.,LTD.

www.huawei.com
-------------------------------------------------------------------------------------------------------------------------------------
This e-mail and its attachments contain confidential information from HUAWEI, 
which
is intended only for the person or entity whose address is listed above. Any 
use of the
information contained herein in any way (including, but not limited to, total 
or partial
disclosure, reproduction, or dissemination) by persons other than the intended
recipient(s) is prohibited. If you receive this e-mail in error, please notify 
the sender by
phone or email immediately and delete it!

________________________________________
From: Shi, Shaofeng [[email protected]]
Sent: Thursday, February 26, 2015 11:32 AM
To: [email protected]
Cc: Kulbhushan Rana
Subject: Re: Error while making cube & Measure option is not responding on GUI

Hi Santosh, hive table importing issue should not impact on cube saving.

If you couldn’t save the cube, firstly please check whether there is error
in the tomcat’s log; If not please check your web browser; We suggest use
Firefox (with firebug add-on) or Chrome, open the JS console (press F12)
and then operate web UI, check whether there is any error reported in
browser.


On 2/26/15, 1:08 PM, "Santoshakhilesh" <[email protected]> wrote:

>Hi Shaofeng ,
>   Thanks for replying.
>   Yes I am checking the yarn exception, But I find that this error comes
>while importing the hive table to kylin.
>   Even if this error comes , hive tables is exported successfully in
>kylin. Is this the reason why cube saving has failed ?
>
>   Next step when I go on creating the cube for following schema  I get
>error at last step while saving and  I am unable to add any measures ,
>clicking on measure option just dont pop up any dialog,
>
>I am using star schema with fact_sales as fact table and dim_* as
>dimension tables.
>
> fact_sales:
> storeid                 int
> itemid                  int
> custid                  int
> qty                     int
> price                   double
>
>dim_customer
> custid                  int
> name                    string
>
> dim_item
> itemid                  int
> category                string
> brand                   string
> color                   string
>
>dim_store
> storeid                 int
> city                    string
> state                   string
>
>JSON is as below.
> The JSON is as below.
>
> {
>   "name": "Retail_Cube",
>   "description": "",
>   "dimensions": [
>     {
>       "name": "RETAIL.FACT_SALES.STOREID",
>       "table": "RETAIL.FACT_SALES",
>       "hierarchy": false,
>       "derived": null,
>       "column": [
>         "STOREID"
>       ],
>       "id": 1
>     },
>     {
>       "name": "RETAIL.FACT_SALES.ITEMID",
>       "table": "RETAIL.FACT_SALES",
>       "hierarchy": false,
>       "derived": null,
>       "column": [
>         "ITEMID"
>       ],
>       "id": 2
>     },
>     {
>       "name": "RETAIL.FACT_SALES.CUSTID",
>       "table": "RETAIL.FACT_SALES",
>       "hierarchy": false,
>       "derived": null,
>       "column": [
>         "CUSTID"
>       ],
>       "id": 3
>     }
>   ],
>   "measures": [
>     {
>       "id": 1,
>       "name": "_COUNT_",
>       "function": {
>         "expression": "COUNT",
>         "returntype": "bigint",
>         "parameter": {
>           "type": "constant",
>           "value": 1
>         }
>       }
>     }
>   ],
>   "rowkey": {
>     "rowkey_columns": [
>       {
>         "column": "STOREID",
>         "length": 0,
>         "dictionary": "true",
>         "mandatory": false
>       },
>       {
>         "column": "ITEMID",
>         "length": 0,
>         "dictionary": "true",
>         "mandatory": false
>       },
>       {
>         "column": "CUSTID",
>         "length": 0,
>         "dictionary": "true",
>         "mandatory": false
>       }
>     ],
>     "aggregation_groups": [
>       [
>         "STOREID",
>         "ITEMID",
>         "CUSTID"
>       ]
>     ]
>   },
>   "notify_list": [],
>   "capacity": "",
>   "hbase_mapping": {
>     "column_family": [
>       {
>         "name": "f1",
>         "columns": [
>           {
>             "qualifier": "m",
>             "measure_refs": [
>               "_COUNT_"
>             ]
>           }
>         ]
>       }
>     ]
>   },
>   "project": "RetailProject",
>   "model_name": "Retail_Cube"
> }
>
>Regards,
>Santosh Akhilesh
>Bangalore R&D
>HUAWEI TECHNOLOGIES CO.,LTD.
>
>www.huawei.com
>--------------------------------------------------------------------------
>-----------------------------------------------------------
>This e-mail and its attachments contain confidential information from
>HUAWEI, which
>is intended only for the person or entity whose address is listed above.
>Any use of the
>information contained herein in any way (including, but not limited to,
>total or partial
>disclosure, reproduction, or dissemination) by persons other than the
>intended
>recipient(s) is prohibited. If you receive this e-mail in error, please
>notify the sender by
>phone or email immediately and delete it!
>
>________________________________________
>From: Shi, Shaofeng [[email protected]]
>Sent: Thursday, February 26, 2015 7:01 AM
>To: [email protected]
>Subject: Re: Error while making cube & Measure option is not responding
>on GUI
>
>Hi Santosh,
>
>It looks like hadoop failed to execute some shell command in the
>container; You need dive into hadoop to see what¹s the concrete error. You
>can use yarn logs command to fetch all logs:
>
>yarn logs -applicationId <app_id>
>
>
>On 2/25/15, 7:39 PM, "Santosh Akhilesh" <[email protected]> wrote:
>
>>Hi Luke / Shaofeng ,
>>           Can you please help me to check this issue.
>>Regards,
>>Santosh Akhilesh
>>
>>On Tue, Feb 24, 2015 at 10:41 PM, Santosh Akhilesh <
>>[email protected]> wrote:
>>
>>> Hi All ,
>>>         is it because of following error in map reduce job ? what could
>>>be
>>> way to resolve this , a google search says that its issue of Yarn class
>>> path , but I am not sure what it is ?
>>>
>>> Kylin Hive Column Cardinality Job table=RETAIL.FACT_SALES
>>> output=/tmp/cardinality/RETAIL.FACT_SALES
>>>
>>> Application application_1424791969399_0008 failed 2 times due to AM
>>> Container for appattempt_1424791969399_0008_000002 exited with
>>>exitCode: 1
>>> For more detailed output, check application tracking page:
>>> http://santosh:8088/proxy/application_1424791969399_0008/Then, click on
>>> links to logs of each attempt.
>>> Diagnostics: Exception from container-launch.
>>> Container id: container_1424791969399_0008_02_000001
>>> Exit code: 1
>>> Stack trace: ExitCodeException exitCode=1:
>>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
>>> at org.apache.hadoop.util.Shell.run(Shell.java:455)
>>> at
>>>
>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715
>>>)
>>> at
>>>
>>>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launc
>>>h
>>>Container(DefaultContainerExecutor.java:211)
>>> at
>>>
>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Cont
>>>a
>>>inerLaunch.call(ContainerLaunch.java:302)
>>> at
>>>
>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Cont
>>>a
>>>inerLaunch.call(ContainerLaunch.java:82)
>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>> at
>>>
>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav
>>>a
>>>:1145)
>>> at
>>>
>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja
>>>v
>>>a:615)
>>> at java.lang.Thread.run(Thread.java:745)
>>> Container exited with a non-zero exit code 1
>>> Failing this attempt. Failing the application.
>>>
>>> ---------- Forwarded message ----------
>>> From: Santoshakhilesh <[email protected]>
>>> Date: Tue, Feb 24, 2015 at 7:41 PM
>>> Subject: FW: Error while making cube & Measure option is not responding
>>>on
>>> GUI
>>> To: "[email protected]" <[email protected]>
>>>
>>>
>>> hi ,
>>>    please someone give me a hand to resolve this issue , thanks.
>>>
>>> Regards,
>>> Santosh Akhilesh
>>> Bangalore R&D
>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>
>>> www.huawei.com
>>>
>>>
>>>------------------------------------------------------------------------
>>>-
>>>------------------------------------------------------------
>>> This e-mail and its attachments contain confidential information from
>>> HUAWEI, which
>>> is intended only for the person or entity whose address is listed
>>>above.
>>> Any use of the
>>> information contained herein in any way (including, but not limited to,
>>> total or partial
>>> disclosure, reproduction, or dissemination) by persons other than the
>>> intended
>>> recipient(s) is prohibited. If you receive this e-mail in error, please
>>> notify the sender by
>>> phone or email immediately and delete it!
>>>
>>> ________________________________________
>>> From: Santoshakhilesh [[email protected]]
>>> Sent: Tuesday, February 24, 2015 12:55 PM
>>> To: [email protected]
>>> Cc: Kulbhushan Rana
>>> Subject: FW: Error while making cube & Measure option is not responding
>>>on
>>> GUI
>>>
>>> 2. If  I ignore and continue and try to save the cube I get an
>>>exception
>>> in Kylin.log , I have checked the path is set correctly and
>>>HCatInputFormat
>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>know
>>> what can I do to resolve this ?
>>>
>>>  -- This was path issue , now no more exception in kylin.log
>>>
>>> But saveing cube still fails with error. And still can't add measures.
>>>
>>> Error Message
>>> Failed to take action.
>>>
>>> In log I can find no exception. Following is the last log in kylin.log
>>>
>>> [pool-3-thread-1]:[2015-02-24
>>>
>>>20:47:15,613][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedule
>>>r
>>>$FetcherRunner.run(DefaultScheduler.java:117)]
>>> - Job Fetcher: 0 running, 0 actual running, 0 ready, 6 others
>>> [http-bio-7070-exec-2]:[2015-02-24
>>>
>>>20:47:51,610][DEBUG][org.apache.kylin.rest.controller.CubeController.des
>>>e
>>>rializeDataModelDesc(CubeController.java:459)]
>>> - Saving cube {
>>>   "name": "",
>>>   "fact_table": "RETAIL.FACT_SALES",
>>>   "lookups": [],
>>>   "filter_condition": "",
>>>   "capacity": "SMALL",
>>>   "partition_desc": {
>>>     "partition_date_column": "",
>>>     "partition_date_start": 0,
>>>     "partition_type": "APPEND"
>>>   },
>>>   "last_modified": 0
>>> }
>>>
>>>
>>> local access logs all with 200 , so seems ok.
>>>
>>> 10.18.146.105 - - [24/Feb/2015:20:46:56 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:07 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:27 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:28 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:34 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:48 +0800] "GET
>>> /kylin/api/user/authentication HTTP/1.1" 200 246
>>> 10.18.146.105 - - [24/Feb/2015:20:47:51 +0800] "POST /kylin/api/cubes
>>> HTTP/1.1" 200 701
>>>
>>>
>>> Regards,
>>> Santosh Akhilesh
>>> Bangalore R&D
>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>
>>> www.huawei.com
>>>
>>>
>>>------------------------------------------------------------------------
>>>-
>>>------------------------------------------------------------
>>> This e-mail and its attachments contain confidential information from
>>> HUAWEI, which
>>> is intended only for the person or entity whose address is listed
>>>above.
>>> Any use of the
>>> information contained herein in any way (including, but not limited to,
>>> total or partial
>>> disclosure, reproduction, or dissemination) by persons other than the
>>> intended
>>> recipient(s) is prohibited. If you receive this e-mail in error, please
>>> notify the sender by
>>> phone or email immediately and delete it!
>>>
>>> ________________________________________
>>> From: Santoshakhilesh [[email protected]]
>>> Sent: Tuesday, February 24, 2015 12:09 PM
>>> To: [email protected]
>>> Cc: Kulbhushan Rana
>>> Subject: Error while making cube & Measure option is not responding on
>>>GUI
>>>
>>> Hi All ,
>>>
>>>     I am building a simple cube for test and using the binary build
>>>0.7.1
>>> . I have following hive tables with columns.
>>>
>>>
>>>
>>> fact_sales:
>>>
>>> storeid                 int
>>> itemid                  int
>>> custid                  int
>>> qty                     int
>>> price                   double
>>>
>>> dim_customer
>>> custid                  int
>>> name                    string
>>>
>>> dim_item
>>>
>>> itemid                  int
>>> category                string
>>> brand                   string
>>> color                   string
>>>
>>> dim_store
>>>
>>> storeid                 int
>>> city                    string
>>> state                   string
>>>
>>> Please help me to answer following issues;
>>>
>>>
>>>
>>> 1. When I go to measure section and click on measure option , there is
>>>no
>>> response , I want add measure on qty and price with sum
>>>
>>> 2. If  I ignore and continue and try to save the cube I get an
>>>exception
>>> in Kylin.log , I have checked the path is set correctly and
>>>HCatInputFormat
>>> this file is present in hive-hcatalog-core-0.14.0.jar . Please let me
>>>know
>>> what can I do to resolve this ?
>>>
>>> 3. Also I have another question since this is a test and data is small
>>>I
>>> have not partitioned the fact table , is it ok to skip partition stage
>>> while cube build ?
>>>
>>>
>>>
>>> Exception
>>>
>>> pool-4-thread-4]:[2015-02-24
>>>
>>>19:26:32,577][ERROR][org.apache.kylin.job.impl.threadpool.DefaultSchedul
>>>e
>>>r$JobRunner.run(DefaultScheduler.java:134)]
>>> - ExecuteException job:c3532a6f-97ea-474a-b36a-218dd517cedb
>>> org.apache.kylin.job.exception.ExecuteException:
>>> org.apache.kylin.job.exception.ExecuteException:
>>> java.lang.NoClassDefFoundError:
>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>  at
>>>
>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecut
>>>a
>>>ble.java:102)
>>>  at
>>>
>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Defa
>>>u
>>>ltScheduler.java:132)
>>>  at
>>>
>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav
>>>a
>>>:1145)
>>>  at
>>>
>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja
>>>v
>>>a:615)
>>>  at java.lang.Thread.run(Thread.java:745)
>>> Caused by: org.apache.kylin.job.exception.ExecuteException:
>>> java.lang.NoClassDefFoundError:
>>> org/apache/hive/hcatalog/mapreduce/HCatInputFormat
>>>
>>>
>>>
>>> The JSON is as below.
>>>
>>>
>>>
>>> {
>>>   "name": "Retail_Cube",
>>>   "description": "",
>>>   "dimensions": [
>>>     {
>>>       "name": "RETAIL.FACT_SALES.STOREID",
>>>       "table": "RETAIL.FACT_SALES",
>>>       "hierarchy": false,
>>>       "derived": null,
>>>       "column": [
>>>         "STOREID"
>>>       ],
>>>       "id": 1
>>>     },
>>>     {
>>>       "name": "RETAIL.FACT_SALES.ITEMID",
>>>       "table": "RETAIL.FACT_SALES",
>>>       "hierarchy": false,
>>>       "derived": null,
>>>       "column": [
>>>         "ITEMID"
>>>       ],
>>>       "id": 2
>>>     },
>>>     {
>>>       "name": "RETAIL.FACT_SALES.CUSTID",
>>>       "table": "RETAIL.FACT_SALES",
>>>       "hierarchy": false,
>>>       "derived": null,
>>>       "column": [
>>>         "CUSTID"
>>>       ],
>>>       "id": 3
>>>     }
>>>   ],
>>>   "measures": [
>>>     {
>>>       "id": 1,
>>>       "name": "_COUNT_",
>>>       "function": {
>>>         "expression": "COUNT",
>>>         "returntype": "bigint",
>>>         "parameter": {
>>>           "type": "constant",
>>>           "value": 1
>>>         }
>>>       }
>>>     }
>>>   ],
>>>   "rowkey": {
>>>     "rowkey_columns": [
>>>       {
>>>         "column": "STOREID",
>>>         "length": 0,
>>>         "dictionary": "true",
>>>         "mandatory": false
>>>       },
>>>       {
>>>         "column": "ITEMID",
>>>         "length": 0,
>>>         "dictionary": "true",
>>>         "mandatory": false
>>>       },
>>>       {
>>>         "column": "CUSTID",
>>>         "length": 0,
>>>         "dictionary": "true",
>>>         "mandatory": false
>>>       }
>>>     ],
>>>     "aggregation_groups": [
>>>       [
>>>         "STOREID",
>>>         "ITEMID",
>>>         "CUSTID"
>>>       ]
>>>     ]
>>>   },
>>>   "notify_list": [],
>>>   "capacity": "",
>>>   "hbase_mapping": {
>>>     "column_family": [
>>>       {
>>>         "name": "f1",
>>>         "columns": [
>>>           {
>>>             "qualifier": "m",
>>>             "measure_refs": [
>>>               "_COUNT_"
>>>             ]
>>>           }
>>>         ]
>>>       }
>>>     ]
>>>   },
>>>   "project": "RetailProject",
>>>   "model_name": "Retail_Cube"
>>> }
>>>
>>>
>>>
>>>
>>>
>>> Regards,
>>> Santosh Akhilesh
>>> Bangalore R&D
>>> HUAWEI TECHNOLOGIES CO.,LTD.
>>>
>>> www.huawei.com
>>>
>>>
>>>------------------------------------------------------------------------
>>>-
>>>------------------------------------------------------------
>>> This e-mail and its attachments contain confidential information from
>>> HUAWEI, which
>>> is intended only for the person or entity whose address is listed
>>>above.
>>> Any use of the
>>> information contained herein in any way (including, but not limited to,
>>> total or partial
>>> disclosure, reproduction, or dissemination) by persons other than the
>>> intended
>>> recipient(s) is prohibited. If you receive this e-mail in error, please
>>> notify the sender by
>>> phone or email immediately and delete it!
>>>
>>>
>>>
>>> --
>>> Regards,
>>> Santosh Akhilesh
>>> +91-0-9845482201
>>>
>>
>>
>>
>>--
>>Regards,
>>Santosh Akhilesh
>>+91-0-9845482201

Reply via email to