Hi dear all:
    I meet this exception when I run IT on sandbox2.4 or run the hive cmd on 
shell. Does any people meet the same problem?
    Here is a discussion about this problem but I think it isn't helpful. 
https://community.hortonworks.com/questions/224847/i-am-getting-errornull-value-not-allowed-for-multi.html
Best regards
yuzhang




==Hive cmd:==========
hive -e "USE default;


CREATE TABLE IF NOT EXISTS default.ci_inner_join_cube_global_dict
( dict_key STRING COMMENT '', 
dict_val INT COMMENT '' 
) 
COMMENT '' 
PARTITIONED BY (dict_column string) 
STORED AS TEXTFILE; 
DROP TABLE IF EXISTS 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by;
 
CREATE TABLE IF NOT EXISTS 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by
 
( 
 dict_key STRING COMMENT '' 
) 
COMMENT '' 
PARTITIONED BY (dict_column string) 
STORED AS SEQUENCEFILE 
;
INSERT OVERWRITE TABLE 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by
 
PARTITION (dict_column = 'TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP') 
SELECT
TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP 
FROM kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd
GROUP BY TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP 
;


" --hiveconf 
mapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.SnappyCodec 
--hiveconf dfs.replication=2 --hiveconf hive.exec.compress.output=true


================  


=======Error========
org.apache.atlas.AtlasServiceException: Metadata service API CREATE_ENTITY 
failed with status 400(Bad Request) Response Body ({"error":"Null value not 
allowed for multiplicty Multiplicity{lower=1, upper=1, 
isUnique=false}","stackTrace":"org.apache.atlas.typesystem.types.ValueConversionException$NullConversionException:
 Null value not allowed for multiplicty Multiplicity{lower=1, upper=1, 
isUnique=false}\n\tat 
org.apache.atlas.typesystem.types.DataTypes$PrimitiveType.convertNull(DataTypes.java:93)\n\tat
 
org.apache.atlas.typesystem.types.DataTypes$StringType.convert(DataTypes.java:469)\n\tat
 
org.apache.atlas.typesystem.types.DataTypes$StringType.convert(DataTypes.java:452)\n\tat
 
org.apache.atlas.typesystem.types.DataTypes$MapType.convert(DataTypes.java:606)\n\tat
 
org.apache.atlas.typesystem.types.DataTypes$MapType.convert(DataTypes.java:562)\n\tat
 
org.apache.atlas.typesystem.persistence.StructInstance.set(StructInstance.java:118)\n\tat
 org.apache.atlas.typesystem.types.ClassType.convert(ClassType.java:141)\n\tat 
org.apache.atlas.services.DefaultMetadataService.deserializeClassInstance(DefaultMetadataService.java:252)\n\tat
 
org.apache.atlas.services.DefaultMetadataService.createEntity(DefaultMetadataService.java:230)\n\tat
 
org.apache.atlas.web.resources.EntityResource.submit(EntityResource.java:96)\n\tat
 sun.reflect.GeneratedMethodAccessor70.invoke(Unknown Source)\n\tat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\tat
 java.lang.reflect.Method.invoke(Method.java:498)\n\tat 
com.sun.jersey.spi.container.JavaMethodInvokerFactory$1.invoke(JavaMethodInvokerFactory.java:60)\n\tat
 
com.sun.jersey.server.impl.model.method.dispatch.AbstractResourceMethodDispatchProvider$ResponseOutInvoker._dispatch(AbstractResourceMethodDispatchProvider.java:205)\n\tat
 
com.sun.jersey.server.impl.model.method.dispatch.ResourceJavaMethodDispatcher.dispatch(ResourceJavaMethodDispatcher.java:75)\n\tat
 
com.sun.jersey.server.impl.uri.rules.HttpMethodRule.accept(HttpMethodRule.java:288)\n\tat
 
com.sun.jersey.server.impl.uri.rules.ResourceClassRule.accept(ResourceClassRule.java:108)\n\tat
 
com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)\n\tat
 
com.sun.jersey.server.impl.uri.rules.RootResourceClassesRule.accept(RootResourceClassesRule.java:84)\n\tat
 
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1469)\n\tat
 
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1400)\n\tat
 
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1349)\n\tat
 
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1339)\n\tat
 
com.sun.jersey.spi.container.servlet.WebComponent.service(WebComponent.java:409)\n\tat
 
com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:558)\n\tat
 
com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:733)\n\tat
 javax.servlet.http.HttpServlet.service(HttpServlet.java:790)\n\tat 
com.google.inject.servlet.ServletDefinition.doServiceImpl(ServletDefinition.java:287)\n\tat
 
com.google.inject.servlet.ServletDefinition.doService(ServletDefinition.java:277)\n\tat
 
com.google.inject.servlet.ServletDefinition.service(ServletDefinition.java:182)\n\tat
 
com.google.inject.servlet.ManagedServletPipeline.service(ManagedServletPipeline.java:91)\n\tat
 
com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:85)\n\tat
 org.apache.atlas.web.filters.AuditFilter.doFilter(AuditFilter.java:67)\n\tat 
com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:82)\n\tat
 
com.google.inject.servlet.ManagedFilterPipeline.dispatch(ManagedFilterPipeline.java:119)\n\tat
 com.google.inject.servlet.GuiceFilter$1.call(GuiceFilter.java:133)\n\tat 
com.google.inject.servlet.GuiceFilter$1.call(GuiceFilter.java:130)\n\tat 
com.google.inject.servlet.GuiceFilter$Context.call(GuiceFilter.java:203)\n\tat 
com.google.inject.servlet.GuiceFilter.doFilter(GuiceFilter.java:130)\n\tat 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)\n\tat
 
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585)\n\tat
 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)\n\tat
 
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:577)\n\tat
 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:223)\n\tat
 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1127)\n\tat
 
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:515)\n\tat 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)\n\tat
 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1061)\n\tat
 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)\n\tat
 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:97)\n\tat
 org.eclipse.jetty.server.Server.handle(Server.java:499)\n\tat 
org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:310)\n\tat 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:257)\n\tat
 
org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:540)\n\tat
 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635)\n\tat
 
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555)\n\tat
 java.lang.Thread.run(Thread.java:748)\n"})
at org.apache.atlas.AtlasClient.callAPIWithResource(AtlasClient.java:365)
at org.apache.atlas.AtlasClient.callAPI(AtlasClient.java:370)
at org.apache.atlas.AtlasClient.createEntity(AtlasClient.java:210)
at 
org.apache.atlas.hive.bridge.HiveMetaStoreBridge.createInstance(HiveMetaStoreBridge.java:132)
at 
org.apache.atlas.hive.bridge.HiveMetaStoreBridge.registerTable(HiveMetaStoreBridge.java:323)
at org.apache.atlas.hive.hook.HiveHook.handleCreateTable(HiveHook.java:271)
at org.apache.atlas.hive.hook.HiveHook.fireAndForget(HiveHook.java:205)
at org.apache.atlas.hive.hook.HiveHook.run(HiveHook.java:172)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1585)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1254)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1118)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1108)
at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:216)
at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:168)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:379)
at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:314)
at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:711)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:684)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:624)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
at org.apache.hadoop.util.RunJar.main(RunJar.java:136)


The command is: 
hive -e "USE default;


CREATE TABLE IF NOT EXISTS default.ci_inner_join_cube_global_dict
( dict_key STRING COMMENT '', 
dict_val INT COMMENT '' 
) 
COMMENT '' 
PARTITIONED BY (dict_column string) 
STORED AS TEXTFILE; 
DROP TABLE IF EXISTS 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by;
 
CREATE TABLE IF NOT EXISTS 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by
 
( 
 dict_key STRING COMMENT '' 
) 
COMMENT '' 
PARTITIONED BY (dict_column string) 
STORED AS SEQUENCEFILE 
;
INSERT OVERWRITE TABLE 
kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd__group_by
 
PARTITION (dict_column = 'TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP') 
SELECT
TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP 
FROM kylin_intermediate_ci_inner_join_cube_325139ef_5dd0_01b4_ae61_bc4dcc99c2bd
GROUP BY TEST_KYLIN_FACT_TEST_COUNT_DISTINCT_BITMAP 
;


" --hiveconf 
mapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.SnappyCodec 
--hiveconf dfs.replication=2 --hiveconf hive.exec.compress.output=true
at 
org.apache.kylin.common.util.CliCommandExecutor.execute(CliCommandExecutor.java:96)
at 
org.apache.kylin.source.hive.CreateMrHiveDictStep.createMrHiveDict(CreateMrHiveDictStep.java:127)
at 
org.apache.kylin.source.hive.CreateMrHiveDictStep.doWork(CreateMrHiveDictStep.java:168)
at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2019-05-04 17:39:17,968 INFO  [Scheduler 1795467682 Job 
54e197e6-9b98-e8f6-ce38-e8c99e3538d6-298] execution.ExecutableManager:471 : job 
id:54e197e6-9b98-e8f6-ce38-e8c99e3538d6-01 from RUNNING to ERROR
2019-05-04 17:39:17,974 DEBUG [pool-4-thread-1] cachesync.Broadcaster:116 : 
Servers in the cluster: [localhost:7070]
2019-05-04 17:39:17,975 DEBUG [pool-4-thread-1] cachesync.Broadcaster:126 : 
Announcing new broadcast to all: BroadcastEvent{entity=execute_output, 
event=update, cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,976 ERROR [pool-5-thread-1] cachesync.Broadcaster:140 : 
Announce broadcast event failed, targetNode localhost:7070 broadcastEvent 
BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}, error msg: 
org.apache.http.conn.HttpHostConnectException: Connection to 
http://localhost:7070 refused
2019-05-04 17:39:17,976 ERROR [pool-5-thread-1] cachesync.Broadcaster:329 : 
Announce broadcast event exceeds retry limit, abandon targetNode localhost:7070 
broadcastEvent BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,980 DEBUG [pool-4-thread-1] cachesync.Broadcaster:116 : 
Servers in the cluster: [localhost:7070]
2019-05-04 17:39:17,980 DEBUG [pool-4-thread-1] cachesync.Broadcaster:126 : 
Announcing new broadcast to all: BroadcastEvent{entity=execute_output, 
event=update, cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,981 ERROR [pool-5-thread-1] cachesync.Broadcaster:140 : 
Announce broadcast event failed, targetNode localhost:7070 broadcastEvent 
BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}, error msg: 
org.apache.http.conn.HttpHostConnectException: Connection to 
http://localhost:7070 refused
2019-05-04 17:39:17,981 ERROR [pool-5-thread-1] cachesync.Broadcaster:329 : 
Announce broadcast event exceeds retry limit, abandon targetNode localhost:7070 
broadcastEvent BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,984 INFO  [Scheduler 1795467682 Job 
54e197e6-9b98-e8f6-ce38-e8c99e3538d6-298] execution.ExecutableManager:471 : job 
id:54e197e6-9b98-e8f6-ce38-e8c99e3538d6 from RUNNING to ERROR
2019-05-04 17:39:17,984 DEBUG [Scheduler 1795467682 Job 
54e197e6-9b98-e8f6-ce38-e8c99e3538d6-298] execution.AbstractExecutable:332 : no 
need to send email, user list is empty
2019-05-04 17:39:17,985 DEBUG [pool-4-thread-1] cachesync.Broadcaster:116 : 
Servers in the cluster: [localhost:7070]
2019-05-04 17:39:17,985 DEBUG [pool-4-thread-1] cachesync.Broadcaster:126 : 
Announcing new broadcast to all: BroadcastEvent{entity=execute_output, 
event=update, cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,986 ERROR [pool-5-thread-1] cachesync.Broadcaster:140 : 
Announce broadcast event failed, targetNode localhost:7070 broadcastEvent 
BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}, error msg: 
org.apache.http.conn.HttpHostConnectException: Connection to 
http://localhost:7070 refused
2019-05-04 17:39:17,986 ERROR [pool-5-thread-1] cachesync.Broadcaster:329 : 
Announce broadcast event exceeds retry limit, abandon targetNode localhost:7070 
broadcastEvent BroadcastEvent{entity=execute_output, event=update, 
cacheKey=54e197e6-9b98-e8f6-ce38-e8c99e3538d6}
2019-05-04 17:39:17,993 INFO  [FetcherRunner 1850834078-290] 
threadpool.DefaultFetcherRunner:85 : Job Fetcher: 0 should running, 0 actual 
running, 0 stopped, 0 ready, 0 already succeed, 1 error, 0 discarded, 0 others
2019-05-04 17:39:21,322 ERROR [main] provision.BuildCubeWithEngine:254 : error
java.lang.RuntimeException: The test 'testInnerJoinCube' is failed.
at 
org.apache.kylin.provision.BuildCubeWithEngine.runTestAndAssertSucceed(BuildCubeWithEngine.java:250)
at 
org.apache.kylin.provision.BuildCubeWithEngine.testCase(BuildCubeWithEngine.java:234)
at 
org.apache.kylin.provision.BuildCubeWithEngine.build(BuildCubeWithEngine.java:211)
at 
org.apache.kylin.provision.BuildCubeWithEngine.main(BuildCubeWithEngine.java:100)
Time elapsed: 618 sec - in org.apache.kylin.provision.BuildCubeWithEngine
2019-05-04 17:39:21,323 ERROR [main] provision.BuildCubeWithEngine:106 : error
java.lang.RuntimeException: The test 'testInnerJoinCube' is failed.
at 
org.apache.kylin.provision.BuildCubeWithEngine.runTestAndAssertSucceed(BuildCubeWithEngine.java:250)
at 
org.apache.kylin.provision.BuildCubeWithEngine.testCase(BuildCubeWithEngine.java:234)
at 
org.apache.kylin.provision.BuildCubeWithEngine.build(BuildCubeWithEngine.java:211)
at 
org.apache.kylin.provision.BuildCubeWithEngine.main(BuildCubeWithEngine.java:100)
2019-05-04 17:39:21,327 INFO  [Thread-13] util.ZKUtil:93 : Going to remove 1 
cached curator clients
2019-05-04 17:39:21,337 INFO  [close-hbase-conn] hbase.HBaseConnection:137 : 
Closing HBase connections...
2019-05-04 17:39:21,338 INFO  [close-hbase-conn] 
client.ConnectionManager$HConnectionImplementation:2068 : Closing master 
protocol: MasterService
2019-05-04 17:39:21,345 INFO  [Thread-13] util.ZKUtil:78 : CuratorFramework for 
zkString sandbox.hortonworks.com is removed due to EXPLICIT
2019-05-04 17:39:21,348 INFO  [close-hbase-conn] 
client.ConnectionManager$HConnectionImplementation:1676 : Closing zookeeper 
sessionid=0x16a7d4772350019
2019-05-04 17:39:21,349 INFO  [Curator-Framework-0] 
imps.CuratorFrameworkImpl:821 : backgroundOperationsLoop exiting
2019-05-04 17:39:21,352 INFO  [close-hbase-conn] zookeeper.ZooKeeper:693 : 
Session: 0x16a7d4772350019 closed
2019-05-04 17:39:21,352 INFO  [main-EventThread] zookeeper.ClientCnxn:522 : 
EventThread shut down for session: 0x16a7d4772350019
2019-05-04 17:39:21,367 INFO  [Thread-13] zookeeper.ZooKeeper:693 : Session: 
0x16a7d477235001a closed
2019-05-04 17:39:21,367 INFO  [main-EventThread] zookeeper.ClientCnxn:522 : 
EventThread shut down for session: 0x16a7d477235001a


==================


| |
yuzhang
|
|
shifengdefan...@163.com
|
签名由网易邮箱大师定制

Reply via email to