Hi ShaoFeng Shi, Thanks for your reply, but here my case is different. I can able to create and build cube with same data using text formatted tables but when i tried to build cube with ORC formatted tables with same data facing an issue.
Regards, Suresh On Mon, Nov 28, 2016 at 7:31 PM, ShaoFeng Shi <shaofeng...@apache.org> wrote: > Hi Suresh, > > Another user also got similar problem, and I replied in the user@ group; > Just minute ago I forwarded it to dev@ group; please take a look and let > me > know whether it is the same: > http://apache-kylin.74782.x6.nabble.com/Fwd-Re-org-apache- > kylin-dict-TrieDictionary-Not-a-valid-value-td6428.html > > 2016-11-28 19:38 GMT+08:00 suresh m <suresh.m8...@gmail.com>: > > > can some see log and help me what is the exact issue facing with ORC > > formatted tables. Why i am unable build cube successfully with ORC > > formatted tables. > > > > On Mon, Nov 28, 2016 at 10:47 AM, suresh m <suresh.m8...@gmail.com> > wrote: > > > > > Please find detail as requested, > > > > > > Log Type: syslog > > > > > > Log Upload Time: Fri Nov 25 15:16:35 +0530 2016 > > > > > > Log Length: 107891 > > > > > > 2016-11-25 15:15:35,185 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Created MRAppMaster for application appattempt_1479580915733_0167_000001 > > > 2016-11-25 15:15:35,592 WARN [main] org.apache.hadoop.util. > NativeCodeLoader: > > Unable to load native-hadoop library for your platform... using > > builtin-java classes where applicable > > > 2016-11-25 15:15:35,630 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Executing with tokens: > > > 2016-11-25 15:15:35,956 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Kind: YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id > { > > id: 167 cluster_timestamp: 1479580915733 } attemptId: 1 } keyId: > 2128280969) > > > 2016-11-25 15:15:35,974 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Using mapred newApiCommitter. > > > 2016-11-25 15:15:35,976 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > OutputCommitter set in config null > > > 2016-11-25 15:15:36,029 INFO [main] org.apache.hadoop.mapreduce. > > lib.output.FileOutputCommitter: File Output Committer Algorithm version > > is 1 > > > 2016-11-25 15:15:36,029 INFO [main] org.apache.hadoop.mapreduce. > > lib.output.FileOutputCommitter: FileOutputCommitter skip cleanup > > _temporary folders under output directory:false, ignore cleanup failures: > > false > > > 2016-11-25 15:15:36,692 WARN [main] org.apache.hadoop.hdfs. > shortcircuit.DomainSocketFactory: > > The short-circuit local reads feature cannot be used because libhadoop > > cannot be loaded. > > > 2016-11-25 15:15:36,702 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > OutputCommitter is org.apache.hadoop.mapreduce. > > lib.output.FileOutputCommitter > > > 2016-11-25 15:15:36,891 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce.jobhistory.EventType for > > class org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler > > > 2016-11-25 15:15:36,891 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce. > v2.app.job.event.JobEventType > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > JobEventDispatcher > > > 2016-11-25 15:15:36,892 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce. > v2.app.job.event.TaskEventType > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > TaskEventDispatcher > > > 2016-11-25 15:15:36,893 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce.v2.app.job.event. > TaskAttemptEventType > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > TaskAttemptEventDispatcher > > > 2016-11-25 15:15:36,893 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce.v2.app.commit. > CommitterEventType > > for class org.apache.hadoop.mapreduce.v2.app.commit. > CommitterEventHandler > > > 2016-11-25 15:15:36,894 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce. > v2.app.speculate.Speculator$EventType > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > SpeculatorEventDispatcher > > > 2016-11-25 15:15:36,894 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce. > > v2.app.rm.ContainerAllocator$EventType for class > > org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter > > > 2016-11-25 15:15:36,895 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce.v2.app.launcher. > ContainerLauncher$EventType > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > ContainerLauncherRouter > > > 2016-11-25 15:15:36,923 INFO [main] org.apache.hadoop.mapreduce. > v2.jobhistory.JobHistoryUtils: > > Default file system is set solely by core-default.xml therefore - > ignoring > > > 2016-11-25 15:15:36,945 INFO [main] org.apache.hadoop.mapreduce. > v2.jobhistory.JobHistoryUtils: > > Default file system is set solely by core-default.xml therefore - > ignoring > > > 2016-11-25 15:15:36,967 INFO [main] org.apache.hadoop.mapreduce. > v2.jobhistory.JobHistoryUtils: > > Default file system is set solely by core-default.xml therefore - > ignoring > > > 2016-11-25 15:15:37,029 INFO [main] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Emitting job history data to the > > timeline server is not enabled > > > 2016-11-25 15:15:37,064 INFO [main] org.apache.hadoop.yarn.event. > AsyncDispatcher: > > Registering class org.apache.hadoop.mapreduce.v2.app.job.event. > JobFinishEvent$Type > > for class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > JobFinishEventHandler > > > 2016-11-25 15:15:37,204 WARN [main] org.apache.hadoop.metrics2. > impl.MetricsConfig: > > Cannot locate configuration: tried hadoop-metrics2-mrappmaster. > > properties,hadoop-metrics2.properties > > > 2016-11-25 15:15:37,300 INFO [main] org.apache.hadoop.metrics2. > impl.MetricsSystemImpl: > > Scheduled snapshot period at 10 second(s). > > > 2016-11-25 15:15:37,300 INFO [main] org.apache.hadoop.metrics2. > impl.MetricsSystemImpl: > > MRAppMaster metrics system started > > > 2016-11-25 15:15:37,308 INFO [main] org.apache.hadoop.mapreduce. > v2.app.job.impl.JobImpl: > > Adding job token for job_1479580915733_0167 to jobTokenSecretManager > > > 2016-11-25 15:15:37,452 INFO [main] org.apache.hadoop.mapreduce. > v2.app.job.impl.JobImpl: > > Not uberizing job_1479580915733_0167 because: not enabled; too much RAM; > > > 2016-11-25 15:15:37,468 INFO [main] org.apache.hadoop.mapreduce. > v2.app.job.impl.JobImpl: > > Input size for job job_1479580915733_0167 = 223589. Number of splits = 1 > > > 2016-11-25 15:15:37,469 INFO [main] org.apache.hadoop.mapreduce. > v2.app.job.impl.JobImpl: > > Number of reduces for job job_1479580915733_0167 = 1 > > > 2016-11-25 15:15:37,469 INFO [main] org.apache.hadoop.mapreduce. > v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from NEW to INITED > > > 2016-11-25 15:15:37,470 INFO [main] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > MRAppMaster launching normal, non-uberized, multi-container job > > job_1479580915733_0167. > > > 2016-11-25 15:15:37,493 INFO [main] org.apache.hadoop.ipc. > CallQueueManager: > > Using callQueue: class java.util.concurrent.LinkedBlockingQueue > > scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler > > > 2016-11-25 15:15:37,506 INFO [Socket Reader #1 for port 60945] > > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 60945 > > > 2016-11-25 15:15:37,525 INFO [main] org.apache.hadoop.yarn. > > factories.impl.pb.RpcServerFactoryPBImpl: Adding protocol > > org.apache.hadoop.mapreduce.v2.api.MRClientProtocolPB to the server > > > 2016-11-25 15:15:37,527 INFO [IPC Server Responder] > > org.apache.hadoop.ipc.Server: IPC Server Responder: starting > > > 2016-11-25 15:15:37,529 INFO [IPC Server listener on 60945] > > org.apache.hadoop.ipc.Server: IPC Server listener on 60945: starting > > > 2016-11-25 15:15:37,529 INFO [main] org.apache.hadoop.mapreduce. > v2.app.client.MRClientService: > > Instantiated MRClientService at hadoopclusterslic73.ad. > > infosys.com/10.122.97.73:60945 > > > 2016-11-25 <http://hadoopclusterslic73.ad.infosys.com/10.122.97.73: > > 609452016-11-25> 15:15:37,614 INFO [main] org.mortbay.log: Logging to > > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via > > org.mortbay.log.Slf4jLog > > > 2016-11-25 15:15:37,623 INFO [main] org.apache.hadoop.security. > > authentication.server.AuthenticationFilter: Unable to initialize > > FileSignerSecretProvider, falling back to use random secrets. > > > 2016-11-25 15:15:37,628 WARN [main] org.apache.hadoop.http. > HttpRequestLog: > > Jetty request log can only be enabled using Log4j > > > 2016-11-25 15:15:37,636 INFO [main] org.apache.hadoop.http. > HttpServer2: > > Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$ > > QuotingInputFilter) > > > 2016-11-25 15:15:37,688 INFO [main] org.apache.hadoop.http. > HttpServer2: > > Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn. > > server.webproxy.amfilter.AmIpFilter) to context mapreduce > > > 2016-11-25 15:15:37,688 INFO [main] org.apache.hadoop.http. > HttpServer2: > > Added filter AM_PROXY_FILTER (class=org.apache.hadoop.yarn. > > server.webproxy.amfilter.AmIpFilter) to context static > > > 2016-11-25 15:15:37,691 INFO [main] org.apache.hadoop.http. > HttpServer2: > > adding path spec: /mapreduce/* > > > 2016-11-25 15:15:37,692 INFO [main] org.apache.hadoop.http. > HttpServer2: > > adding path spec: /ws/* > > > 2016-11-25 15:15:38,181 INFO [main] org.apache.hadoop.yarn.webapp. > WebApps: > > Registered webapp guice modules > > > 2016-11-25 15:15:38,183 INFO [main] org.apache.hadoop.http. > HttpServer2: > > Jetty bound to port 34311 > > > 2016-11-25 15:15:38,183 INFO [main] org.mortbay.log: jetty-6.1.26.hwx > > > 2016-11-25 15:15:38,263 INFO [main] org.mortbay.log: Extract > > jar:file:/hadoop/yarn/local/filecache/16/mapreduce.tar.gz/ > > hadoop/share/hadoop/yarn/hadoop-yarn-common-2.7.3.2.5. > > 0.0-1245.jar!/webapps/mapreduce to /hadoop/yarn/local/usercache/ > > hdfs/appcache/application_1479580915733_0167/container_ > > e125_1479580915733_0167_01_000001/tmp/Jetty_0_0_0_0_ > > 34311_mapreduce____2ncvaf/webapp > > > 2016-11-25 15:15:39,882 INFO [main] org.mortbay.log: Started > HttpServer2$ > > SelectChannelConnectorWithSafeStartup@0.0.0.0:34311 > > > 2016-11-25 15:15:39,882 INFO [main] org.apache.hadoop.yarn.webapp. > WebApps: > > Web app mapreduce started at 34311 > > > 2016-11-25 15:15:39,933 INFO [main] org.apache.hadoop.ipc. > CallQueueManager: > > Using callQueue: class java.util.concurrent.LinkedBlockingQueue > > scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler > > > 2016-11-25 15:15:39,936 INFO [Socket Reader #1 for port 57220] > > org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 57220 > > > 2016-11-25 15:15:39,943 INFO [IPC Server listener on 57220] > > org.apache.hadoop.ipc.Server: IPC Server listener on 57220: starting > > > 2016-11-25 15:15:39,953 INFO [IPC Server Responder] > > org.apache.hadoop.ipc.Server: IPC Server Responder: starting > > > 2016-11-25 15:15:39,978 INFO [main] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: nodeBlacklistingEnabled:true > > > 2016-11-25 15:15:39,978 INFO [main] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: maxTaskFailuresPerNode is 3 > > > 2016-11-25 15:15:39,978 INFO [main] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: blacklistDisablePercent is 33 > > > 2016-11-25 15:15:40,082 WARN [main] org.apache.hadoop.ipc.Client: > Failed > > to connect to server: hadoopclusterslic71.ad. > infosys.com/10.122.97.71:8030: > > retries get failed due to exceeded maximum allowed retries number: 0 > > > java.net.ConnectException: Connection refused > > > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > > > at sun.nio.ch.SocketChannelImpl.finishConnect( > > SocketChannelImpl.java:717) > > > at org.apache.hadoop.net.SocketIOWithTimeout.connect( > > SocketIOWithTimeout.java:206) > > > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531) > > > at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495) > > > at org.apache.hadoop.ipc.Client$Connection.setupConnection( > > Client.java:650) > > > at org.apache.hadoop.ipc.Client$Connection.setupIOstreams( > > Client.java:745) > > > at org.apache.hadoop.ipc.Client$Connection.access$3200(Client. > > java:397) > > > at org.apache.hadoop.ipc.Client.getConnection(Client.java:1618) > > > at org.apache.hadoop.ipc.Client.call(Client.java:1449) > > > at org.apache.hadoop.ipc.Client.call(Client.java:1396) > > > at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker. > > invoke(ProtobufRpcEngine.java:233) > > > at com.sun.proxy.$Proxy80.registerApplicationMaster(Unknown > > Source) > > > at org.apache.hadoop.yarn.api.impl.pb.client. > > ApplicationMasterProtocolPBClientImpl.registerApplicationMaster( > > ApplicationMasterProtocolPBClientImpl.java:106) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at sun.reflect.NativeMethodAccessorImpl.invoke( > > NativeMethodAccessorImpl.java:62) > > > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > > DelegatingMethodAccessorImpl.java:43) > > > at java.lang.reflect.Method.invoke(Method.java:497) > > > at org.apache.hadoop.io.retry.RetryInvocationHandler. > invokeMethod( > > RetryInvocationHandler.java:278) > > > at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( > > RetryInvocationHandler.java:194) > > > at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke( > > RetryInvocationHandler.java:176) > > > at com.sun.proxy.$Proxy81.registerApplicationMaster(Unknown > > Source) > > > at org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator. > > register(RMCommunicator.java:160) > > > at org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator. > > serviceStart(RMCommunicator.java:121) > > > at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator. > > serviceStart(RMContainerAllocator.java:250) > > > at org.apache.hadoop.service.AbstractService.start( > > AbstractService.java:193) > > > at org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ > > ContainerAllocatorRouter.serviceStart(MRAppMaster.java:881) > > > at org.apache.hadoop.service.AbstractService.start( > > AbstractService.java:193) > > > at org.apache.hadoop.service.CompositeService.serviceStart( > > CompositeService.java:120) > > > at org.apache.hadoop.mapreduce.v2.app.MRAppMaster. > > serviceStart(MRAppMaster.java:1151) > > > at org.apache.hadoop.service.AbstractService.start( > > AbstractService.java:193) > > > at org.apache.hadoop.mapreduce.v2.app.MRAppMaster$5.run( > > MRAppMaster.java:1557) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapreduce.v2.app.MRAppMaster. > > initAndStartAppMaster(MRAppMaster.java:1553) > > > at org.apache.hadoop.mapreduce.v2.app.MRAppMaster.main( > > MRAppMaster.java:1486) > > > 2016-11-25 15:15:40,089 INFO [main] org.apache.hadoop.yarn.client. > > ConfiguredRMFailoverProxyProvider: Failing over to rm2 > > > 2016-11-25 15:15:40,183 INFO [main] org.apache.hadoop.mapreduce. > v2.app.rm.RMCommunicator: > > maxContainerCapability: <memory:28672, vCores:3> > > > 2016-11-25 15:15:40,183 INFO [main] org.apache.hadoop.mapreduce. > v2.app.rm.RMCommunicator: > > queue: default > > > 2016-11-25 15:15:40,186 INFO [main] org.apache.hadoop.mapreduce. > > v2.app.launcher.ContainerLauncherImpl: Upper limit on the thread pool > > size is 500 > > > 2016-11-25 15:15:40,186 INFO [main] org.apache.hadoop.mapreduce. > > v2.app.launcher.ContainerLauncherImpl: The thread pool initial size is > 10 > > > 2016-11-25 15:15:40,189 INFO [main] org.apache.hadoop.yarn.client. > > api.impl.ContainerManagementProtocolProxy: yarn.client.max-cached- > nodemanagers-proxies > > : 0 > > > 2016-11-25 15:15:40,202 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from INITED to SETUP > > > 2016-11-25 15:15:40,212 INFO [CommitterEvent Processor #0] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: JOB_SETUP > > > 2016-11-25 15:15:40,226 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from SETUP to RUNNING > > > 2016-11-25 15:15:40,291 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:40,328 INFO [eventHandlingThread] > > org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Event > > Writer setup for JobId: job_1479580915733_0167, File: > > hdfs://SLICHDP:8020/user/hdfs/.staging/job_1479580915733_ > > 0167/job_1479580915733_0167_1.jhist > > > 2016-11-25 15:15:40,351 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic73.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:40,357 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_m_000000 Task Transitioned from NEW to SCHEDULED > > > 2016-11-25 15:15:40,358 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_r_000000 Task Transitioned from NEW to SCHEDULED > > > 2016-11-25 15:15:40,359 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from NEW > > to UNASSIGNED > > > 2016-11-25 15:15:40,359 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_r_000000_0 TaskAttempt Transitioned from NEW > > to UNASSIGNED > > > 2016-11-25 15:15:40,401 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: mapResourceRequest:<memory:3072, > vCores:1> > > > 2016-11-25 15:15:40,416 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: reduceResourceRequest:<memory:4096, > > vCores:1> > > > 2016-11-25 15:15:41,191 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:0 ContRel:0 > > HostLocal:0 RackLocal:0 > > > 2016-11-25 15:15:41,225 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=4 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:38912, > vCores:1> > > knownNMs=2 > > > 2016-11-25 15:15:41,225 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:38912, vCores:1> > > > 2016-11-25 15:15:41,226 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:15:42,235 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got > allocated > > containers 1 > > > 2016-11-25 15:15:42,237 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > > container container_e125_1479580915733_0167_01_000002 to > > attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:42,238 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:34816, vCores:0> > > > 2016-11-25 15:15:42,238 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:15:42,238 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:15:42,286 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic73.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:42,311 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The job-jar > > file on the remote FS is hdfs://SLICHDP/user/hdfs/. > > staging/job_1479580915733_0167/job.jar > > > 2016-11-25 15:15:42,315 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: The > job-conf > > file on the remote FS is /user/hdfs/.staging/job_ > > 1479580915733_0167/job.xml > > > 2016-11-25 15:15:42,336 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Adding #0 > > tokens and #1 secret keys for NM use for launching container > > > 2016-11-25 15:15:42,336 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Size of > > containertokens_dob is 1 > > > 2016-11-25 15:15:42,336 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Putting > > shuffle token in serviceData > > > 2016-11-25 15:15:42,441 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from > > UNASSIGNED to ASSIGNED > > > 2016-11-25 15:15:42,455 INFO [ContainerLauncher #0] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container > > container_e125_1479580915733_0167_01_000002 taskAttempt > > attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:42,457 INFO [ContainerLauncher #0] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Launching attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:42,457 INFO [ContainerLauncher #0] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic73.ad.infosys.com:45454 > > > 2016-11-25 15:15:42,531 INFO [ContainerLauncher #0] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Shuffle port returned by ContainerManager for > attempt_1479580915733_0167_m_000000_0 > > : 13562 > > > 2016-11-25 15:15:42,533 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > TaskAttempt: > > [attempt_1479580915733_0167_m_000000_0] using containerId: > > [container_e125_1479580915733_0167_01_000002 on NM: [ > > hadoopclusterslic73.ad.infosys.com:45454] > > > 2016-11-25 15:15:42,536 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from > > ASSIGNED to RUNNING > > > 2016-11-25 15:15:42,536 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_m_000000 Task Transitioned from SCHEDULED to > > RUNNING > > > 2016-11-25 15:15:43,241 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=4 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:34816, > vCores:0> > > knownNMs=2 > > > 2016-11-25 15:15:44,790 INFO [Socket Reader #1 for port 57220] > > SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for > > job_1479580915733_0167 (auth:SIMPLE) > > > 2016-11-25 15:15:44,870 INFO [IPC Server handler 5 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : > > jvm_1479580915733_0167_m_137438953472002 asked for a task > > > 2016-11-25 15:15:44,870 INFO [IPC Server handler 5 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: > > jvm_1479580915733_0167_m_137438953472002 given task: > > attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:51,923 INFO [IPC Server handler 12 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_0 is : 0.667 > > > 2016-11-25 15:15:52,099 INFO [IPC Server handler 5 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_0 is : 0.667 > > > 2016-11-25 15:15:52,137 ERROR [IPC Server handler 12 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: > > attempt_1479580915733_0167_m_000000_0 - exited : java.io.IOException: > > Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:15:52,138 INFO [IPC Server handler 12 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report > from > > attempt_1479580915733_0167_m_000000_0: Error: java.io.IOException: > Failed > > to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:15:52,141 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_0: Error: > > java.io.IOException: Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:15:52,142 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from > > RUNNING to FAIL_CONTAINER_CLEANUP > > > 2016-11-25 15:15:52,155 INFO [ContainerLauncher #1] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container > > container_e125_1479580915733_0167_01_000002 taskAttempt > > attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:52,156 INFO [ContainerLauncher #1] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > KILLING attempt_1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:52,156 INFO [ContainerLauncher #1] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic73.ad.infosys.com:45454 > > > 2016-11-25 15:15:52,195 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from > > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP > > > 2016-11-25 15:15:52,204 INFO [CommitterEvent Processor #1] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: TASK_ABORT > > > 2016-11-25 15:15:52,215 WARN [CommitterEvent Processor #1] > > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not > > delete hdfs://SLICHDP/kylin/kylin_metadata/kylin-b2f43555-7105- > > 4912-b0bf-c40a0b405a05/ORC_ALERT_C/cuboid/_temporary/1/_ > temporary/attempt_ > > 1479580915733_0167_m_000000_0 > > > 2016-11-25 15:15:52,218 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_0 TaskAttempt Transitioned from > > FAIL_TASK_CLEANUP to FAILED > > > 2016-11-25 15:15:52,224 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:52,224 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic73.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:52,226 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from NEW > > to UNASSIGNED > > > 2016-11-25 15:15:52,226 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: 1 failures on node > hadoopclusterslic73.ad. > > infosys.com > > > 2016-11-25 15:15:52,230 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Added attempt_1479580915733_0167_m_ > 000000_1 > > to list of failed maps > > > 2016-11-25 15:15:52,291 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:15:52,299 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=1 resourcelimit=<memory:38912, > vCores:1> > > knownNMs=2 > > > 2016-11-25 15:15:52,299 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received > > completed container container_e125_1479580915733_0167_01_000002 > > > 2016-11-25 15:15:52,300 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:38912, vCores:1> > > > 2016-11-25 15:15:52,300 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_0: Container killed by > > the ApplicationMaster. > > > Container killed on request. Exit code is 143 > > > Container exited with a non-zero exit code 143 > > > > > > 2016-11-25 15:15:52,300 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:15:52,300 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:1 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:15:53,303 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got > allocated > > containers 1 > > > 2016-11-25 15:15:53,303 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning > > container Container: [ContainerId: container_e125_1479580915733_ > 0167_01_000003, > > NodeId: hadoopclusterslic72.ad.infosys.com:45454, NodeHttpAddress: > > hadoopclusterslic72.ad.infosys.com:8042, Resource: <memory:4096, > > vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: > > 10.122.97.72:45454 }, ] to fast fail map > > > 2016-11-25 15:15:53,303 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > from > > earlierFailedMaps > > > 2016-11-25 15:15:53,304 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > > container container_e125_1479580915733_0167_01_000003 to > > attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:15:53,304 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:34816, vCores:0> > > > 2016-11-25 15:15:53,304 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:15:53,304 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:15:53,304 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:15:53,304 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from > > UNASSIGNED to ASSIGNED > > > 2016-11-25 15:15:53,305 INFO [ContainerLauncher #2] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container > > container_e125_1479580915733_0167_01_000003 taskAttempt > > attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:15:53,305 INFO [ContainerLauncher #2] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Launching attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:15:53,305 INFO [ContainerLauncher #2] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:15:53,318 INFO [ContainerLauncher #2] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Shuffle port returned by ContainerManager for > attempt_1479580915733_0167_m_000000_1 > > : 13562 > > > 2016-11-25 15:15:53,318 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > TaskAttempt: > > [attempt_1479580915733_0167_m_000000_1] using containerId: > > [container_e125_1479580915733_0167_01_000003 on NM: [ > > hadoopclusterslic72.ad.infosys.com:45454] > > > 2016-11-25 15:15:53,318 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from > > ASSIGNED to RUNNING > > > 2016-11-25 15:15:54,309 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:34816, > vCores:0> > > knownNMs=2 > > > 2016-11-25 15:15:55,797 INFO [Socket Reader #1 for port 57220] > > SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for > > job_1479580915733_0167 (auth:SIMPLE) > > > 2016-11-25 15:15:55,819 INFO [IPC Server handler 13 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : > > jvm_1479580915733_0167_m_137438953472003 asked for a task > > > 2016-11-25 15:15:55,819 INFO [IPC Server handler 13 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: > > jvm_1479580915733_0167_m_137438953472003 given task: > > attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:16:02,857 INFO [IPC Server handler 8 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_1 is : 0.667 > > > 2016-11-25 15:16:03,332 INFO [IPC Server handler 13 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_1 is : 0.667 > > > 2016-11-25 15:16:03,347 ERROR [IPC Server handler 8 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: > > attempt_1479580915733_0167_m_000000_1 - exited : java.io.IOException: > > Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:03,347 INFO [IPC Server handler 8 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report > from > > attempt_1479580915733_0167_m_000000_1: Error: java.io.IOException: > Failed > > to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:03,349 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_1: Error: > > java.io.IOException: Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:03,350 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from > > RUNNING to FAIL_CONTAINER_CLEANUP > > > 2016-11-25 15:16:03,351 INFO [ContainerLauncher #3] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container > > container_e125_1479580915733_0167_01_000003 taskAttempt > > attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:16:03,355 INFO [ContainerLauncher #3] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > KILLING attempt_1479580915733_0167_m_000000_1 > > > 2016-11-25 15:16:03,355 INFO [ContainerLauncher #3] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:16:03,369 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from > > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP > > > 2016-11-25 15:16:03,369 INFO [CommitterEvent Processor #2] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: TASK_ABORT > > > 2016-11-25 15:16:03,375 WARN [CommitterEvent Processor #2] > > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not > > delete hdfs://SLICHDP/kylin/kylin_metadata/kylin-b2f43555-7105- > > 4912-b0bf-c40a0b405a05/ORC_ALERT_C/cuboid/_temporary/1/_ > temporary/attempt_ > > 1479580915733_0167_m_000000_1 > > > 2016-11-25 15:16:03,375 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_1 TaskAttempt Transitioned from > > FAIL_TASK_CLEANUP to FAILED > > > 2016-11-25 15:16:03,375 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:03,375 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic73.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:03,376 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: 1 failures on node > hadoopclusterslic72.ad. > > infosys.com > > > 2016-11-25 15:16:03,376 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from NEW > > to UNASSIGNED > > > 2016-11-25 15:16:03,380 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Added attempt_1479580915733_0167_m_ > 000000_2 > > to list of failed maps > > > 2016-11-25 15:16:04,341 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:2 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:16:04,344 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:34816, > vCores:0> > > knownNMs=2 > > > 2016-11-25 15:16:04,344 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:34816, vCores:0> > > > 2016-11-25 15:16:04,344 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:16:05,352 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received > > completed container container_e125_1479580915733_0167_01_000003 > > > 2016-11-25 15:16:05,352 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got > allocated > > containers 1 > > > 2016-11-25 15:16:05,352 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning > > container Container: [ContainerId: container_e125_1479580915733_ > 0167_01_000004, > > NodeId: hadoopclusterslic72.ad.infosys.com:45454, NodeHttpAddress: > > hadoopclusterslic72.ad.infosys.com:8042, Resource: <memory:4096, > > vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: > > 10.122.97.72:45454 }, ] to fast fail map > > > 2016-11-25 15:16:05,352 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > from > > earlierFailedMaps > > > 2016-11-25 15:16:05,352 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_1: Container killed by > > the ApplicationMaster. > > > Container killed on request. Exit code is 143 > > > Container exited with a non-zero exit code 143 > > > > > > 2016-11-25 15:16:05,353 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > > container container_e125_1479580915733_0167_01_000004 to > > attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:05,353 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:34816, vCores:0> > > > 2016-11-25 15:16:05,353 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:16:05,353 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:16:05,353 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:05,354 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from > > UNASSIGNED to ASSIGNED > > > 2016-11-25 15:16:05,356 INFO [ContainerLauncher #4] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container > > container_e125_1479580915733_0167_01_000004 taskAttempt > > attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:05,356 INFO [ContainerLauncher #4] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Launching attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:05,357 INFO [ContainerLauncher #4] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:16:05,371 INFO [ContainerLauncher #4] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Shuffle port returned by ContainerManager for > attempt_1479580915733_0167_m_000000_2 > > : 13562 > > > 2016-11-25 15:16:05,371 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > TaskAttempt: > > [attempt_1479580915733_0167_m_000000_2] using containerId: > > [container_e125_1479580915733_0167_01_000004 on NM: [ > > hadoopclusterslic72.ad.infosys.com:45454] > > > 2016-11-25 15:16:05,372 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from > > ASSIGNED to RUNNING > > > 2016-11-25 15:16:06,362 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:34816, > vCores:0> > > knownNMs=2 > > > 2016-11-25 15:16:07,537 INFO [Socket Reader #1 for port 57220] > > SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for > > job_1479580915733_0167 (auth:SIMPLE) > > > 2016-11-25 15:16:07,567 INFO [IPC Server handler 24 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : > > jvm_1479580915733_0167_m_137438953472004 asked for a task > > > 2016-11-25 15:16:07,567 INFO [IPC Server handler 24 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: > > jvm_1479580915733_0167_m_137438953472004 given task: > > attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:14,753 INFO [IPC Server handler 6 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_2 is : 0.667 > > > 2016-11-25 15:16:15,241 INFO [IPC Server handler 13 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_2 is : 0.667 > > > 2016-11-25 15:16:15,258 ERROR [IPC Server handler 8 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: > > attempt_1479580915733_0167_m_000000_2 - exited : java.io.IOException: > > Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:15,258 INFO [IPC Server handler 8 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report > from > > attempt_1479580915733_0167_m_000000_2: Error: java.io.IOException: > Failed > > to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:15,261 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_2: Error: > > java.io.IOException: Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:15,273 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from > > RUNNING to FAIL_CONTAINER_CLEANUP > > > 2016-11-25 15:16:15,274 INFO [ContainerLauncher #5] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container > > container_e125_1479580915733_0167_01_000004 taskAttempt > > attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:15,274 INFO [ContainerLauncher #5] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > KILLING attempt_1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:15,274 INFO [ContainerLauncher #5] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:16:15,289 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from > > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP > > > 2016-11-25 15:16:15,292 INFO [CommitterEvent Processor #3] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: TASK_ABORT > > > 2016-11-25 15:16:15,300 WARN [CommitterEvent Processor #3] > > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not > > delete hdfs://SLICHDP/kylin/kylin_metadata/kylin-b2f43555-7105- > > 4912-b0bf-c40a0b405a05/ORC_ALERT_C/cuboid/_temporary/1/_ > temporary/attempt_ > > 1479580915733_0167_m_000000_2 > > > 2016-11-25 15:16:15,300 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_2 TaskAttempt Transitioned from > > FAIL_TASK_CLEANUP to FAILED > > > 2016-11-25 15:16:15,301 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:15,301 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic73.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:15,301 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from NEW > > to UNASSIGNED > > > 2016-11-25 15:16:15,301 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: 2 failures on node > hadoopclusterslic72.ad. > > infosys.com > > > 2016-11-25 15:16:15,307 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Added attempt_1479580915733_0167_m_ > 000000_3 > > to list of failed maps > > > 2016-11-25 15:16:15,412 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Before > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:16:15,420 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=1 resourcelimit=<memory:38912, > vCores:1> > > knownNMs=2 > > > 2016-11-25 15:16:15,420 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Received > > completed container container_e125_1479580915733_0167_01_000004 > > > 2016-11-25 15:16:15,420 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:38912, vCores:1> > > > 2016-11-25 15:16:15,420 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:16:15,420 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:1 ScheduledReds:0 AssignedMaps:0 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:3 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:16:15,421 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_2: Container killed by > > the ApplicationMaster. > > > Container killed on request. Exit code is 143 > > > Container exited with a non-zero exit code 143 > > > > > > 2016-11-25 15:16:16,432 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Got > allocated > > containers 1 > > > 2016-11-25 15:16:16,432 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigning > > container Container: [ContainerId: container_e125_1479580915733_ > 0167_01_000005, > > NodeId: hadoopclusterslic72.ad.infosys.com:45454, NodeHttpAddress: > > hadoopclusterslic72.ad.infosys.com:8042, Resource: <memory:4096, > > vCores:1>, Priority: 5, Token: Token { kind: ContainerToken, service: > > 10.122.97.72:45454 }, ] to fast fail map > > > 2016-11-25 15:16:16,432 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > from > > earlierFailedMaps > > > 2016-11-25 15:16:16,433 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Assigned > > container container_e125_1479580915733_0167_01_000005 to > > attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:16,433 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: > Recalculating > > schedule, headroom=<memory:34816, vCores:0> > > > 2016-11-25 15:16:16,433 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Reduce slow > > start threshold not met. completedMapsForReduceSlowstart 1 > > > 2016-11-25 15:16:16,433 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: After > > Scheduling: PendingReds:1 ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 > > AssignedReds:0 CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 > > HostLocal:1 RackLocal:0 > > > 2016-11-25 15:16:16,433 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.yarn.util.RackResolver: Resolved > hadoopclusterslic72.ad. > > infosys.com to /default-rack > > > 2016-11-25 15:16:16,434 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from > > UNASSIGNED to ASSIGNED > > > 2016-11-25 15:16:16,436 INFO [ContainerLauncher #6] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_LAUNCH for container > > container_e125_1479580915733_0167_01_000005 taskAttempt > > attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:16,436 INFO [ContainerLauncher #6] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Launching attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:16,436 INFO [ContainerLauncher #6] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:16:16,516 INFO [ContainerLauncher #6] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Shuffle port returned by ContainerManager for > attempt_1479580915733_0167_m_000000_3 > > : 13562 > > > 2016-11-25 15:16:16,517 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > TaskAttempt: > > [attempt_1479580915733_0167_m_000000_3] using containerId: > > [container_e125_1479580915733_0167_01_000005 on NM: [ > > hadoopclusterslic72.ad.infosys.com:45454] > > > 2016-11-25 15:16:16,517 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from > > ASSIGNED to RUNNING > > > 2016-11-25 15:16:17,436 INFO [RMCommunicator Allocator] > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: > > getResources() for application_1479580915733_0167: ask=1 release= 0 > > newContainers=0 finishedContainers=0 resourcelimit=<memory:34816, > vCores:0> > > knownNMs=2 > > > 2016-11-25 15:16:19,664 INFO [Socket Reader #1 for port 57220] > > SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for > > job_1479580915733_0167 (auth:SIMPLE) > > > 2016-11-25 15:16:19,692 INFO [IPC Server handler 6 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : > > jvm_1479580915733_0167_m_137438953472005 asked for a task > > > 2016-11-25 15:16:19,692 INFO [IPC Server handler 6 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: > > jvm_1479580915733_0167_m_137438953472005 given task: > > attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:27,222 INFO [IPC Server handler 13 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_3 is : 0.667 > > > 2016-11-25 15:16:27,952 INFO [IPC Server handler 7 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of > TaskAttempt > > attempt_1479580915733_0167_m_000000_3 is : 0.667 > > > 2016-11-25 15:16:27,971 ERROR [IPC Server handler 11 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: > > attempt_1479580915733_0167_m_000000_3 - exited : java.io.IOException: > > Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:27,971 INFO [IPC Server handler 11 on 57220] > > org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report > from > > attempt_1479580915733_0167_m_000000_3: Error: java.io.IOException: > Failed > > to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:27,974 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics > > report from attempt_1479580915733_0167_m_000000_3: Error: > > java.io.IOException: Failed to build cube in mapper 0 > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:145) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:149) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:422) > > > at org.apache.hadoop.security.UserGroupInformation.doAs( > > UserGroupInformation.java:1724) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162) > > > Caused by: java.util.concurrent.ExecutionException: > > java.lang.RuntimeException: java.io.IOException: java.io.IOException: > > java.lang.IllegalArgumentException: Value not exists! > > > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > > > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > > > at org.apache.kylin.engine.mr.steps.InMemCuboidMapper. > > cleanup(InMemCuboidMapper.java:143) > > > ... 8 more > > > Caused by: java.lang.RuntimeException: java.io.IOException: > > java.io.IOException: java.lang.IllegalArgumentException: Value not > exists! > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:82) > > > at java.util.concurrent.Executors$RunnableAdapter. > > call(Executors.java:511) > > > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > > > at java.util.concurrent.ThreadPoolExecutor.runWorker( > > ThreadPoolExecutor.java:1142) > > > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > > ThreadPoolExecutor.java:617) > > > at java.lang.Thread.run(Thread.java:745) > > > Caused by: java.io.IOException: java.io.IOException: java.lang. > IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:126) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder. > > build(DoggedCubeBuilder.java:73) > > > at org.apache.kylin.cube.inmemcubing.AbstractInMemCubeBuilder$1. > > run(AbstractInMemCubeBuilder.java:80) > > > ... 5 more > > > Caused by: java.io.IOException: java.lang.IllegalArgumentException: > > Value not exists! > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.abort(DoggedCubeBuilder.java:194) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.checkException(DoggedCubeBuilder.java:167) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > BuildOnce.build(DoggedCubeBuilder.java:114) > > > ... 7 more > > > Caused by: java.lang.IllegalArgumentException: Value not exists! > > > at org.apache.kylin.common.util.Dictionary.getIdFromValueBytes( > > Dictionary.java:162) > > > at org.apache.kylin.dict.TrieDictionary.getIdFromValueImpl( > > TrieDictionary.java:167) > > > at org.apache.kylin.common.util.Dictionary.getIdFromValue( > > Dictionary.java:98) > > > at org.apache.kylin.dimension.DictionaryDimEnc$ > > DictionarySerializer.serialize(DictionaryDimEnc.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:121) > > > at org.apache.kylin.cube.gridtable.CubeCodeSystem. > > encodeColumnValue(CubeCodeSystem.java:110) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:93) > > > at org.apache.kylin.gridtable.GTRecord.setValues(GTRecord. > java:81) > > > at org.apache.kylin.cube.inmemcubing. > > InMemCubeBuilderInputConverter.convert(InMemCubeBuilderInputConverter > > .java:74) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:542) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder$ > > InputConverter$1.next(InMemCubeBuilder.java:523) > > > at org.apache.kylin.gridtable.GTAggregateScanner.iterator( > > GTAggregateScanner.java:139) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > createBaseCuboid(InMemCubeBuilder.java:339) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:166) > > > at org.apache.kylin.cube.inmemcubing.InMemCubeBuilder. > > build(InMemCubeBuilder.java:135) > > > at org.apache.kylin.cube.inmemcubing.DoggedCubeBuilder$ > > SplitThread.run(DoggedCubeBuilder.java:282) > > > > > > 2016-11-25 15:16:27,975 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from > > RUNNING to FAIL_CONTAINER_CLEANUP > > > 2016-11-25 15:16:27,976 INFO [ContainerLauncher #7] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > Processing the event EventType: CONTAINER_REMOTE_CLEANUP for container > > container_e125_1479580915733_0167_01_000005 taskAttempt > > attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:27,997 INFO [ContainerLauncher #7] > > org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl: > > KILLING attempt_1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:27,997 INFO [ContainerLauncher #7] > > org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy: > > Opening proxy : hadoopclusterslic72.ad.infosys.com:45454 > > > 2016-11-25 15:16:28,009 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from > > FAIL_CONTAINER_CLEANUP to FAIL_TASK_CLEANUP > > > 2016-11-25 15:16:28,011 INFO [CommitterEvent Processor #4] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: TASK_ABORT > > > 2016-11-25 15:16:28,013 WARN [CommitterEvent Processor #4] > > org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: Could not > > delete hdfs://SLICHDP/kylin/kylin_metadata/kylin-b2f43555-7105- > > 4912-b0bf-c40a0b405a05/ORC_ALERT_C/cuboid/_temporary/1/_ > temporary/attempt_ > > 1479580915733_0167_m_000000_3 > > > 2016-11-25 15:16:28,014 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_m_000000_3 TaskAttempt Transitioned from > > FAIL_TASK_CLEANUP to FAILED > > > 2016-11-25 15:16:28,026 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_m_000000 Task Transitioned from RUNNING to > FAILED > > > 2016-11-25 15:16:28,026 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Num completed > Tasks: > > 1 > > > 2016-11-25 15:16:28,027 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: Job failed as tasks > > failed. failedMaps:1 failedReduces:0 > > > 2016-11-25 15:16:28,027 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: 3 failures on node > hadoopclusterslic72.ad. > > infosys.com > > > 2016-11-25 15:16:28,027 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerRequestor: Blacklisted host hadoopclusterslic72.ad. > > infosys.com > > > 2016-11-25 15:16:28,032 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from RUNNING to FAIL_WAIT > > > 2016-11-25 15:16:28,033 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_r_000000 Task Transitioned from SCHEDULED to > > KILL_WAIT > > > 2016-11-25 15:16:28,033 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: > > attempt_1479580915733_0167_r_000000_0 TaskAttempt Transitioned from > > UNASSIGNED to KILLED > > > 2016-11-25 15:16:28,034 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: > > task_1479580915733_0167_r_000000 Task Transitioned from KILL_WAIT to > > KILLED > > > 2016-11-25 15:16:28,034 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from FAIL_WAIT to FAIL_ABORT > > > 2016-11-25 15:16:28,037 INFO [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Processing the event EventType: > > CONTAINER_DEALLOCATE > > > 2016-11-25 15:16:28,037 ERROR [Thread-53] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Could not deallocate container for task > > attemptId attempt_1479580915733_0167_r_000000_0 > > > 2016-11-25 15:16:28,043 INFO [CommitterEvent Processor #0] > > org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler: > > Processing the event EventType: JOB_ABORT > > > 2016-11-25 15:16:28,058 INFO [AsyncDispatcher event handler] > > org.apache.hadoop.mapreduce.v2.app.job.impl.JobImpl: > > job_1479580915733_0167Job Transitioned from FAIL_ABORT to FAILED > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > We are finishing cleanly so this is the last retry > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Notify RMCommunicator isAMLastRetry: true > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.rm.RMCommunicator: > > RMCommunicator notified that shouldUnregistered is: true > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Notify JHEH isAMLastRetry: true > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: JobHistoryEventHandler notified that > > forceJobCompletion is true > > > 2016-11-25 15:16:28,092 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Calling stop for all the services > > > 2016-11-25 15:16:28,093 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Stopping JobHistoryEventHandler. Size > > of the outstanding queue size is 2 > > > 2016-11-25 15:16:28,097 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: In stop, writing event TASK_FAILED > > > 2016-11-25 15:16:28,099 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: In stop, writing event JOB_FAILED > > > 2016-11-25 15:16:28,177 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Copying > hdfs://SLICHDP:8020/user/hdfs/ > > .staging/job_1479580915733_0167/job_1479580915733_0167_1.jhist to > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167- > > 1480067133013-hdfs-Kylin_Cube_Builder_ORC_ALERT_C- > > 1480067188027-0-0-FAILED-default-1480067140199.jhist_tmp > > > 2016-11-25 15:16:28,248 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Copied to done location: > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167- > > 1480067133013-hdfs-Kylin_Cube_Builder_ORC_ALERT_C- > > 1480067188027-0-0-FAILED-default-1480067140199.jhist_tmp > > > 2016-11-25 15:16:28,253 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Copying > hdfs://SLICHDP:8020/user/hdfs/ > > .staging/job_1479580915733_0167/job_1479580915733_0167_1_conf.xml to > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167_conf.xml_ > > tmp > > > 2016-11-25 15:16:28,320 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Copied to done location: > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167_conf.xml_ > > tmp > > > 2016-11-25 15:16:28,338 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Moved tmp to done: > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_ > 1479580915733_0167.summary_tmp > > to hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_ > 1479580915733_0167.summary > > > 2016-11-25 15:16:28,350 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Moved tmp to done: > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167_conf.xml_ > tmp > > to hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_ > 1479580915733_0167_conf.xml > > > 2016-11-25 15:16:28,353 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Moved tmp to done: > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167- > > 1480067133013-hdfs-Kylin_Cube_Builder_ORC_ALERT_C- > > 1480067188027-0-0-FAILED-default-1480067140199.jhist_tmp to > > hdfs://SLICHDP:8020/mr-history/tmp/hdfs/job_1479580915733_0167- > > 1480067133013-hdfs-Kylin_Cube_Builder_ORC_ALERT_C- > > 1480067188027-0-0-FAILED-default-1480067140199.jhist > > > 2016-11-25 15:16:28,353 INFO [Thread-74] org.apache.hadoop.mapreduce. > > jobhistory.JobHistoryEventHandler: Stopped JobHistoryEventHandler. > > super.stop() > > > 2016-11-25 15:16:28,357 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.rm.RMCommunicator: > > Setting job diagnostics to Task failed task_1479580915733_0167_m_000000 > > > Job failed as tasks failed. failedMaps:1 failedReduces:0 > > > > > > 2016-11-25 15:16:28,357 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.rm.RMCommunicator: > > History url is http://hadoopclusterslic73.ad. > infosys.com:19888/jobhistory/ > > job/job_1479580915733_0167 > > > 2016-11-25 <http://hadoopclusterslic73.ad.infosys.com:19888/ > > jobhistory/job/job_1479580915733_01672016-11-25> 15:16:28,373 INFO > > [Thread-74] org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: > Waiting > > for application to be successfully unregistered. > > > 2016-11-25 15:16:29,375 INFO [Thread-74] org.apache.hadoop.mapreduce. > > v2.app.rm.RMContainerAllocator: Final Stats: PendingReds:1 > > ScheduledMaps:0 ScheduledReds:0 AssignedMaps:1 AssignedReds:0 > > CompletedMaps:0 CompletedReds:0 ContAlloc:4 ContRel:0 HostLocal:1 > > RackLocal:0 > > > 2016-11-25 15:16:29,377 INFO [Thread-74] org.apache.hadoop.mapreduce. > v2.app.MRAppMaster: > > Deleting staging directory hdfs://SLICHDP /user/hdfs/.staging/job_ > > 1479580915733_0167 > > > 2016-11-25 15:16:29,380 INFO [Thread-74] org.apache.hadoop.ipc.Server: > > Stopping server on 57220 > > > 2016-11-25 15:16:29,387 INFO [TaskHeartbeatHandler PingChecker] > > org.apache.hadoop.mapreduce.v2.app.TaskHeartbeatHandler: > > TaskHeartbeatHandler thread interrupted > > > 2016-11-25 15:16:29,387 INFO [IPC Server listener on 57220] > > org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 57220 > > > > > > > > > On Fri, Nov 25, 2016 at 7:36 PM, ShaoFeng Shi <shaofeng...@apache.org> > > > wrote: > > > > > >> Didn't hear of that. Hive table's file format is transparent for > Kylin; > > >> Even if the table is a view, Kylin can build from it. > > >> > > >> What's the detail error you got when using ORC table? If you can > provide > > >> the detail information, that would be better. > > >> > > >> 2016-11-25 18:22 GMT+08:00 suresh m <suresh.m8...@gmail.com>: > > >> > > >> > Hi Facing an issue where i can able to build cube with text format > but > > >> > unable to building cube with ORC tables. > > >> > > > >> > Let me know kylin having any issues with ORC format.? > > >> > > > >> > Hive having limitation that Text format tables not having > possibility > > >> to > > >> > enabling ACID properties since text format not supporting ACID. But > > for > > >> me > > >> > ACID properties is important to handle my data, this i can do with > ORC > > >> but > > >> > kylin throwing errors with ORC format. > > >> > > > >> > > > >> > Regards, > > >> > Suresh > > >> > > > >> > > >> > > >> > > >> -- > > >> Best regards, > > >> > > >> Shaofeng Shi 史少锋 > > >> > > > > > > > > > > > > -- > Best regards, > > Shaofeng Shi 史少锋 >