[ 
https://issues.apache.org/jira/browse/HIVE-16331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Zhang resolved HIVE-16331.
-----------------------------
    Resolution: Fixed

looks duplicate to HIVE-12522

> create orc table fails in auto merge job when hive.exec.scratchdir set to 
> viewfs path 
> --------------------------------------------------------------------------------------
>
>                 Key: HIVE-16331
>                 URL: https://issues.apache.org/jira/browse/HIVE-16331
>             Project: Hive
>          Issue Type: Bug
>          Components: Hive
>            Reporter: Yi Zhang
>
> if hive.exec.sracthdir set to viewfs path, but fs.defaultFS not in viewfs, 
> when create ORC table in hive/tez, if auto merge job is kicked off, the auto 
> merge job fails with following error:
> ```
> 2017-03-29 23:10:57,892 INFO [main]: org.apache.hadoop.hive.ql.Driver: 
> Launching Job 3 out of 3
> 2017-03-29 23:10:57,894 INFO [main]: org.apache.hadoop.hive.ql.Driver: 
> Starting task [Stage-4:MAPRED] in serial mode
> 2017-03-29 23:10:57,894 INFO [main]: 
> org.apache.hadoop.hive.ql.exec.tez.TezSessionPoolManager: The current user: 
> yizhang, session user: yizhang
> 2017-03-29 23:10:57,894 INFO [main]: 
> org.apache.hadoop.hive.ql.exec.tez.TezSessionPoolManager: Current queue name 
> is hadoop-sync incoming queue name is hadoop-sync
> 2017-03-29 23:10:57,949 INFO [main]: hive.ql.Context: New scratch dir is 
> viewfs://ns-default/tmp/hive_scratchdir/yizhang/5da3d082-33b3-4194-97e2-005549d1b3c4/hive_2017-03-29_23-09-55_489_4030642791346631679-1
> 2017-03-29 23:10:57,949 DEBUG [main]: 
> org.apache.hadoop.hive.ql.exec.tez.DagUtils: TezDir path set 
> viewfs://ns-default/tmp/hive_scratchdir/yizhang/5da3d082-33b3-4194-97e2-005549d1b3c4/hive_2017-03-29_23-09-55_489_4030642791346631679-1/yizhang/_tez_scratch_dir
>  for user: yizhang
> 2017-03-29 23:10:57,950 DEBUG [main]: org.apache.hadoop.hdfs.DFSClient: 
> /tmp/hive_scratchdir/yizhang/5da3d082-33b3-4194-97e2-005549d1b3c4/hive_2017-03-29_23-09-55_489_4030642791346631679-1/yizhang/_tez_scratch_dir:
>  masked=rwxr-xr-x
> 2017-03-29 23:10:57,950 DEBUG [main]: org.apache.hadoop.ipc.Client: The ping 
> interval is 60000 ms.
> 2017-03-29 23:10:57,950 DEBUG [main]: org.apache.hadoop.ipc.Client: 
> Connecting to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020
> 2017-03-29 23:10:57,951 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang: starting, having connections 3
> 2017-03-29 23:10:57,951 DEBUG [IPC Parameter Sending Thread #0]: 
> org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang sending #373
> 2017-03-29 23:10:57,954 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang got value #373
> 2017-03-29 23:10:57,955 DEBUG [main]: 
> org.apache.hadoop.ipc.ProtobufRpcEngine: Call: mkdirs took 5ms
> 2017-03-29 23:10:57,955 INFO [main]: org.apache.hadoop.hive.ql.exec.Task: 
> Session is already open
> 2017-03-29 23:10:57,955 DEBUG [IPC Parameter Sending Thread #0]: 
> org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang sending #374
> 2017-03-29 23:10:57,956 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang got value #374
> 2017-03-29 23:10:57,956 DEBUG [main]: 
> org.apache.hadoop.ipc.ProtobufRpcEngine: Call: getFileInfo took 1ms
> 2017-03-29 23:10:57,956 DEBUG [IPC Parameter Sending Thread #0]: 
> org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang sending #375
> 2017-03-29 23:10:57,961 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang got value #375
> 2017-03-29 23:10:57,961 DEBUG [main]: 
> org.apache.hadoop.ipc.ProtobufRpcEngine: Call: getFileInfo took 5ms
> 2017-03-29 23:10:57,962 DEBUG [IPC Parameter Sending Thread #0]: 
> org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang sending #376
> 2017-03-29 23:10:57,962 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang got value #376
> 2017-03-29 23:10:57,962 DEBUG [main]: 
> org.apache.hadoop.ipc.ProtobufRpcEngine: Call: getFileInfo took 1ms
> 2017-03-29 23:10:57,963 INFO [main]: 
> org.apache.hadoop.hive.ql.exec.tez.DagUtils: Resource modification time: 
> 1490828996812
> 2017-03-29 23:10:57,963 DEBUG [main]: org.apache.hadoop.hive.ql.exec.Task: 
> Adding local resource: scheme: "viewfs" host: "ns-default" port: -1 file: 
> "/tmp/hive_scratchdir/yizhang/_tez_session_dir/c753ae3e-9051-436f-b2b9-d8c9d7670be3/hoodie-mr-0.2.1.jar"
> 2017-03-29 23:10:57,963 INFO [main]: 
> org.apache.hadoop.hive.ql.log.PerfLogger: <PERFLOG method=TezBuildDag 
> from=org.apache.hadoop.hive.ql.exec.tez.TezTask>
> 2017-03-29 23:10:57,963 INFO [main]: 
> org.apache.hadoop.hive.ql.log.PerfLogger: <PERFLOG 
> method=TezCreateVertex.File Merge 
> from=org.apache.hadoop.hive.ql.exec.tez.TezTask>
> 2017-03-29 23:10:57,966 INFO [main]: hive.ql.Context: New scratch dir is 
> viewfs://ns-default/tmp/hive_scratchdir/yizhang/5da3d082-33b3-4194-97e2-005549d1b3c4/hive_2017-03-29_23-09-55_489_4030642791346631679-1
> 2017-03-29 23:10:57,970 DEBUG [main]: org.apache.hadoop.hdfs.DFSClient: 
> /tmp/hive_scratchdir/yizhang/5da3d082-33b3-4194-97e2-005549d1b3c4/hive_2017-03-29_23-09-55_489_4030642791346631679-1/yizhang/_tez_scratch_dir/b3356fae-c055-4980-a2ef-7603d162fab6:
>  masked=rwxr-xr-x
> 2017-03-29 23:10:57,970 DEBUG [IPC Parameter Sending Thread #0]: 
> org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang sending #377
> 2017-03-29 23:10:57,972 DEBUG [IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang]: org.apache.hadoop.ipc.Client: IPC Client (85121323) connection to 
> hadooplithiumnamenode01-sjc1.prod.uber.internal/10.67.143.155:8020 from 
> yizhang got value #377
> 2017-03-29 23:10:57,972 DEBUG [main]: 
> org.apache.hadoop.ipc.ProtobufRpcEngine: Call: mkdirs took 2ms
> 2017-03-29 23:10:57,974 INFO [main]: 
> org.apache.hadoop.hive.ql.exec.tez.DagUtils: Vertex has custom input? false
> 2017-03-29 23:10:57,975 ERROR [main]: org.apache.hadoop.hive.ql.exec.Task: 
> Failed to execute tez graph.
> java.lang.IllegalArgumentException: Wrong FS: 
> hdfs://nameservice1/tmp/hive_stagingdir/yizhang_hive_2017-03-29_23-09-55_489_4030642791346631679-1/_tmp.-ext-10001,
>  expected: viewfs://ns-default/
>         at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:657)
>         at 
> org.apache.hadoop.fs.viewfs.ViewFileSystem.getUriPath(ViewFileSystem.java:117)
>         at 
> org.apache.hadoop.fs.viewfs.ViewFileSystem.getFileStatus(ViewFileSystem.java:346)
>         at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1412)
>         at 
> org.apache.hadoop.hive.ql.exec.tez.DagUtils.createVertex(DagUtils.java:576)
>         at 
> org.apache.hadoop.hive.ql.exec.tez.DagUtils.createVertex(DagUtils.java:1073)
>        at org.apache.hadoop.hive.ql.exec.tez.TezTask.build(TezTask.java:329)
>         at 
> org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:154)
>         at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:172)
>         at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
>         at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1903)
>         at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1630)
>         at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1392)
>         at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1180)
>         at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1168)
>         at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:220)
>         at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:172)
>         at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:383)
>         at 
> org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:775)
>         at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:693)
>         at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:628)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>         at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> 2017-03-29 23:10:57,987 ERROR [main]: org.apache.hadoop.hive.ql.Driver: 
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.tez.TezTask
> 2017-03-29 23:10:57,987 DEBUG [main]: org.apache.hadoop.hive.ql.Driver: 
> Shutting down query --explain
> create  table raw_trifle_tmp stored as orc  as
> SELECT max(trans_amount) as max_trans_amount, profile_uuid, state, 
> gateway_name
> FROM  yizhang_prod1.raw_trifle_parq
> GROUP BY profile_uuid, state, gateway_name
> ```



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to