[ https://issues.apache.org/jira/browse/IMPALA-9328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Quanlong Huang resolved IMPALA-9328. ------------------------------------ Resolution: Not A Problem > Data loading fail when set USE_CDP_HIVE=true > -------------------------------------------- > > Key: IMPALA-9328 > URL: https://issues.apache.org/jira/browse/IMPALA-9328 > Project: IMPALA > Issue Type: Test > Reporter: Quanlong Huang > Priority: Major > Attachments: hive-server2.log.gz, > load-tpcds-core-hive-generated-text-none-none.sql.log.gz, tez_app_status.png > > > Permanently fail to reload test data after rebasing to the latest master > branch with USE_CDP_HIVE=true. > {code:java} > Executing: create-load-data.sh > Generating HBase data (logging to > /home/quanlong/workspace/Impala/logs/data_loading/create-hbase.log)... > **** Timeout Timer Started (pid 1113, ppid 1027) for 9000 s! **** > Generating HBase data OK (Took: 0 min 13 sec) > Creating /test-warehouse HDFS directory (logging to > /home/quanlong/workspace/Impala/logs/data_loading/create-test-warehouse-dir.log)... > > Creating /test-warehouse HDFS directory OK (Took: 0 min 2 sec) > Derived params for create-load-data.sh: > EXPLORATION_STRATEGY=exhaustive > SKIP_METADATA_LOAD=0 > SKIP_SNAPSHOT_LOAD=0 > SNAPSHOT_FILE= > CM_HOST= > REMOTE_LOAD= > Starting Impala cluster (logging to > /home/quanlong/workspace/Impala/logs/data_loading/start-impala-cluster.log)... > > Starting Impala cluster OK (Took: 0 min 7 sec) > Setting up HDFS environment (logging to > /home/quanlong/workspace/Impala/logs/data_loading/setup-hdfs-env.log)... > Setting up HDFS environment OK (Took: 0 min 7 sec) > Running warm up Hive statements > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > ERROR StatusLogger No log4j2 configuration file found. Using default > configuration: logging only errors to the console. Set system property > 'log4j2.debug' to show Log4j2 internal initialization logging. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > Connecting to jdbc:hive2://localhost:11050/default; > Connected to: Apache Hive (version 3.1.2000.7.0.2.0-212) > Driver: Hive JDBC (version 3.1.2000.7.0.2.0-212) > Transaction isolation: TRANSACTION_REPEATABLE_READ > INFO : Compiling > command(queryId=quanlong_20200127014446_97930032-2762-46a1-862c-2afdfd9a3891): > create database if not exists functional > INFO : Semantic Analysis Completed (retrial = false) > INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null) > INFO : Completed compiling > command(queryId=quanlong_20200127014446_97930032-2762-46a1-862c-2afdfd9a3891); > Time taken: 0.768 seconds > INFO : Executing > command(queryId=quanlong_20200127014446_97930032-2762-46a1-862c-2afdfd9a3891): > create database if not exists functional > INFO : Starting task [Stage-0:DDL] in serial mode > INFO : Completed executing > command(queryId=quanlong_20200127014446_97930032-2762-46a1-862c-2afdfd9a3891); > Time taken: 0.117 seconds > INFO : OK > No rows affected (1.02 seconds) > Beeline version 3.1.2000.7.0.2.0-212 by Apache Hive > Closing: 0: jdbc:hive2://localhost:11050/default; > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > ERROR StatusLogger No log4j2 configuration file found. Using default > configuration: logging only errors to the console. Set system property > 'log4j2.debug' to show Log4j2 internal initialization logging. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > Connecting to jdbc:hive2://localhost:11050/default; > Connected to: Apache Hive (version 3.1.2000.7.0.2.0-212) > Driver: Hive JDBC (version 3.1.2000.7.0.2.0-212) > Transaction isolation: TRANSACTION_REPEATABLE_READ > INFO : Compiling > command(queryId=quanlong_20200127014450_59394581-7916-427b-8011-4fa41db4357d): > create table if not exists hive_warm_up_tbl (i int) > INFO : Semantic Analysis Completed (retrial = false) > INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null) > INFO : Completed compiling > command(queryId=quanlong_20200127014450_59394581-7916-427b-8011-4fa41db4357d); > Time taken: 0.049 seconds > INFO : Executing > command(queryId=quanlong_20200127014450_59394581-7916-427b-8011-4fa41db4357d): > create table if not exists hive_warm_up_tbl (i int) > INFO : Starting task [Stage-0:DDL] in serial mode > INFO : Completed executing > command(queryId=quanlong_20200127014450_59394581-7916-427b-8011-4fa41db4357d); > Time taken: 0.107 seconds > INFO : OK > No rows affected (0.218 seconds) > Beeline version 3.1.2000.7.0.2.0-212 by Apache Hive > Closing: 0: jdbc:hive2://localhost:11050/default; > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > ERROR StatusLogger No log4j2 configuration file found. Using default > configuration: logging only errors to the console. Set system property > 'log4j2.debug' to show Log4j2 internal initialization logging. > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/apache-hive-3.1.2000.7.0.2.0-212-bin/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in > [jar:file:/home/quanlong/workspace/Impala/toolchain/cdp_components-1617729/hadoop-3.1.1.7.0.2.0-212/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] > Connecting to jdbc:hive2://localhost:11050/default; > Connected to: Apache Hive (version 3.1.2000.7.0.2.0-212) > Driver: Hive JDBC (version 3.1.2000.7.0.2.0-212) > Transaction isolation: TRANSACTION_REPEATABLE_READ > INFO : Compiling > command(queryId=quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e): > insert overwrite table hive_warm_up_tbl values (1) > INFO : Semantic Analysis Completed (retrial = false) > INFO : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:col1, > type:int, comment:null)], properties:null) > INFO : Completed compiling > command(queryId=quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e); > Time taken: 0.867 seconds > INFO : Executing > command(queryId=quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e): > insert overwrite table hive_warm_up_tbl values (1) > INFO : Query ID = > quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e > INFO : Total jobs = 3 > INFO : Launching Job 1 out of 3 > INFO : Starting task [Stage-1:MAPRED] in serial mode > INFO : Subscribed to counters: [] for queryId: > quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e > INFO : Tez session hasn't been created yet. Opening session > INFO : Dag name: insert overwrite table hive_warm_up_tb...(1) (Stage-1) > INFO : Status: Running (Executing on YARN cluster with App id > application_1580060567575_0001) > INFO : Starting task [Stage-7:CONDITIONAL] in serial mode > INFO : Stage-4 is selected by condition resolver. > INFO : Stage-3 is filtered out by condition resolver. > INFO : Stage-5 is filtered out by condition resolver. > INFO : Starting task [Stage-4:MOVE] in serial mode > INFO : Moving data to directory > hdfs://localhost:20500/test-warehouse/hive_warm_up_tbl/.hive-staging_hive_2020-01-27_01-44-53_579_8871862015774311819-1/-ext-10000 > from > hdfs://localhost:20500/test-warehouse/hive_warm_up_tbl/.hive-staging_hive_2020-01-27_01-44-53_579_8871862015774311819-1/-ext-10002 > INFO : Starting task [Stage-2:DEPENDENCY_COLLECTION] in serial mode > INFO : Starting task [Stage-0:MOVE] in serial mode > INFO : Loading data to table default.hive_warm_up_tbl from > hdfs://localhost:20500/test-warehouse/hive_warm_up_tbl/.hive-staging_hive_2020-01-27_01-44-53_579_8871862015774311819-1/-ext-10000 > INFO : Completed executing > command(queryId=quanlong_20200127014453_1984fe85-2f81-419a-94dc-9357f3dd223e); > Time taken: 8.625 seconds > INFO : OK > ---------------------------------------------------------------------------------------------- > VERTICES MODE STATUS TOTAL COMPLETED RUNNING PENDING > FAILED KILLED > ---------------------------------------------------------------------------------------------- > Map 1 .......... container SUCCEEDED 1 1 0 0 > 0 0 > ---------------------------------------------------------------------------------------------- > VERTICES: 01/01 [==========================>>] 100% ELAPSED TIME: 3.00 s > > ---------------------------------------------------------------------------------------------- > No rows affected (9.539 seconds) > Beeline version 3.1.2000.7.0.2.0-212 by Apache Hive > Closing: 0: jdbc:hive2://localhost:11050/default; > Loading custom schemas (logging to > /home/quanlong/workspace/Impala/logs/data_loading/load-custom-schemas.log)... > Loading custom schemas OK (Took: 0 min 3 sec) > Started Loading functional-query data in background; pid 5027. > Started Loading TPC-H data in background; pid 5028. > Loading functional-query data (logging to > /home/quanlong/workspace/Impala/logs/data_loading/load-functional-query.log)... > > Started Loading TPC-DS data in background; pid 5031. > Loading TPC-H data (logging to > /home/quanlong/workspace/Impala/logs/data_loading/load-tpch.log)... > Loading TPC-DS data (logging to > /home/quanlong/workspace/Impala/logs/data_loading/load-tpcds.log)... > FAILED (Took: 5 min 35 sec) > 'load-data tpch core' failed. Tail of log: > 01:45:27 Impala Load Files: > 01:45:27 load-tpch-core-impala-generated-kudu-none-none.sql > 01:45:27 load-tpch-core-impala-generated-parquet-none-none.sql > 01:45:27 > 01:45:27 Impala Invalidate Files: > 01:45:27 invalidate-tpch-core-impala-generated.sql > 01:45:27 > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-seq-snap-block.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-rc-none-none.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-parquet-none-none.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-avro-none-none.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-avro-snap-block.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-seq-gzip-block.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-kudu-none-none.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-orc-def-block.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-text-none-none.sql > 01:45:27 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-text-gzip-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-seq-snap-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-parquet-none-none.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-text-none-none.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-text-gzip-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-avro-none-none.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-avro-snap-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-seq-gzip-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-orc-def-block.sql > 01:45:29 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-rc-none-none.sql > 01:45:32 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/create-tpch-core-impala-generated-kudu-none-none.sql > 01:45:32 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-text-none-none.sql > 01:45:39 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-text-none-none.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-text-gzip-block.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-avro-snap-block.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-avro-none-none.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-seq-gzip-block.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-seq-snap-block.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-orc-def-block.sql > 01:45:39 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-rc-none-none.sql > 01:46:39 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-text-gzip-block.sql > 01:47:18 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-avro-snap-block.sql > 01:48:05 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-seq-gzip-block.sql > 01:48:39 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-orc-def-block.sql > 01:49:08 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-rc-none-none.sql > 01:49:36 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-seq-snap-block.sql > 01:50:10 Finished execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-hive-generated-avro-none-none.sql > 01:50:10 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/invalidate-tpch-core-impala-generated.sql > 01:50:12 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/invalidate-tpch-core-impala-generated.sql > 01:50:12 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-impala-generated-kudu-none-none.sql > 01:50:12 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-impala-generated-parquet-none-none.sql > 01:50:41 Error executing impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-impala-generated-parquet-none-none.sql > See: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpch/load-tpch-core-impala-generated-parquet-none-none.sql.log > FAILED (Took: 5 min 59 sec) > 'load-data functional-query exhaustive' failed. Tail of log: > at > org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) > at com.sun.proxy.$Proxy10.addBlock(Unknown Source) > at > org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1085) > at > org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1866) > at > org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1668) > at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716) > put: File /test-warehouse/child_table/child_table.txt._COPYING_ could only be > written to 0 of the 1 minReplication nodes. There are 3 datanode(s) running > and 3 node(s) are excluded in this operation. > pty. Data will be loaded. > HDFS path: /test-warehouse/unsupported_types_rc_snap does not exists or is > empty. Data will be loaded. > Skipping 'functional_rc_snap.unsupported_partition_types' due to include > constraint match. > Skipping 'functional_rc_snap.old_rcfile_table' due to include constraint > match. > Skipping 'functional_rc_snap.bad_text_lzo' due to include constraint match. > Skipping 'functional_rc_snap.bad_text_gzip' due to include constraint match. > Skipping 'functional_rc_snap.bad_seq_snap' due to include constraint match. > Skipping 'functional_rc_snap.bad_avro_snap_strings' due to include constraint > match. > Skipping 'functional_rc_snap.bad_avro_snap_floats' due to include constraint > match. > Skipping 'functional_rc_snap.bad_avro_decimal_schema' due to include > constraint match. > Skipping 'functional_rc_snap.bad_avro_date_out_of_range' due to include > constraint match. > Skipping 'functional_rc_snap.hive2_bad_avro_date_pre_gregorian' due to > include constraint match. > Skipping 'functional_rc_snap.hive3_avro_date_pre_gregorian' due to include > constraint match. > Skipping 'functional_rc_snap.bad_parquet' due to include constraint match. > Skipping 'functional_rc_snap.bad_parquet_strings_negative_len' due to include > constraint match. > Skipping 'functional_rc_snap.bad_parquet_strings_out_of_bounds' due to > include constraint match. > Skipping 'functional_rc_snap.bad_magic_number' due to include constraint > match. > Skipping 'functional_rc_snap.alltypesagg_hive_13_1' due to include constraint > match. > Skipping 'functional_rc_snap.bad_metadata_len' due to include constraint > match. > Skipping 'functional_rc_snap.bad_dict_page_offset' due to include constraint > match. > Skipping 'functional_rc_snap.bad_compressed_size' due to include constraint > match. > Skipping 'functional_rc_snap.kite_required_fields' due to include constraint > match. > Skipping 'functional_rc_snap.bad_column_metadata' due to include constraint > match. > Skipping 'functional_rc_snap.bad_serde' due to include constraint match. > Skipping 'functional_rc_snap.rcfile_lazy_binary_serde' due to include > constraint match. > Skipping 'functional_rc_snap.decimal_tbl' due to include constraint match. > Skipping 'functional_rc_snap.decimal_rtf_tbl' due to include constraint match. > Skipping 'functional_rc_snap.decimal_rtf_tiny_tbl' due to include constraint > match. > Skipping 'functional_rc_snap.decimal_tiny' due to include constraint match. > HDFS path: /test-warehouse/parent_table_rc_snap does not exists or is empty. > Data will be loaded. > Empty insert for table parent_table. Skipping insert generation > HDFS path: /test-warehouse/parent_table_2_rc_snap does not exists or is > empty. Data will be loaded. > Empty insert for table parent_table_2. Skipping insert generation > Traceback (most recent call last): > File > "/home/quanlong/workspace/Impala/testdata/bin/generate-schema-statements.py", > line 859, in <module> > test_vectors, sections, include_constraints, exclude_constraints, > only_constraints) > File > "/home/quanlong/workspace/Impala/testdata/bin/generate-schema-statements.py", > line 618, in generate_statements > load = eval_section(section['LOAD']) > File > "/home/quanlong/workspace/Impala/testdata/bin/generate-schema-statements.py", > line 549, in eval_section > assert p.returncode == 0 > AssertionError > 01:51:05 Error generating schema statements for workload: functional-query > Background task Loading functional-query data (pid 5027) failed. > Background task Loading TPC-H data (pid 5028) failed. > FAILED (Took: 6 min 24 sec) > 'load-data tpcds core' failed. Tail of log: > HDFS path: /test-warehouse/tpcds.income_band_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.inventory_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.item_orc_def does not exists or is empty. > Data will be loaded. > HDFS path: /test-warehouse/tpcds.promotion_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.ship_mode_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.store_orc_def does not exists or is empty. > Data will be loaded. > HDFS path: /test-warehouse/tpcds.store_returns_orc_def does not exists or is > empty. Data will be loaded. > Skipping 'tpcds_orc_def.store_sales_unpartitioned' due to include constraint > match. > HDFS path: /test-warehouse/tpcds.store_sales_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.time_dim_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.warehouse_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.web_page_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.web_returns_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.web_sales_orc_def does not exists or is > empty. Data will be loaded. > HDFS path: /test-warehouse/tpcds.web_site_orc_def does not exists or is > empty. Data will be loaded. > 01:46:46 Avro schema dir > (/home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/avro_schemas) > does not exist. Skipping copy to HDFS. > 01:46:46 Impala Create Files: > 01:46:46 create-tpcds-core-impala-generated-orc-def-block.sql > 01:46:46 create-tpcds-core-impala-generated-parquet-none-none.sql > 01:46:46 create-tpcds-core-impala-generated-text-none-none.sql > 01:46:46 create-tpcds-core-impala-generated-seq-snap-block.sql > 01:46:46 > 01:46:46 Hive Load Text Files: > 01:46:46 load-tpcds-core-hive-generated-text-none-none.sql > 01:46:46 > 01:46:46 Hive Load Non-Text Files: > 01:46:46 load-tpcds-core-hive-generated-seq-snap-block.sql > 01:46:46 load-tpcds-core-hive-generated-orc-def-block.sql > 01:46:46 > 01:46:46 Impala Load Files: > 01:46:46 load-tpcds-core-impala-generated-parquet-none-none.sql > 01:46:46 > 01:46:46 Impala Invalidate Files: > 01:46:46 invalidate-tpcds-core-impala-generated.sql > 01:46:46 > 01:46:46 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-orc-def-block.sql > 01:46:46 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-text-none-none.sql > 01:46:46 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-seq-snap-block.sql > 01:46:46 Beginning execution of impala SQL on localhost: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-parquet-none-none.sql > 01:46:47 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-text-none-none.sql > 01:46:48 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-seq-snap-block.sql > 01:46:48 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-orc-def-block.sql > 01:46:48 Finished execution of impala SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/create-tpcds-core-impala-generated-parquet-none-none.sql > 01:46:48 Beginning execution of hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/load-tpcds-core-hive-generated-text-none-none.sql > 01:51:30 Error executing hive SQL: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/load-tpcds-core-hive-generated-text-none-none.sql > See: > /home/quanlong/workspace/Impala/logs/data_loading/sql/tpcds/load-tpcds-core-hive-generated-text-none-none.sql.log > Background task Loading TPC-DS data (pid 5031) failed. > ERROR in /home/quanlong/workspace/Impala/testdata/bin/create-load-data.sh at > line 85: fi > Generated: > /home/quanlong/workspace/Impala/logs/extra_junit_xml_logs/generate_junitxml.buildall.create-load-data.20200126_17_51_30.xml > {code} > Looks like due to failure in writing HDFS. > {code:java} > Caused by: org.apache.hadoop.ipc.RemoteException(java.io.IOException): File > /test-warehouse/tpcds.store_sales/.hive-staging_hive_2020-01-27_01-51-15_365_6638450221736290958-3/_task_tmp.-ext-10002/ss_sold_date_sk=2451846/_tmp.000000_3 > could only be written to 0 of the 1 minReplication nodes. There are 3 > datanode(s) running and 3 node(s) are excluded in this operation. > {code} > Attached load-tpcds-core-hive-generated-text-none-none.sql.log and > hive-server2.log. It's weird that the Tez job is shown succeed in Yarn > ([http://localhost:8088|http://localhost:8088/]). Attached the screenshot. -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org For additional commands, e-mail: issues-all-h...@impala.apache.org