Joe McDonnell created IMPALA-13144:
--------------------------------------

             Summary: TestIcebergTable.test_migrated_table_field_id_resolution 
fails with Disk I/O error
                 Key: IMPALA-13144
                 URL: https://issues.apache.org/jira/browse/IMPALA-13144
             Project: IMPALA
          Issue Type: Bug
          Components: Backend
    Affects Versions: Impala 4.5.0
            Reporter: Joe McDonnell


A couple test jobs hit a failure on 
TestIcebergTable.test_migrated_table_field_id_resolution:
{noformat}
query_test/test_iceberg.py:270: in test_migrated_table_field_id_resolution
    vector, unique_database)
common/impala_test_suite.py:725: in run_test_case
    result = exec_fn(query, user=test_section.get('USER', '').strip() or None)
common/impala_test_suite.py:660: in __exec_in_impala
    result = self.__execute_query(target_impalad_client, query, user=user)
common/impala_test_suite.py:1013: in __execute_query
    return impalad_client.execute(query, user=user)
common/impala_connection.py:216: in execute
    fetch_profile_after_close=fetch_profile_after_close)
beeswax/impala_beeswax.py:191: in execute
    handle = self.__execute_query(query_string.strip(), user=user)
beeswax/impala_beeswax.py:384: in __execute_query
    self.wait_for_finished(handle)
beeswax/impala_beeswax.py:405: in wait_for_finished
    raise ImpalaBeeswaxException("Query aborted:" + error_log, None)
E   ImpalaBeeswaxException: ImpalaBeeswaxException:
E    Query aborted:Disk I/O error on 
impala-ec2-centos79-m6i-4xlarge-xldisk-153e.vpc.cloudera.com:27000: Failed to 
open HDFS file 
hdfs://localhost:20500/test-warehouse/iceberg_migrated_alter_test/000000_0
E   Error(2): No such file or directory
E   Root cause: RemoteException: File does not exist: 
/test-warehouse/iceberg_migrated_alter_test/000000_0
E       at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:87)
E       at 
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:77)
E       at 
org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:159)
E       at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:2040)
E       at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:738)
E       at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:454)
E       at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
E       at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:533)
E       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070)
E       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:994)
E       at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:922)
E       at java.security.AccessController.doPrivileged(Native Method)
E       at javax.security.auth.Subject.doAs(Subject.java:422)
E       at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
E       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2899){noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to