[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
(https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
(https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console). In 
the following the failed test cases are given.
{code:java}

19:33:57 [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances


19:33:57 webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads

19:43:36 [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop

19:43:37 [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: rc/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none]

19:47:28  reason: Expected results differ across file formats

19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.

19:33:57 [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances

19:33:57 webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
 19:43:36 [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop

19:43:37 [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': False, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: avro/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': False, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: avro/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL data_errors/test_data_errors.py::TestHdfsScanNodeErrors:
{code:java}
// code placeholder
:({code}
)::test_hdfs_scan_node_errors[protocol: beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': True, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: rc/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': False, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: orc/def/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': True, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: kudu/none]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': True, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: avro/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': True, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: kudu/none]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': False, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: seq/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': True, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: avro/snap/block]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:

{'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
'disable_codegen': False, 'abort_on_error': 1, 
'exec_single_node_rows_threshold': 0}

| table_format: kudu/none]
 19:47:28 reason: Expected results differ across file formats
 19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option:


[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.

 

*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances

*19:33:57* webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads

*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop

*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

  was:
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.

 

*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances 

 

*19:33:57* webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads

*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 

 

*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: rc/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
(https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
(https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console). In 
the following the failed test cases are given.

19:33:57 [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances 

19:33:57 webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
19:43:36 [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 

19:43:37 [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: rc/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none]
19:47:28  reason: Expected results differ across file formats
19:47:28 XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: {'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.

 

*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances 

 

*19:33:57* webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads

*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop

  was:
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.
*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances *19:33:57* 
webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 
*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
rc/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.

 

*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances 

 

*19:33:57* webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads

*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 

 

*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: rc/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none]

*19:47:28*  reason: Expected results differ across file formats

*19:47:28* XFAIL 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
* After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.
 * *19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances *19:33:57* 
webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
 * *19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 
*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states
 * *19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block] *19:47:28*  reason: Expected results differ across file 
formats *19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block] *19:47:28*  reason: Expected results differ across file 
formats *19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: rc/snap/block] 
*19:47:28*  reason: Expected results differ across file formats *19:47:28* 
XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block] *19:47:28*  reason: Expected results differ across file formats 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none] *19:47:28*  
reason: Expected results differ across file formats *19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block] 
*19:47:28*  reason: Expected results differ across file formats *19:47:28* 
XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none] *19:47:28*  
reason: Expected results differ across file formats *19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block] *19:47:28*  reason: Expected results differ across file formats 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block] 
*19:47:28*  reason: Expected results differ across file formats *19:47:28* 
XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none] *19:47:28*  reason: Expected results differ across file formats 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the failed test cases are given.
*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances *19:33:57* 
webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 
*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
rc/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: 

[jira] [Updated] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang-Yu Rao updated IMPALA-8466:

Description: 
After sending my patch set to Jenkins for dry-run tests 
([https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console]), unrelated 
failures occurred. Specifically, some tests under ubuntu-16.04-dockerised-tests 
failed 
([https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/90/console]). In 
the following the messages related to the failed test cases are given.
*19:33:57* [gw7] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances *19:33:57* 
webserver/test_web_pages.py::TestWebPage::test_io_mgr_threads
*19:43:36* [gw11] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_instances_mt_dop 
*19:43:37* [gw8] xfail 
webserver/test_web_pages.py::TestWebPage::test_backend_states 
*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
rc/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
orc/def/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: kudu/none]*19:47:28*   
reason: Expected results differ across file formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
seq/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'exec_single_node_rows_threshold': 0} | table_format: 
avro/snap/block]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 beeswax | exec_option: \{'batch_size': 1, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
kudu/none]*19:47:28*   reason: Expected results differ across file 
formats*19:47:28* XFAIL 
data_errors/test_data_errors.py::TestHdfsScanNodeErrors::()::test_hdfs_scan_node_errors[protocol:
 

[jira] [Commented] (IMPALA-8381) Remove branch from ParquetPlainEncoder::Decode()

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8381?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827373#comment-16827373
 ] 

ASF subversion and git services commented on IMPALA-8381:
-

Commit 6a703741d8fdc359833a0d593ca8b121cd5d890d in impala's branch 
refs/heads/master from Daniel Becker
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=6a70374 ]

IMPALA-8381: Optimize ParquetPlainEncoder::DecodeBatch() for simple types

Refactored the ParquetPlainEncoder::Decode() and
ParquetPlainEncoder::DecodeBatch() methods to increase performance in
batch decoding.

The `Decode` and `DecodeBatch` methods retain their behaviour and
outward interface, but the internal structure changes.

We change how we split up the `Decode` template specialisations. The
generic unspecialised template is used for numerical parquet types
(INT32, INT64, INT96, FLOAT and DOUBLE) and various specialisations are
used for BYTE_ARRAY and FIXED_LEN_BYTE_ARRAY.

We add a new method template, DecodeNoCheck, which does the actual
decoding without bounds checking. It is called by the generic Decode
method template internally. For all parquet types except for BYTE_ARRAY,
DecodeBatch performs the bounds check once for the whole batch at the
same time and calls DecodeNoCheck, so we save the cost of bounds
checking for every decoded value. For BYTE_ARRAY, this cannot be done
and we have to perform the checks for every value.

In the non-BYTE_ARRAY version of DecodeBatch, we explicitly unroll the
loop in batches of 8 to increase performance.

The overall performance increase is up to 2x for small strides (8 bytes,
INT32) but decreases as the stride increases, and disappears from around
40 bytes. With bigger strides, there is no performance difference from
the previous implementation.

Testing:
  Added tests to parquet-plain-test.cc to test the `Decode` and the
  `DecodeBatch` methods both in single-value decoding and batch
  decoding.

Change-Id: I57b7d2573bb6dfd038e581acb3bd8ea1565aa20d
Reviewed-on: http://gerrit.cloudera.org:8080/12985
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Remove branch from ParquetPlainEncoder::Decode()
> 
>
> Key: IMPALA-8381
> URL: https://issues.apache.org/jira/browse/IMPALA-8381
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Csaba Ringhofer
>Assignee: Daniel Becker
>Priority: Minor
>  Labels: newbie, parquet, performance, ramp-up
>
> Removing the "if" at
> https://github.com/apache/impala/blob/5670f96b828d57f9e36510bb9af02bcc31de775c/be/src/exec/parquet/parquet-common.h#L203
> can lead to 1.5x speed up in plain decoding (type=int32, stride=16). For 
> primitive types, the same check can be done for a whole batch, so the speedup 
> can be gained for large batches without loosing safety. The only Parquet type 
> where this check is needed per element is BYTE_ARRAY (typically used for 
> STRING columns), which already has a template specialization for  
> ParquetPlainEncoder::Decode().



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8444) Analysis perf regression after IMPALA-7616

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827374#comment-16827374
 ] 

ASF subversion and git services commented on IMPALA-8444:
-

Commit 349142635250f92412003a1e62829fc4b441dc75 in impala's branch 
refs/heads/master from Fredy Wijaya
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=3491426 ]

IMPALA-8444: Fix performance regression when building privilege name

This patch fixes the performance regression when building privilege name
by rewriting PrincipalPrivilege.buildPrivilegeName() with a simple
string concatentation instead of using a list that gets converted into a
string.

Below is the result of running a benchmark using JMH comparing the old
and new implementations:

Result "org.apache.impala.BuildPrivilegeNameBenchmark.fast":
  0.344 ±(99.9%) 0.004 us/op [Average]
  (min, avg, max) = (0.336, 0.344, 0.355), stdev = 0.005
  CI (99.9%): [0.339, 0.348] (assumes normal distribution)

Result "org.apache.impala.BuildPrivilegeNameBenchmark.slow":
  0.831 ±(99.9%) 0.011 us/op [Average]
  (min, avg, max) = (0.807, 0.831, 0.856), stdev = 0.015
  CI (99.9%): [0.820, 0.842] (assumes normal distribution)

Benchmark Mode  Cnt  Score   Error Units
BuildPrivilegeNameBenchmark.fast  avgt   25  0.344 ± 0.004 us/op
BuildPrivilegeNameBenchmark.slow  avgt   25  0.831 ± 0.011 us/op

This patch also updates SentryAuthorizationPolicy.listPrivileges() to
reuse the privilege names that have already been built instead of
building them again. While fixing this, I found a bug where Principal
stores the PrincipalPrivilege in a case insensitive way. This is true
for all privilege scopes, except URI. This patch fixes the issue by
making privilege name to be case sensitive instead.

This patch removes incorrect synchronization in
SentryAuthorizationPolicy.listPrivileges() that can cause the operation
to run in serial in a highly concurrent workload.

Testing:
- Ran all FE tests
- Ran all E2E authorization tests
- Added E2E test for privilege name case sensitivity bug

Change-Id: I942d9b55f07c8972f69e532567d9b7d80fceb6e5
Reviewed-on: http://gerrit.cloudera.org:8080/13095
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Analysis perf regression after IMPALA-7616
> --
>
> Key: IMPALA-8444
> URL: https://issues.apache.org/jira/browse/IMPALA-8444
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.2.0
>Reporter: bharath v
>Assignee: Fredy Wijaya
>Priority: Critical
> Fix For: Impala 3.3.0
>
>
> The patch for IMPALA-7616 caused a performance regression in analysis time 
> when run in an environment with ~1k roles and 10.5k. privileges. The 
> regression is evident when run as a role that has a large number of 
> privileges.
> Following is the stack to look for when jstacking the coordnator.
> {noformat}
> "Thread-21" #49 prio=5 os_prio=0 tid=0x0cd6e000 nid=0x6a3d runnable 
> [0x7fa28e4a]
>java.lang.Thread.State: RUNNABLE
> at java.lang.String.toLowerCase(String.java:2670)
> at 
> org.apache.impala.catalog.PrincipalPrivilege.buildPrivilegeName(PrincipalPrivilege.java:82)
> at 
> org.apache.impala.catalog.PrincipalPrivilege.getName(PrincipalPrivilege.java:143)
> at 
> org.apache.impala.catalog.AuthorizationPolicy.listPrivileges(AuthorizationPolicy.java:423)
> - locked <0x7fa376987100> (a 
> org.apache.impala.catalog.AuthorizationPolicy)
> at 
> org.apache.impala.catalog.AuthorizationPolicy.listPrivileges(AuthorizationPolicy.java:443)
> - locked <0x7fa376987100> (a 
> org.apache.impala.catalog.AuthorizationPolicy)
> at 
> org.apache.sentry.provider.cache.SimpleCacheProviderBackend.getPrivileges(SimpleCacheProviderBackend.java:75)
> at 
> org.apache.sentry.policy.db.SimpleDBPolicyEngine.getPrivileges(SimpleDBPolicyEngine.java:98)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.getPrivileges(ResourceAuthorizationProvider.java:147)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.doHasAccess(ResourceAuthorizationProvider.java:120)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.hasAccess(ResourceAuthorizationProvider.java:107)
> at 
> org.apache.impala.authorization.AuthorizationChecker.hasAccess(AuthorizationChecker.java:215)
> at 
> org.apache.impala.authorization.AuthorizationChecker.checkAccess(AuthorizationChecker.java:128)
> at 
> org.apache.impala.analysis.AnalysisContext.authorizePrivilegeRequest(AnalysisContext.java:592)
> at 
> org.apache.impala.analysis.AnalysisContext.authorize(AnalysisContext.java:564)
> at 
> 

[jira] [Commented] (IMPALA-8369) Impala should be able to interoperate with Hive 3.1.0

2019-04-26 Thread Vihang Karajgaonkar (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827368#comment-16827368
 ] 

Vihang Karajgaonkar commented on IMPALA-8369:
-

No, {{HiveServer2}} still remains HiveServer2 in Hive-3

> Impala should be able to interoperate with Hive 3.1.0
> -
>
> Key: IMPALA-8369
> URL: https://issues.apache.org/jira/browse/IMPALA-8369
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: impala-acid
>
> Currently, Impala only works with Hive 2.1.1. Since Hive 3.1.0 has been 
> released for a while it would be good to add support for Hive 3.1.0 (HMS 
> 3.1.0). This patch will focus on ability to connect to HMS 3.1.0 and run 
> existing tests. It will not focus on adding support for newer features like 
> ACID in Hive 3.1.0 which can be taken up as separate JIRA.
> It would be good to make changes to Impala source code such that it can work 
> with both Hive 2.1.0 and Hive 3.1.0 without the need to create a separate 
> branch. However, this should be a aspirational goal. If we hit a blocker we 
> should investigate alternative approaches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8466) Unrelated failures occurred after sending a patch set to Jenkins for dry-run tests

2019-04-26 Thread Fang-Yu Rao (JIRA)
Fang-Yu Rao created IMPALA-8466:
---

 Summary: Unrelated failures occurred after sending a patch set to 
Jenkins for dry-run tests
 Key: IMPALA-8466
 URL: https://issues.apache.org/jira/browse/IMPALA-8466
 Project: IMPALA
  Issue Type: Bug
  Components: Catalog, Infrastructure
Reporter: Fang-Yu Rao
Assignee: Tim Armstrong


After sending my patch set to Jenkins for dry-run tests, unrelated failures 
occurred. The messages could be found at 
[https://jenkins.impala.io/job/gerrit-verify-dryrun/4085/console].



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8446) Create a unit test for Admission Controller

2019-04-26 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman resolved IMPALA-8446.

   Resolution: Fixed
Fix Version/s: Impala 2.13.0

> Create a unit test for Admission Controller
> ---
>
> Key: IMPALA-8446
> URL: https://issues.apache.org/jira/browse/IMPALA-8446
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
> Fix For: Impala 2.13.0
>
>
> This will allow construction of white box tests that exercise Admission 
> Controller code.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8444) Analysis perf regression after IMPALA-7616

2019-04-26 Thread Fredy Wijaya (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fredy Wijaya resolved IMPALA-8444.
--
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Analysis perf regression after IMPALA-7616
> --
>
> Key: IMPALA-8444
> URL: https://issues.apache.org/jira/browse/IMPALA-8444
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.2.0
>Reporter: bharath v
>Assignee: Fredy Wijaya
>Priority: Critical
> Fix For: Impala 3.3.0
>
>
> The patch for IMPALA-7616 caused a performance regression in analysis time 
> when run in an environment with ~1k roles and 10.5k. privileges. The 
> regression is evident when run as a role that has a large number of 
> privileges.
> Following is the stack to look for when jstacking the coordnator.
> {noformat}
> "Thread-21" #49 prio=5 os_prio=0 tid=0x0cd6e000 nid=0x6a3d runnable 
> [0x7fa28e4a]
>java.lang.Thread.State: RUNNABLE
> at java.lang.String.toLowerCase(String.java:2670)
> at 
> org.apache.impala.catalog.PrincipalPrivilege.buildPrivilegeName(PrincipalPrivilege.java:82)
> at 
> org.apache.impala.catalog.PrincipalPrivilege.getName(PrincipalPrivilege.java:143)
> at 
> org.apache.impala.catalog.AuthorizationPolicy.listPrivileges(AuthorizationPolicy.java:423)
> - locked <0x7fa376987100> (a 
> org.apache.impala.catalog.AuthorizationPolicy)
> at 
> org.apache.impala.catalog.AuthorizationPolicy.listPrivileges(AuthorizationPolicy.java:443)
> - locked <0x7fa376987100> (a 
> org.apache.impala.catalog.AuthorizationPolicy)
> at 
> org.apache.sentry.provider.cache.SimpleCacheProviderBackend.getPrivileges(SimpleCacheProviderBackend.java:75)
> at 
> org.apache.sentry.policy.db.SimpleDBPolicyEngine.getPrivileges(SimpleDBPolicyEngine.java:98)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.getPrivileges(ResourceAuthorizationProvider.java:147)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.doHasAccess(ResourceAuthorizationProvider.java:120)
> at 
> org.apache.sentry.provider.common.ResourceAuthorizationProvider.hasAccess(ResourceAuthorizationProvider.java:107)
> at 
> org.apache.impala.authorization.AuthorizationChecker.hasAccess(AuthorizationChecker.java:215)
> at 
> org.apache.impala.authorization.AuthorizationChecker.checkAccess(AuthorizationChecker.java:128)
> at 
> org.apache.impala.analysis.AnalysisContext.authorizePrivilegeRequest(AnalysisContext.java:592)
> at 
> org.apache.impala.analysis.AnalysisContext.authorize(AnalysisContext.java:564)
> at 
> org.apache.impala.analysis.AnalysisContext.analyzeAndAuthorize(AnalysisContext.java:415)
> at 
> org.apache.impala.service.Frontend.doCreateExecRequest(Frontend.java:1240)
> at 
> org.apache.impala.service.Frontend.getTExecRequest(Frontend.java:1210)
> at 
> org.apache.impala.service.Frontend.createExecRequest(Frontend.java:1182)
> at 
> org.apache.impala.service.JniFrontend.createExecRequest(JniFrontend.java:158)
> {noformat}
> Issue worsens when running concurrent workloads, because the underlying 
> Sentry {{listPrivileges()}} is synchronized and that serializes all the query 
> analysis requests.
> {noformat}
>   public synchronized Set listPrivileges(Set groups,  <
>   ActiveRoleSet roleSet) {
> Set privileges = Sets.newHashSet();
> if (roleSet != ActiveRoleSet.ALL) {
>   throw new UnsupportedOperationException("Impala does not support role 
> subsets.");
> }
> {noformat}
> Notes:
> - If the authorization metadata footprint is small, this issue can be ignored.
> - One workaround is to run the query using a role that has very small number 
> of privileges (revoke privileges that are not necessary to run a given query).
> - Another workaround is to disable Sentry authorization.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-6590) Disable expr rewrites and codegen for VALUES() statements

2019-04-26 Thread Abhishek Rawat (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6590?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-6590 started by Abhishek Rawat.
--
> Disable expr rewrites and codegen for VALUES() statements
> -
>
> Key: IMPALA-6590
> URL: https://issues.apache.org/jira/browse/IMPALA-6590
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 2.8.0, Impala 2.9.0, Impala 2.10.0, Impala 2.11.0
>Reporter: Alexander Behm
>Assignee: Abhishek Rawat
>Priority: Major
>  Labels: perf, planner, ramp-up, regression
>
> The analysis of statements with big VALUES clauses like INSERT INTO  
> VALUES is slow due to expression rewrites like constant folding. The 
> performance of such statements has regressed since the introduction of expr 
> rewrites and constant folding in IMPALA-1788.
> We should skip expr rewrites for VALUES altogether since it mostly provides 
> no benefit but can have a large overhead due to evaluation of expressions in 
> the backend (constant folding). These expressions are ultimately evaluated 
> and materialized in the backend anyway, so there's no point in folding them 
> during analysis.
> Similarly, there is no point in doing codegen for these exprs in the backend 
> union node.
> *Workaround*
> {code}
> SET ENABLE_EXPR_REWRITES=FALSE;
> SET DISABLE_CODEGEN=TRUE;
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-6088) Rack aware broadcast operator

2019-04-26 Thread Ruslan Dautkhanov (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-6088?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827317#comment-16827317
 ] 

Ruslan Dautkhanov commented on IMPALA-6088:
---

Would be also interesting to look if something like Torrent protocol helps here 
too.
This Torrent protocol can take into account rack location, if available. 

Spark uses this to avoid -query coordinator- Spark Driver to be a network 
bottleneck 

https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/broadcast/TorrentBroadcast.scala

{quote}

* The driver divides the serialized object into small chunks and
 * stores those chunks in the BlockManager of the driver.

 * On each executor, the executor first attempts to fetch the object from its 
BlockManager. If
 * it does not exist, it then uses remote fetches to fetch the small chunks 
from the driver and/or
 * other executors if available. Once it gets the chunks, it puts the chunks in 
its own
 * BlockManager, ready for other executors to fetch from.

 * This prevents the driver from being the bottleneck in sending out multiple 
copies of the
 * broadcast data (one per executor).

{quote}

> Rack aware broadcast operator
> -
>
> Key: IMPALA-6088
> URL: https://issues.apache.org/jira/browse/IMPALA-6088
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Distributed Exec
>Reporter: Mostafa Mokhtar
>Priority: Major
>
> When conducting large scale experiments on a 6 rack cluster with aggregator 
> core network topology overall cluster bandwidth utilization was limited. 
> With aggregator core networks nodes and racks are not equidistant, which 
> means a broadcast operation can be inefficient as the broadcasting node needs 
> to send the same data N times to each node on a remote rack. 
> Ideally Rowbatches should be sent once per remote rack then a node on each 
> remote rack would broadcast within its rack. 
> Table below represent rack to rack latency for the 90% of operations, ration 
> between best and worst case is 7.3x 
> | ||  va||vc||vd1||   vd3||   ve|
> |va|  4,238|  4,290|  9,692|  8,897|  8,208|
> |vc|  9,290|  4,396|  30,952| 13,529| 14,578|
> |vd1| 9,131|  29,066| 4,346|  17,265| 16,849|
> |vd3| 7,409|  15,517| 17,265| 4,370|  4,687|
> |ve|  4,914|  16,894| 16,430| 4,713|  4,472|



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-8465) hs2.test_json_endpoints.TestJsonEndpoints fails on deployed clusters

2019-04-26 Thread Tim Armstrong (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-8465:
-

Assignee: Tim Armstrong

> hs2.test_json_endpoints.TestJsonEndpoints fails on deployed clusters
> 
>
> Key: IMPALA-8465
> URL: https://issues.apache.org/jira/browse/IMPALA-8465
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: David Knupp
>Assignee: Tim Armstrong
>Priority: Major
>
> Changes to tests.common.ImpalaCluster in [commit 
> 2ca7f8e7|https://github.com/apache/impala/commit/2ca7f8e7c0781a1914275b3506cf8a7748c44c85#diff-6fea89ad0e6c440b0373bb136d7510b5]
>  introduced a regression in this test.
> {noformat}
> hs2/test_json_endpoints.py:51: in test_waiting_in_flight_queries
> queries_json = self._get_json_queries(http_addr)
> hs2/test_json_endpoints.py:33: in _get_json_queries
> return cluster.impalads[0].service.get_debug_webpage_json("/queries")
> E   IndexError: list index out of range
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8465) hs2.test_json_endpoints.TestJsonEndpoints fails on deployed clusters

2019-04-26 Thread David Knupp (JIRA)
David Knupp created IMPALA-8465:
---

 Summary: hs2.test_json_endpoints.TestJsonEndpoints fails on 
deployed clusters
 Key: IMPALA-8465
 URL: https://issues.apache.org/jira/browse/IMPALA-8465
 Project: IMPALA
  Issue Type: Bug
  Components: Infrastructure
Affects Versions: Impala 3.3.0
Reporter: David Knupp


Changes to tests.common.ImpalaCluster in [commit 
2ca7f8e7|https://github.com/apache/impala/commit/2ca7f8e7c0781a1914275b3506cf8a7748c44c85#diff-6fea89ad0e6c440b0373bb136d7510b5]
 introduced a regression in this test.
{noformat}
hs2/test_json_endpoints.py:51: in test_waiting_in_flight_queries
queries_json = self._get_json_queries(http_addr)
hs2/test_json_endpoints.py:33: in _get_json_queries
return cluster.impalads[0].service.get_debug_webpage_json("/queries")
E   IndexError: list index out of range
{noformat}




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8464) JdbcTest.testMetaDataGetColumnComments() fails with local catalog

2019-04-26 Thread Tim Armstrong (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827265#comment-16827265
 ] 

Tim Armstrong commented on IMPALA-8464:
---

[~tlipcon] [~bharathv] any thoughts on what might be happening here?

> JdbcTest.testMetaDataGetColumnComments() fails with local catalog
> -
>
> Key: IMPALA-8464
> URL: https://issues.apache.org/jira/browse/IMPALA-8464
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 3.1.0, Impala 3.2.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
>
> {noformat}
> Error Message
> Incorrect table comment expected:<[]> but was:<[table comment]>
> Stacktrace
> org.junit.ComparisonFailure: Incorrect table comment expected:<[]> but 
> was:<[table comment]>
>   at org.junit.Assert.assertEquals(Assert.java:115)
>   at 
> org.apache.impala.service.JdbcTest.testMetaDataGetColumnComments(JdbcTest.java:484)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:236)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:386)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:323)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:143)
> Standard Output
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
> org.apache.hive.jdbc.HiveDriver
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
> jdbc:hive2://localhost:21050/default;auth=noSasl
> 19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
> 19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
> org.apache.hive.jdbc.HiveDriver
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
> jdbc:hive2://localhost:21050/default;auth=noSasl
> 19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
> 19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
> org.apache.hive.jdbc.HiveDriver
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
> jdbc:hive2://localhost:21050/default;auth=noSasl
> 19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
> 19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
> org.apache.hive.jdbc.HiveDriver
> 19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
> jdbc:hive2://localhost:21050/default;auth=noSasl
> 19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: 

[jira] [Commented] (IMPALA-8369) Impala should be able to interoperate with Hive 3.1.0

2019-04-26 Thread Alex Rodoni (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8369?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827259#comment-16827259
 ] 

Alex Rodoni commented on IMPALA-8369:
-

[~vihangk1] [~tlipcon] When this is committed, should all occurrences of 
"HiveServer2" be changed to "HiveServer3" or just "HiveServer" in docs?

> Impala should be able to interoperate with Hive 3.1.0
> -
>
> Key: IMPALA-8369
> URL: https://issues.apache.org/jira/browse/IMPALA-8369
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: impala-acid
>
> Currently, Impala only works with Hive 2.1.1. Since Hive 3.1.0 has been 
> released for a while it would be good to add support for Hive 3.1.0 (HMS 
> 3.1.0). This patch will focus on ability to connect to HMS 3.1.0 and run 
> existing tests. It will not focus on adding support for newer features like 
> ACID in Hive 3.1.0 which can be taken up as separate JIRA.
> It would be good to make changes to Impala source code such that it can work 
> with both Hive 2.1.0 and Hive 3.1.0 without the need to create a separate 
> branch. However, this should be a aspirational goal. If we hit a blocker we 
> should investigate alternative approaches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-8457) Fix python when Impala isn't fully built

2019-04-26 Thread Thomas Tauber-Marshall (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8457?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-8457 started by Thomas Tauber-Marshall.
--
> Fix python when Impala isn't fully built
> 
>
> Key: IMPALA-8457
> URL: https://issues.apache.org/jira/browse/IMPALA-8457
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Thomas Tauber-Marshall
>Priority: Critical
>
> A common pattern is to check out Impala, eg. in a Jenkins job, and then run 
> Impala's functional python tests. If the tests are being run against a real 
> cluster that has already been setup (as opposed to being run against the 
> mini-cluster), it would be nice to be able to run any of our python tests 
> without having to do a full build of Impala.
> This usually works, but occasionally gets broken eg. due to people adding 
> imports of thrift structs to python files, since the thrift definitions 
> aren't available without a full build. We dealt with this once recently in 
> IMPALA-8199, and it has now cropped up again as a result of IMPALA-8158
> Going forward, I think a good solution would be to add a flag to buildall.sh, 
> say '-testonly', that ensures that everything needed to run the python tests 
> is done, eg. in this case something like 'make thrift-deps' should get run.
> We could incorporate running this into our pre-commit job to ensure that it 
> doesn't get broken again in the future.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8464) JdbcTest.testMetaDataGetColumnComments() fails with local catalog

2019-04-26 Thread Tim Armstrong (JIRA)
Tim Armstrong created IMPALA-8464:
-

 Summary: JdbcTest.testMetaDataGetColumnComments() fails with local 
catalog
 Key: IMPALA-8464
 URL: https://issues.apache.org/jira/browse/IMPALA-8464
 Project: IMPALA
  Issue Type: Bug
  Components: Catalog
Affects Versions: Impala 3.2.0, Impala 3.1.0
Reporter: Tim Armstrong
Assignee: Tim Armstrong


{noformat}
Error Message

Incorrect table comment expected:<[]> but was:<[table comment]>

Stacktrace

org.junit.ComparisonFailure: Incorrect table comment expected:<[]> but 
was:<[table comment]>
at org.junit.Assert.assertEquals(Assert.java:115)
at 
org.apache.impala.service.JdbcTest.testMetaDataGetColumnComments(JdbcTest.java:484)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:236)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:386)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:323)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:143)

Standard Output

19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
org.apache.hive.jdbc.HiveDriver
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
jdbc:hive2://localhost:21050/default;auth=noSasl
19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
org.apache.hive.jdbc.HiveDriver
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
jdbc:hive2://localhost:21050/default;auth=noSasl
19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
org.apache.hive.jdbc.HiveDriver
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
jdbc:hive2://localhost:21050/default;auth=noSasl
19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Using JDBC Driver Name: 
org.apache.hive.jdbc.HiveDriver
19/04/26 18:35:36 INFO testutil.ImpalaJdbcClient: Connecting to: 
jdbc:hive2://localhost:21050/default;auth=noSasl
19/04/26 18:35:36 INFO jdbc.Utils: Supplied authorities: localhost:21050
19/04/26 18:35:36 INFO jdbc.Utils: Resolved authority: localhost:21050
19/04/26 18:36:01 INFO service.FeSupport: Loading libfesupport.so
19/04/26 18:36:01 INFO service.FeSupport: Loaded libfesupport.so
19/04/26 18:36:01 INFO util.JvmPauseMonitor: Starting JVM pause monitor
{noformat}

Not sure what is going wrong here.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (IMPALA-8463) skip.header.line.count ignored on local catalog

2019-04-26 Thread Tim Armstrong (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-8463 started by Tim Armstrong.
-
> skip.header.line.count ignored on local catalog
> ---
>
> Key: IMPALA-8463
> URL: https://issues.apache.org/jira/browse/IMPALA-8463
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 3.1.0, Impala 3.2.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Critical
>
> test_text_scanner_with_header() fails when run against local catalog.
> {noformat}
> [tarmstrong-box2.ca.cloudera.com:21000] functional> select c1, c2 from 
> table_with_header
>   > ;
> Query: select c1, c2 from table_with_header
> Query submitted at: 2019-04-25 22:47:52 (Coordinator: 
> http://2c766e5e2187:25000)
> Query progress can be monitored at: 
> http://2c766e5e2187:25000/query_plan?query_id=68403d93ac914925:394ee6a1
> +--+--+
> | c1   | c2   |
> +--+--+
> | 5| 6|
> | 1| 2|
> | 3| 4|
> | NULL | NULL |
> +--+--+
> WARNINGS: Error converting column: 0 to INT
> Error converting column: 1 to DOUBLE
> Error parsing row: file: 
> hdfs://172.19.0.1:20500/test-warehouse/table_with_header/table_with_header.csv,
>  before offset: 22
> {noformat}
> This is suspicious in LocalFsTable.java
> {code}
>   @Override
>   public int parseSkipHeaderLineCount(StringBuilder error) {
> // TODO Auto-generated method stub
> return 0;
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8463) skip.header.line.count ignored on local catalog

2019-04-26 Thread Tim Armstrong (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-8463:
--
Target Version: Impala 3.3.0

> skip.header.line.count ignored on local catalog
> ---
>
> Key: IMPALA-8463
> URL: https://issues.apache.org/jira/browse/IMPALA-8463
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 3.1.0, Impala 3.2.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Critical
>
> test_text_scanner_with_header() fails when run against local catalog.
> {noformat}
> [tarmstrong-box2.ca.cloudera.com:21000] functional> select c1, c2 from 
> table_with_header
>   > ;
> Query: select c1, c2 from table_with_header
> Query submitted at: 2019-04-25 22:47:52 (Coordinator: 
> http://2c766e5e2187:25000)
> Query progress can be monitored at: 
> http://2c766e5e2187:25000/query_plan?query_id=68403d93ac914925:394ee6a1
> +--+--+
> | c1   | c2   |
> +--+--+
> | 5| 6|
> | 1| 2|
> | 3| 4|
> | NULL | NULL |
> +--+--+
> WARNINGS: Error converting column: 0 to INT
> Error converting column: 1 to DOUBLE
> Error parsing row: file: 
> hdfs://172.19.0.1:20500/test-warehouse/table_with_header/table_with_header.csv,
>  before offset: 22
> {noformat}
> This is suspicious in LocalFsTable.java
> {code}
>   @Override
>   public int parseSkipHeaderLineCount(StringBuilder error) {
> // TODO Auto-generated method stub
> return 0;
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8406) Failed REFRESH can partially modify table without bumping version number

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8406?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827133#comment-16827133
 ] 

ASF subversion and git services commented on IMPALA-8406:
-

Commit 114712031f2293af5b3d9509776f88c23d3fa0fc in impala's branch 
refs/heads/master from Todd Lipcon
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=1147120 ]

IMPALA-8454 (part 1): Refactor file descriptor loading code

This refactors various file-descriptor loading code out of HdfsTable
into new standalone classes. In order to support ACID tables, we'll need
to make various changes to these bits of code, and having them extracted
and cleaned up will make that easier.

This consolidates all of the places in which we list partition
directories into one method which does the appropriate thing regardless
of situation.

This has a small behavior change related to IMPALA-8406: previously, we
had a bug where, while refreshing a table, if one or more partitions
failed to refresh, the other partitions might still get refreshed
despite an error being returned. Those other partitions wouldn't be
available immediately until some other operation caused the table's
catalog version number to increase. This was buggy behavior.

Rather than tackle that problem in this "refactor" patch, this patch
just slightly improves the behavior: we'll either atomically update or
not update all partitions, but we might still add new partitions noticed
by the REFRESH, and might still update other HMS metadata.

This patch may end up slightly improving various other code paths that
refresh file descriptor lists. We used to have slightly different ways
of doing this in three different places, with different sets of
optimizations. Now we do it all in one place, and we pull all the known
tricks.

Change-Id: I59edf493b9ba38be5f556b4795a7684d9c9e3a07
Reviewed-on: http://gerrit.cloudera.org:8080/12950
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Failed REFRESH can partially modify table without bumping version number
> 
>
> Key: IMPALA-8406
> URL: https://issues.apache.org/jira/browse/IMPALA-8406
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 3.2.0
>Reporter: Todd Lipcon
>Priority: Major
>
> Currently, various incremental operations in the catalogd modify Table 
> objects in place, including REFRESH, which modifies each partition. In this 
> case, if one partition fails to refresh (eg due to incorrect partitions or 
> some other file access problem), other partitions can still be modified, 
> either because they were modified first (in a non-parallel operation) or 
> modified in parallel (for REFRESH).
> In this case, the REFRESH operation will throw an Exception back to the user, 
> but in fact it has modified the catalog entry. The version number, however, 
> is not bumped, which breaks some invariants of the catalog that an object 
> doesn't change without changing version numbers.
> This also produces some unexpected behavior such as:
> - SHOW FILES IN t;
> - REFRESH t; -- gets a failure
> - SHOW FILES in t; -- see the same result as originally
> - ALTER TABLE t SET UNCACHED; -- bumps the version number due to unrelated 
> operation
> - SHOW FILES IN t; -- the set of files has changed due to the earlier 
> partially-complete REFRESH



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8454) Recursively list files within transactional tables

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827132#comment-16827132
 ] 

ASF subversion and git services commented on IMPALA-8454:
-

Commit 114712031f2293af5b3d9509776f88c23d3fa0fc in impala's branch 
refs/heads/master from Todd Lipcon
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=1147120 ]

IMPALA-8454 (part 1): Refactor file descriptor loading code

This refactors various file-descriptor loading code out of HdfsTable
into new standalone classes. In order to support ACID tables, we'll need
to make various changes to these bits of code, and having them extracted
and cleaned up will make that easier.

This consolidates all of the places in which we list partition
directories into one method which does the appropriate thing regardless
of situation.

This has a small behavior change related to IMPALA-8406: previously, we
had a bug where, while refreshing a table, if one or more partitions
failed to refresh, the other partitions might still get refreshed
despite an error being returned. Those other partitions wouldn't be
available immediately until some other operation caused the table's
catalog version number to increase. This was buggy behavior.

Rather than tackle that problem in this "refactor" patch, this patch
just slightly improves the behavior: we'll either atomically update or
not update all partitions, but we might still add new partitions noticed
by the REFRESH, and might still update other HMS metadata.

This patch may end up slightly improving various other code paths that
refresh file descriptor lists. We used to have slightly different ways
of doing this in three different places, with different sets of
optimizations. Now we do it all in one place, and we pull all the known
tricks.

Change-Id: I59edf493b9ba38be5f556b4795a7684d9c9e3a07
Reviewed-on: http://gerrit.cloudera.org:8080/12950
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Recursively list files within transactional tables
> --
>
> Key: IMPALA-8454
> URL: https://issues.apache.org/jira/browse/IMPALA-8454
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Catalog
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
>Priority: Major
>
> For transactional tables, the data files are not directly within the 
> partition directories, but instead are stored within subdirectories 
> corresponding to writeIds, compactions, etc. To support this, we need to be 
> able to recursively load file lists within partition directories.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-4018) Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... FORMAT )

2019-04-26 Thread Greg Rahn (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16827103#comment-16827103
 ] 

Greg Rahn commented on IMPALA-4018:
---

[~gaborkaszab], [~maze] -

I'm strongly against introducing any behavior that makes the new 
CAST(...FORMAT...) syntax not follow the SQL:2016 contract, even via a feature 
flag.  I am, however, ok with introducing a flag to bring the legacy date masks 
and functions into ISO SQL compliance (and provide a UX that matches the 
expectations of DBMS users).  Here is the main reason why: SQL tools need to be 
"taught" how to convert to and from these types and formats and there has to be 
only correct way to do so. There is no way to teach a tool conditional logic 
based on feature flags. We're working with vendors to add the new SQL:2016 
syntax into their tools so there can be only one correct way and that way has 
to adhere to the contract that other SQL systems will adopt and use as well. I 
hope this makes it clear on why we need to follow and converge on ISO/ANSI SQL 
behaviors.



> Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... 
> FORMAT )
> 
>
> Key: IMPALA-4018
> URL: https://issues.apache.org/jira/browse/IMPALA-4018
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Frontend
>Affects Versions: Impala 2.2.4
>Reporter: Greg Rahn
>Assignee: Gabor Kaszab
>Priority: Critical
>  Labels: ansi-sql, compatibility, sql-language
>
> *Summary*
> The format masks/templates for currently are implemented using the [Java 
> SimpleDateFormat 
> patterns|http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html],
>  and although this is what Hive has implemented, it is not what most standard 
> SQL systems implement.  For example see 
> [Vertica|https://my.vertica.com/docs/7.2.x/HTML/Content/Authoring/SQLReferenceManual/Functions/Formatting/TemplatePatternsForDateTimeFormatting.htm],
>  
> [Netezza|http://www.ibm.com/support/knowledgecenter/SSULQD_7.2.1/com.ibm.nz.dbu.doc/r_dbuser_ntz_sql_extns_templ_patterns_date_time_conv.html],
>   
> [Oracle|https://docs.oracle.com/database/121/SQLRF/sql_elements004.htm#SQLRF00212],
>  and 
> [PostgreSQL|https://www.postgresql.org/docs/9.5/static/functions-formatting.html#FUNCTIONS-FORMATTING-DATETIME-TABLE].
>  
> *Examples of incompatibilities*
> {noformat}
> -- PostgreSQL/Netezza/Vertica/Oracle
> select to_timestamp('May 15, 2015 12:00:00', 'mon dd,  hh:mi:ss');
> -- Impala
> select to_timestamp('May 15, 2015 12:00:00', 'MMM dd,  HH:mm:ss');
> -- PostgreSQL/Netezza/Vertica/Oracle
> select to_timestamp('2015-02-14 20:19:07','-mm-dd hh24:mi:ss');
> -- Impala
> select to_timestamp('2015-02-14 20:19:07','-MM-dd HH:mm:ss');
> -- Vertica/Oracle
> select to_timestamp('2015-02-14 20:19:07.123456','-mm-dd hh24:mi:ss.ff');
> -- Impala
> select to_timestamp('2015-02-14 20:19:07.123456','-MM-dd 
> HH:mm:ss.SS');
> {noformat}
> *Considerations*
> Because this is a change in default behavior for to_timestamp(), if possible, 
> having a feature flag to revert to the legacy Java SimpleDateFormat patterns 
> should be strongly considered.  This would allow users to chose the behavior 
> they desire and scope it to a session if need be.
> SQL:2016 defines the following datetime templates
> {noformat}
>  ::=
>   {  }...
>  ::=
> 
>   | 
>  ::=
> 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>  ::=
> 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
> | 
>  ::=
>    | YYY | YY | Y
>  ::=
>    | RR
>  ::=
>   MM
>  ::=
>   DD
>  ::=
>   DDD
>  ::=
>   HH | HH12
>  ::=
>   HH24
>  ::=
>   MI
>  ::=
>   SS
>  ::=
>   S
>  ::=
>   FF1 | FF2 | FF3 | FF4 | FF5 | FF6 | FF7 | FF8 | FF9
>  ::=
>   A.M. | P.M.
>  ::=
>   TZH
>  ::=
>   TZM
> {noformat}
> SQL:2016 also introduced the FORMAT clause for CAST which is the standard way 
> to do string <> datetime conversions
> {noformat}
>  ::=
>   CAST 
>AS 
>   [ FORMAT  ]
>   
>  ::=
> 
>   | 
>  ::=
> 
> | 
>  ::=
>   
> {noformat}
> For example:
> {noformat}
> CAST( AS  [FORMAT ])
> CAST( AS  [FORMAT ])
> cast(dt as string format 'DD-MM-')
> cast('01-05-2017' as date format 'DD-MM-')
> {noformat}
> *Update*
> Here is the proposal for the new datetime patterns and their semantics:
> https://docs.google.com/document/d/1V7k6-lrPGW7_uhqM-FhKl3QsxwCRy69v2KIxPsGjc1k/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-4018) Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... FORMAT )

2019-04-26 Thread Gabor Kaszab (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gabor Kaszab updated IMPALA-4018:
-
Description: 
*Summary*
The format masks/templates for currently are implemented using the [Java 
SimpleDateFormat 
patterns|http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html],
 and although this is what Hive has implemented, it is not what most standard 
SQL systems implement.  For example see 
[Vertica|https://my.vertica.com/docs/7.2.x/HTML/Content/Authoring/SQLReferenceManual/Functions/Formatting/TemplatePatternsForDateTimeFormatting.htm],
 
[Netezza|http://www.ibm.com/support/knowledgecenter/SSULQD_7.2.1/com.ibm.nz.dbu.doc/r_dbuser_ntz_sql_extns_templ_patterns_date_time_conv.html],
  
[Oracle|https://docs.oracle.com/database/121/SQLRF/sql_elements004.htm#SQLRF00212],
 and 
[PostgreSQL|https://www.postgresql.org/docs/9.5/static/functions-formatting.html#FUNCTIONS-FORMATTING-DATETIME-TABLE].
 

*Examples of incompatibilities*
{noformat}
-- PostgreSQL/Netezza/Vertica/Oracle
select to_timestamp('May 15, 2015 12:00:00', 'mon dd,  hh:mi:ss');
-- Impala
select to_timestamp('May 15, 2015 12:00:00', 'MMM dd,  HH:mm:ss');

-- PostgreSQL/Netezza/Vertica/Oracle
select to_timestamp('2015-02-14 20:19:07','-mm-dd hh24:mi:ss');
-- Impala
select to_timestamp('2015-02-14 20:19:07','-MM-dd HH:mm:ss');

-- Vertica/Oracle
select to_timestamp('2015-02-14 20:19:07.123456','-mm-dd hh24:mi:ss.ff');
-- Impala
select to_timestamp('2015-02-14 20:19:07.123456','-MM-dd HH:mm:ss.SS');
{noformat}

*Considerations*
Because this is a change in default behavior for to_timestamp(), if possible, 
having a feature flag to revert to the legacy Java SimpleDateFormat patterns 
should be strongly considered.  This would allow users to chose the behavior 
they desire and scope it to a session if need be.

SQL:2016 defines the following datetime templates

{noformat}
 ::=
  {  }...
 ::=

  | 
 ::=

  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
 ::=

  | 
  | 
  | 
  | 
  | 
  | 
| 
 ::=
   | YYY | YY | Y
 ::=
   | RR
 ::=
  MM
 ::=
  DD
 ::=
  DDD
 ::=
  HH | HH12
 ::=
  HH24
 ::=
  MI
 ::=
  SS
 ::=
  S
 ::=
  FF1 | FF2 | FF3 | FF4 | FF5 | FF6 | FF7 | FF8 | FF9
 ::=
  A.M. | P.M.
 ::=
  TZH
 ::=
  TZM
{noformat}

SQL:2016 also introduced the FORMAT clause for CAST which is the standard way 
to do string <> datetime conversions
{noformat}
 ::=
  CAST 
   AS 
  [ FORMAT  ]
  
 ::=

  | 
 ::=

| 
 ::=
  
{noformat}
For example:
{noformat}
CAST( AS  [FORMAT ])
CAST( AS  [FORMAT ])
cast(dt as string format 'DD-MM-')
cast('01-05-2017' as date format 'DD-MM-')
{noformat}

*Update*
Here is the proposal for the new datetime patterns and their semantics:
https://docs.google.com/document/d/1V7k6-lrPGW7_uhqM-FhKl3QsxwCRy69v2KIxPsGjc1k/


  was:
*Summary*
The format masks/templates for currently are implemented using the [Java 
SimpleDateFormat 
patterns|http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html],
 and although this is what Hive has implemented, it is not what most standard 
SQL systems implement.  For example see 
[Vertica|https://my.vertica.com/docs/7.2.x/HTML/Content/Authoring/SQLReferenceManual/Functions/Formatting/TemplatePatternsForDateTimeFormatting.htm],
 
[Netezza|http://www.ibm.com/support/knowledgecenter/SSULQD_7.2.1/com.ibm.nz.dbu.doc/r_dbuser_ntz_sql_extns_templ_patterns_date_time_conv.html],
  
[Oracle|https://docs.oracle.com/database/121/SQLRF/sql_elements004.htm#SQLRF00212],
 and 
[PostgreSQL|https://www.postgresql.org/docs/9.5/static/functions-formatting.html#FUNCTIONS-FORMATTING-DATETIME-TABLE].
 

*Examples of incompatibilities*
{noformat}
-- PostgreSQL/Netezza/Vertica/Oracle
select to_timestamp('May 15, 2015 12:00:00', 'mon dd,  hh:mi:ss');
-- Impala
select to_timestamp('May 15, 2015 12:00:00', 'MMM dd,  HH:mm:ss');

-- PostgreSQL/Netezza/Vertica/Oracle
select to_timestamp('2015-02-14 20:19:07','-mm-dd hh24:mi:ss');
-- Impala
select to_timestamp('2015-02-14 20:19:07','-MM-dd HH:mm:ss');

-- Vertica/Oracle
select to_timestamp('2015-02-14 20:19:07.123456','-mm-dd hh24:mi:ss.ff');
-- Impala
select to_timestamp('2015-02-14 20:19:07.123456','-MM-dd HH:mm:ss.SS');
{noformat}

*Considerations*
Because this is a change in default behavior for to_timestamp(), if possible, 
having a feature flag to revert to the legacy Java SimpleDateFormat patterns 
should be strongly considered.  This would allow users to chose the behavior 
they desire and scope it to a session if need be.

SQL:2016 defines the following datetime templates

{noformat}
 ::=
  {  }...
 ::=

  | 
 ::=

  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
  | 
 ::=

  | 
  | 
  | 
  | 
  | 
  | 
| 
 ::=
   | YYY | YY | Y
 ::=
   | RR
 ::=
  MM
 ::=
  DD
 ::=
  DDD
 ::=
  HH | 

[jira] [Commented] (IMPALA-4018) Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... FORMAT )

2019-04-26 Thread Gabor Kaszab (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-4018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16826993#comment-16826993
 ] 

Gabor Kaszab commented on IMPALA-4018:
--

Hey [~grahn],

Meanwhile, I have reached out to the Spark community and [~maze] did an initial 
run through on the proposal doc. His initial opinion was that it feels 
unnatural to have the existing functionality toggle between the old and new 
patterns while CAST(..FORMAT..) uses the new patterns only. We agreed that this 
introduces a source of confusion on the user's side and would be much cleaner 
to implement the toggling mechanism for CAST(..FORMAT..) as well. This way all 
the functions within Impala would use the same pattern matching and the flag 
would change all of them and the users won't experience that e.g. 
to_timestamp() uses the old patterns while CAST(..FORMAT..) uses the new ones: 
Depending on the value of the flag all of them would use either the old or the 
new functionality.

What do you think? Can we revive the conversation we had around this?

Cheers,
Gabor


> Add support for SQL:2016 datetime templates/patterns/masks to CAST(... AS ... 
> FORMAT )
> 
>
> Key: IMPALA-4018
> URL: https://issues.apache.org/jira/browse/IMPALA-4018
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Frontend
>Affects Versions: Impala 2.2.4
>Reporter: Greg Rahn
>Assignee: Gabor Kaszab
>Priority: Critical
>  Labels: ansi-sql, compatibility, sql-language
>
> *Summary*
> The format masks/templates for currently are implemented using the [Java 
> SimpleDateFormat 
> patterns|http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html],
>  and although this is what Hive has implemented, it is not what most standard 
> SQL systems implement.  For example see 
> [Vertica|https://my.vertica.com/docs/7.2.x/HTML/Content/Authoring/SQLReferenceManual/Functions/Formatting/TemplatePatternsForDateTimeFormatting.htm],
>  
> [Netezza|http://www.ibm.com/support/knowledgecenter/SSULQD_7.2.1/com.ibm.nz.dbu.doc/r_dbuser_ntz_sql_extns_templ_patterns_date_time_conv.html],
>   
> [Oracle|https://docs.oracle.com/database/121/SQLRF/sql_elements004.htm#SQLRF00212],
>  and 
> [PostgreSQL|https://www.postgresql.org/docs/9.5/static/functions-formatting.html#FUNCTIONS-FORMATTING-DATETIME-TABLE].
>  
> *Examples of incompatibilities*
> {noformat}
> -- PostgreSQL/Netezza/Vertica/Oracle
> select to_timestamp('May 15, 2015 12:00:00', 'mon dd,  hh:mi:ss');
> -- Impala
> select to_timestamp('May 15, 2015 12:00:00', 'MMM dd,  HH:mm:ss');
> -- PostgreSQL/Netezza/Vertica/Oracle
> select to_timestamp('2015-02-14 20:19:07','-mm-dd hh24:mi:ss');
> -- Impala
> select to_timestamp('2015-02-14 20:19:07','-MM-dd HH:mm:ss');
> -- Vertica/Oracle
> select to_timestamp('2015-02-14 20:19:07.123456','-mm-dd hh24:mi:ss.ff');
> -- Impala
> select to_timestamp('2015-02-14 20:19:07.123456','-MM-dd 
> HH:mm:ss.SS');
> {noformat}
> *Considerations*
> Because this is a change in default behavior for to_timestamp(), if possible, 
> having a feature flag to revert to the legacy Java SimpleDateFormat patterns 
> should be strongly considered.  This would allow users to chose the behavior 
> they desire and scope it to a session if need be.
> SQL:2016 defines the following datetime templates
> {noformat}
>  ::=
>   {  }...
>  ::=
> 
>   | 
>  ::=
> 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
>  ::=
> 
>   | 
>   | 
>   | 
>   | 
>   | 
>   | 
> | 
>  ::=
>    | YYY | YY | Y
>  ::=
>    | RR
>  ::=
>   MM
>  ::=
>   DD
>  ::=
>   DDD
>  ::=
>   HH | HH12
>  ::=
>   HH24
>  ::=
>   MI
>  ::=
>   SS
>  ::=
>   S
>  ::=
>   FF1 | FF2 | FF3 | FF4 | FF5 | FF6 | FF7 | FF8 | FF9
>  ::=
>   A.M. | P.M.
>  ::=
>   TZH
>  ::=
>   TZM
> {noformat}
> SQL:2016 also introduced the FORMAT clause for CAST which is the standard way 
> to do string <> datetime conversions
> {noformat}
>  ::=
>   CAST 
>AS 
>   [ FORMAT  ]
>   
>  ::=
> 
>   | 
>  ::=
> 
> | 
>  ::=
>   
> {noformat}
> For example:
> {noformat}
> CAST( AS  [FORMAT ])
> CAST( AS  [FORMAT ])
> cast(dt as string format 'DD-MM-')
> cast('01-05-2017' as date format 'DD-MM-')
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8381) Remove branch from ParquetPlainEncoder::Decode()

2019-04-26 Thread Daniel Becker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8381?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16826924#comment-16826924
 ] 

Daniel Becker commented on IMPALA-8381:
---

We made some benchmarks comparing the old and the new implementations. The code 
for the benchmarks is not in the patch for this Jira as the benchmarks were 
originally written for the delta encoding (implementation in progress: 
[https://gerrit.cloudera.org/#/c/12621/]).

The benchmarks read plain encoded values to a "scratch batch" with the given 
stride. `OutType` means the type of the values in the scratch batch. We use the 
Impala benchmark framework. For the values, bigger is better.

Baseline (old implementation):
||ParquetType||OutType||Stride||p10||p50||p90||
|INT32|int32_t|8|136|140|141|
|INT32|int32_t|12|136|139|139|
|INT32|int32_t|16|137|139|140|
|INT32|int32_t|20|137|139|140|
|INT32|int32_t|30|125|128|129|
|INT32|int32_t|40|92.3|93.9|94.7|
|INT32|int32_t|50|72|72.7|73.1|
|INT32|int32_t|80|54|54.4|55|
|INT32|int32_t|100|51.7|52.2|52.7|
|INT32|int32_t|120|49.4|49.9|50.4|
|INT32|int32_t|150|46.3|46.6|47|
|INT32|int32_t|180|46.9|47.6|48.4|
|INT32|int32_t|200|46.9|47.9|48.6|
|INT32|int32_t|400|51.7|52.9|54.4|
||ParquetType||OutType||Stride||p10||p50||p90||
|INT64|int64_t|8|137|140|140|
|INT64|int64_t|12|135|140|141|
|INT64|int64_t|16|137|139|141|
|INT64|int64_t|20|132|138|139|
|INT64|int64_t|30|115|117|117|
|INT64|int64_t|40|88.2|89.5|90.1|
|INT64|int64_t|50|69.5|70|70.5|
|INT64|int64_t|80|52.7|53|53.4|
|INT64|int64_t|100|48|48.5|48.9|
|INT64|int64_t|120|48.1|48.6|49.1|
|INT64|int64_t|150|43.1|43.6|43.8|
|INT64|int64_t|180|42.2|42.4|43|
|INT64|int64_t|200|43.9|44.4|44.8|
|INT64|int64_t|400|42.5|42.8|43.2|

New implementation:
||ParquetType||OutType||Stride||p10||p50||p90||
|INT32|int32_t|8|281|284|286|
|INT32|int32_t|12|257|261|263|
|INT32|int32_t|16|237|240|242|
|INT32|int32_t|20|209|212|215|
|INT32|int32_t|30|136|138|139|
|INT32|int32_t|40|95.7|96.5|97.2|
|INT32|int32_t|50|73.2|73.9|74.5|
|INT32|int32_t|80|54|54.5|55.2|
|INT32|int32_t|100|51.9|52.4|52.8|
|INT32|int32_t|120|49.6|50|50.4|
|INT32|int32_t|150|45.9|46.6|47|
|INT32|int32_t|180|47.3|48.6|49.5|
|INT32|int32_t|200|47.4|48.6|49.7|
|INT32|int32_t|400|50.7|52.9|54.4|
||ParquetType||OutType||Stride||p10||p50||p90||
|INT64|int64_t|8|264|268|271|
|INT64|int64_t|12|221|224|226|
|INT64|int64_t|16|216|217|219|
|INT64|int64_t|20|184|186|188|
|INT64|int64_t|30|120|122|123|
|INT64|int64_t|40|90.2|91.3|91.9|
|INT64|int64_t|50|69.3|69.9|70.4|
|INT64|int64_t|80|52.7|53.2|53.6|
|INT64|int64_t|100|48.2|48.8|49.1|
|INT64|int64_t|120|48.2|48.7|49|
|INT64|int64_t|150|43.2|43.6|44|
|INT64|int64_t|180|42.2|42.7|43|
|INT64|int64_t|200|44|44.5|44.7|
|INT64|int64_t|400|43.9|44.5|45.1|

> Remove branch from ParquetPlainEncoder::Decode()
> 
>
> Key: IMPALA-8381
> URL: https://issues.apache.org/jira/browse/IMPALA-8381
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Csaba Ringhofer
>Assignee: Daniel Becker
>Priority: Minor
>  Labels: newbie, parquet, performance, ramp-up
>
> Removing the "if" at
> https://github.com/apache/impala/blob/5670f96b828d57f9e36510bb9af02bcc31de775c/be/src/exec/parquet/parquet-common.h#L203
> can lead to 1.5x speed up in plain decoding (type=int32, stride=16). For 
> primitive types, the same check can be done for a whole batch, so the speedup 
> can be gained for large batches without loosing safety. The only Parquet type 
> where this check is needed per element is BYTE_ARRAY (typically used for 
> STRING columns), which already has a template specialization for  
> ParquetPlainEncoder::Decode().



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8446) Create a unit test for Admission Controller

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16826686#comment-16826686
 ] 

ASF subversion and git services commented on IMPALA-8446:
-

Commit 2dda47ec27e8f15810907d77c6e41c275355cea6 in impala's branch 
refs/heads/master from Andrew Sherman
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=2dda47e ]

IMPALA-8446: Create a unit test for Admission Controller.

This test allows construction of white box tests that exercise Admission
Controller code.

The initial test cases are
(1) a test which does the following:
  + creates a RequestPoolService which reads some Admission Controller
configuration files
  + creates an Admission Controller object
  + creates a QuerySchedule for a query
  + tests if the query can be admitted by the Admission Controller
  + simulates activity in the cluster which consumes memory
  + tests that the earlier query cannot now be admitted by the Admission
Controller
(2) a test which verifies that the configuration files are read
correctly.

TESTING:
Ran end-to-end tests cleanly and checked that the new tests were
executed.

Change-Id: I8a840590b868f2df1a06f3f397b7b0fc2b29462c
Reviewed-on: http://gerrit.cloudera.org:8080/13078
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Create a unit test for Admission Controller
> ---
>
> Key: IMPALA-8446
> URL: https://issues.apache.org/jira/browse/IMPALA-8446
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> This will allow construction of white box tests that exercise Admission 
> Controller code.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8119) Improve experience for configuring JVM heap size in docker container.

2019-04-26 Thread Tim Armstrong (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-8119.
---
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Improve experience for configuring JVM heap size in docker container.
> -
>
> Key: IMPALA-8119
> URL: https://issues.apache.org/jira/browse/IMPALA-8119
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: docker
> Fix For: Impala 3.3.0
>
>
> The heap size can be configured via JAVA_TOOL_OPTIONS. Configuring the JVM 
> heap size is likely a common task when deploying Impala containers to 
> production because the catalogd and coordinators need to be sized to fit the 
> working set of metadata - it would be worthwhile to have a more explicit and 
> clearly-documented way to set the heap size.
> We could also try to be fancy and choose a heap size based on the container 
> memory limit and the role of the container.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8119) Improve experience for configuring JVM heap size in docker container.

2019-04-26 Thread ASF subversion and git services (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8119?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16826685#comment-16826685
 ] 

ASF subversion and git services commented on IMPALA-8119:
-

Commit 28d5a8299c27fd06779e108943a9aa128c5af15f in impala's branch 
refs/heads/master from Tim Armstrong
[ https://gitbox.apache.org/repos/asf?p=impala.git;h=28d5a82 ]

IMPALA-8119: document how to set heap size in docker

JAVA_TOOL_OPTIONS is a standard mechanism to pass arguments to a JVM.
Let's just document this as the canonical way to pass in the heap size.

Change-Id: Ie6ddba3c42a698b52d7c4e2ff6a9c73068e198b2
Reviewed-on: http://gerrit.cloudera.org:8080/13119
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Improve experience for configuring JVM heap size in docker container.
> -
>
> Key: IMPALA-8119
> URL: https://issues.apache.org/jira/browse/IMPALA-8119
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: docker
>
> The heap size can be configured via JAVA_TOOL_OPTIONS. Configuring the JVM 
> heap size is likely a common task when deploying Impala containers to 
> production because the catalogd and coordinators need to be sized to fit the 
> working set of metadata - it would be worthwhile to have a more explicit and 
> clearly-documented way to set the heap size.
> We could also try to be fancy and choose a heap size based on the container 
> memory limit and the role of the container.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org