[jira] [Commented] (IMPALA-9243) Coordinator Web UI should list which executors have been blacklisted

2019-12-16 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9243?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16997452#comment-16997452
 ] 

Lars Volker commented on IMPALA-9243:
-

I think it would be good to have a counter for the number of times that a node 
was blacklisted (or total time) in addition to a flag. That way it's possible 
to spot bad nodes and check for thresholds.

> Coordinator Web UI should list which executors have been blacklisted
> 
>
> Key: IMPALA-9243
> URL: https://issues.apache.org/jira/browse/IMPALA-9243
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: Sahil Takiar
>Priority: Major
>
> Currently, information about which nodes are blacklisted only shows up in 
> runtime profiles and Coordinator logs. It would be nice to display 
> blacklisting information in the Web UI as well so that a user can view which 
> nodes are blacklisted at any given time.
> One potential place to put the blacklisting information is in the /backends 
> page, which already lists out all the backends part of the cluster. A new 
> column called "Status" which can have values of either "Active" or 
> "Blacklisted" would be nice (perhaps we should re-factor the "Quiescing" 
> column to use the new "Status" column as well). This is similar to what the 
> Spark Web UI does for blacklisted nodes: 
> [https://ndu0e1pobsf1dobtvj5nls3q-wpengine.netdna-ssl.com/wp-content/uploads/2019/08/BLACKLIST-SCHEDULING.png]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9151) Number of executors during planning needs to account for suspended executor groupa

2019-12-03 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-9151.
-
Fix Version/s: Impala 3.4.0
   Resolution: Fixed

> Number of executors during planning needs to account for suspended executor 
> groupa
> --
>
> Key: IMPALA-9151
> URL: https://issues.apache.org/jira/browse/IMPALA-9151
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.4.0
>
>
> When configuring Impala with executor groups, the planner might see a 
> {{ExecutorMembershipSnapshot}} that has no executors in it. This can happen 
> if the first executor group has not started up yet, or if all executor groups 
> have been shutdown. If this happens, the planner will make sub-optimal 
> decisions, e.g. decide on a broadcast join vs a PHJ. In the former case, we 
> should have a configurable fallback cluster size to use during planning. In 
> the latter case, we should hang on to the last executor group size that we 
> had observed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9208) impala still failed to read orc format table using "enable_orc_scanner=true"

2019-12-03 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9208?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16987231#comment-16987231
 ] 

Lars Volker commented on IMPALA-9208:
-

You need to pass the flag as a parameter to the impalad when starting your 
cluster. Please use u...@impala.apache.org  for any issues with using Impala.

> impala still failed to read orc format table using "enable_orc_scanner=true"
> 
>
> Key: IMPALA-9208
> URL: https://issues.apache.org/jira/browse/IMPALA-9208
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 3.3.0
>Reporter: lv haiyang
>Priority: Major
>
> [bigdata@sdw1 ~]$ impala-shell --var=enable_orc_scanner=true
> Starting Impala Shell without Kerberos authentication
> Opened TCP connection to sdw1:21000
> Connected to sdw1:21000
> Server version: impalad version 3.4.0-SNAPSHOT RELEASE (build 
> 190ad89ebd46161f63a520c2a6d75ad56701e145)
> ***
> Welcome to the Impala shell.
> (Impala Shell v3.4.0-SNAPSHOT (190ad89) built on Tue Nov 19 08:52:31 UTC 2019)
> The SET command shows the current value of all shell and query options.
> ***
> [sdw1:21000] default> use company;
> Query: use company
> [sdw1:21000] company> select * from orc_table;
> Query: select * from orc_table
> Query submitted at: 2019-12-02 11:53:41 (Coordinator: http://sdw1:25000)
> ERROR: NotImplementedException: ORC scans are disabled by 
> --enable_orc_scanner flag



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-9202) Fix flakiness in TestExecutorGroups

2019-11-27 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9202?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-9202:

Labels: broken-build flaky flaky-test  (was: flaky-test)

> Fix flakiness in TestExecutorGroups
> ---
>
> Key: IMPALA-9202
> URL: https://issues.apache.org/jira/browse/IMPALA-9202
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.4.0
>Reporter: Bikramjeet Vig
>Assignee: Bikramjeet Vig
>Priority: Minor
>  Labels: broken-build, flaky, flaky-test
>
> test_executor_groups.TestExecutorGroups.test_admission_slots failed on an 
> assertion recently with the following stacktrace
> {noformat}
> custom_cluster/test_executor_groups.py:215: in test_admission_slots
> assert ("Initial admission queue reason: Not enough admission control 
> slots "
> E   assert 'Initial admission queue reason: Not enough admission control 
> slots available on host' in 'Query (id=0445eb75d4842dce:219df01e):\n  
> DEBUG MODE WARNING: Query profile created while running a DEBUG buil...0)\n   
>   - NumRowsFetchedFromCache: 0 (0)\n - RowMaterializationRate: 0\n - 
> RowMaterializationTimer: 0.000ns\n
> {noformat}
> On investigating the logs, it seems like the query did in fact get queued 
> with the expected reason. The only reason I can think of that it failed to 
> appear on profile is that the profile was fetched before the admission reason 
> could be added to the profile. This happened in an ASAN build so I am 
> assuming the slowness in execution contributed to widening the window in 
> which this can happen.
> from the logs:
> {noformat}
> I1104 18:18:34.144309 113361 impala-server.cc:1046] 
> 0445eb75d4842dce:219df01e] Registered query 
> query_id=0445eb75d4842dce:219df01e 
> session_id=da467385483f4fb3:16683a81d25fe79e
> I1104 18:18:34.144951 113361 Frontend.java:1256] 
> 0445eb75d4842dce:219df01e] Analyzing query: select * from 
> functional_parquet.alltypestiny  where month < 3 and id + 
> random() < sleep(500); db: default
> I1104 18:18:34.149049 113361 BaseAuthorizationChecker.java:96] 
> 0445eb75d4842dce:219df01e] Authorization check took 4 ms
> I1104 18:18:34.149219 113361 Frontend.java:1297] 
> 0445eb75d4842dce:219df01e] Analysis and authorization finished.
> I1104 18:18:34.163229 113885 scheduler.cc:548] 
> 0445eb75d4842dce:219df01e] Exec at coord is false
> I1104 18:18:34.163945 113885 admission-controller.cc:1295] 
> 0445eb75d4842dce:219df01e] Trying to admit 
> id=0445eb75d4842dce:219df01e in pool_name=default-pool 
> executor_group_name=default-pool-group1 per_host_mem_estimate=176.02 MB 
> dedicated_coord_mem_estimate=100.02 MB max_requests=-1 (configured 
> statically) max_queued=200 (configured statically) max_mem=-1.00 B 
> (configured statically)
> I1104 18:18:34.164203 113885 admission-controller.cc:1307] 
> 0445eb75d4842dce:219df01e] Stats: agg_num_running=1, 
> agg_num_queued=0, agg_mem_reserved=8.00 KB,  
> local_host(local_mem_admitted=452.05 MB, num_admitted_running=1, 
> num_queued=0, backend_mem_reserved=8.00 KB)
> I1104 18:18:34.164383 113885 admission-controller.cc:902] 
> 0445eb75d4842dce:219df01e] Queuing, query 
> id=0445eb75d4842dce:219df01e reason: Not enough admission control 
> slots available on host 
> impala-ec2-centos74-r4-4xlarge-ondemand-1f88.vpc.cloudera.com:22002. Needed 1 
> slots but 1/1 are already in use.
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8863) Add support to run tests over HS2-HTTP

2019-11-27 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8863.
-
Fix Version/s: Impala 3.4.0
   Resolution: Fixed

> Add support to run tests over HS2-HTTP
> --
>
> Key: IMPALA-8863
> URL: https://issues.apache.org/jira/browse/IMPALA-8863
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: impyla
> Fix For: Impala 3.4.0
>
>
> Once https://github.com/cloudera/impyla/issues/357 gets addressed, we should 
> run at least some of our tests over hs2-http using Impyla to better test 
> Impyla's HTTP endpoint support.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8863) Add support to run tests over HS2-HTTP

2019-11-27 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8863?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8863:

Target Version: Impala 3.4.0  (was: Impala 3.3.0)

> Add support to run tests over HS2-HTTP
> --
>
> Key: IMPALA-8863
> URL: https://issues.apache.org/jira/browse/IMPALA-8863
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: impyla
> Fix For: Impala 3.4.0
>
>
> Once https://github.com/cloudera/impyla/issues/357 gets addressed, we should 
> run at least some of our tests over hs2-http using Impyla to better test 
> Impyla's HTTP endpoint support.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9122) TestEventProcessing.test_insert_events flaky in precommit

2019-11-26 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16982732#comment-16982732
 ] 

Lars Volker commented on IMPALA-9122:
-

Hit this again here: https://jenkins.impala.io/job/gerrit-verify-dryrun/5292/

> TestEventProcessing.test_insert_events flaky in precommit
> -
>
> Key: IMPALA-9122
> URL: https://issues.apache.org/jira/browse/IMPALA-9122
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Reporter: Tim Armstrong
>Assignee: Anurag Mantripragada
>Priority: Critical
>  Labels: broken-build, flaky
>
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/8613/
> {noformat}
> custom_cluster.test_event_processing.TestEventProcessing.test_insert_events 
> (from pytest)
> Failing for the past 1 build (Since Failed#8613 )
> Took 1 min 20 sec.
> add description
> Error Message
> assert  >(18421) 
> is True  +  where  TestEventProcessing.wait_for_insert_event_processing of 
> > = 
>  0x7f7fa86ec6d0>.wait_for_insert_event_processing
> Stacktrace
> custom_cluster/test_event_processing.py:82: in test_insert_events
> self.run_test_insert_events()
> custom_cluster/test_event_processing.py:143: in run_test_insert_events
> assert self.wait_for_insert_event_processing(last_synced_event_id) is True
> E   assert  of  0x7f7fa86ec6d0>>(18421) is True
> E+  where  TestEventProcessing.wait_for_insert_event_processing of 
> > = 
>  0x7f7fa86ec6d0>.wait_for_insert_event_processing
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9032) Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms timeout

2019-11-20 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-9032.
-
Resolution: Duplicate

> Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms 
> timeout
> -
>
> Key: IMPALA-9032
> URL: https://issues.apache.org/jira/browse/IMPALA-9032
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Lars Volker
>Priority: Major
>
> This looks like a bug to me but I'm not entirely sure. I'm trying to run our 
> tests over hs2-http (IMPALA-8863) and after the change for IMPALA-7312 to 
> introduce a non-blocking mode for FetchResults() it looks like we sometimes 
> return an empty result way before {{fetch_rows_timeout_ms}} has elapsed. This 
> triggers a bug in Impyla 
> ([#369|https://github.com/cloudera/impyla/issues/369]), but it also seems 
> like something we should investigate and fix in Impala.
> {noformat}
> I1007 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697988 56527 scheduler.cc:468] 6d4cba4d2e8ccc42:66ce26a8] 
> Exec at coord is falseI1007 22:10:10.698014 54090 impala-hs2-server.cc:663] 
> GetOperationStatus(): query_id=0d43fd73ce4403fd:da25dde9I1007 
> 22:10:10.698173   127 control-service.cc:142] 
> 0646e91fd6a0a953:02949ff3] ExecQueryFInstances(): 
> query_id=0646e91fd6a0a953:02949ff3 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.698356 56527 admission-controller.cc:1270] 
> 6d4cba4d2e8ccc42:66ce26a8] Trying to admit 
> id=6d4cba4d2e8ccc42:66ce26a8 in pool_name=root.default 
> executor_group_name=default per_host_mem_estimate=52.02 MB 
> dedicated_coord_mem_estimate=110.02 MB max_requests=-1 (configured 
> statically) max_queued=200 (configured statically) max_mem=29.30 GB 
> (configured statically)I1007 22:10:10.698386 56527 
> admission-controller.cc:1282] 6d4cba4d2e8ccc42:66ce26a8] Stats: 
> agg_num_running=9, agg_num_queued=0, agg_mem_reserved=8.34 GB,  
> local_host(local_mem_admitted=9.09 GB, num_admitted_running=9, num_queued=0, 
> backend_mem_reserved=6.70 GB)I1007 22:10:10.698415 56527 
> admission-controller.cc:871] 6d4cba4d2e8ccc42:66ce26a8] Admitting 
> query id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.698479 56527 
> impala-server.cc:1713] 6d4cba4d2e8ccc42:66ce26a8] Registering query 
> locationsI1007 22:10:10.698529 56527 coordinator.cc:97] 
> 6d4cba4d2e8ccc42:66ce26a8] Exec() 
> query_id=6d4cba4d2e8ccc42:66ce26a8 stmt=select count(*) from alltypes 
> where month=1I1007 22:10:10.698992 56527 coordinator.cc:361] 
> 6d4cba4d2e8ccc42:66ce26a8] starting execution on 3 backends for 
> query_id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.699383 56523 
> coordinator.cc:375] 0646e91fd6a0a953:02949ff3] started execution on 1 
> backends for query_id=0646e91fd6a0a953:02949ff3I1007 22:10:10.699409 
> 56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec at coord is 
> falseI1007 22:10:10.700017   127 control-service.cc:142] 
> 6d4cba4d2e8ccc42:66ce26a8] ExecQueryFInstances(): 
> query_id=6d4cba4d2e8ccc42:66ce26a8 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.700147 56534 scheduler.cc:468] 
> e1495f928c2cd4f6:eeda82aa] Exec at coord is falseI1007 
> 22:10:10.700234   325 TAcceptQueueServer.cpp:340] New connection to server 
> hiveserver2-http-frontend from client I1007 
> 22:10:10.700286   329 TAcceptQueueServer.cpp:227] TAcceptQueueServer: 
> hiveserver2-http-frontend started connection setup for client  172.18.0.1 Port: 51580>I1007 22:10:10.700314   329 
> TAcceptQueueServer.cpp:245] TAcceptQueueServer: hiveserver2-http-frontend 
> finished connection setup for client I1007 
> 22:10:10.700371 56550 impala-hs2-server.cc:844] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c #results=1 has_more=trueI1007 
> 22:10:10.700508 56551 impala-server.cc:1969] Connection 
> 8249c7defcb10124:1bc65ed9ea562aab from client 172.18.0.1:51576 to server 
> hiveserver2-http-frontend closed. The connection had 1 associated 
> session(s).I1007 22:10:10.700688 53748 impala-beeswax-server.cc:260] close(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700711 53748 
> impala-server.cc:1129] UnregisterQuery(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700721 53748 
> impala-server.cc:1234] Cancel(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700742 53748 
> coordinator.cc:716] 

[jira] [Commented] (IMPALA-9091) query_test.test_scanners.TestScannerReservation.test_scanners flaky

2019-11-20 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16978689#comment-16978689
 ] 

Lars Volker commented on IMPALA-9091:
-

I hit this again here: 
https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/8899/testReport/junit/query_test.test_scanners/TestScannerReservation/test_scanners_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/

> query_test.test_scanners.TestScannerReservation.test_scanners flaky
> ---
>
> Key: IMPALA-9091
> URL: https://issues.apache.org/jira/browse/IMPALA-9091
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Reporter: Tim Armstrong
>Assignee: Zoltán Borók-Nagy
>Priority: Critical
>  Labels: flaky
>
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/8463
> {noformat}
> E   AssertionError: Did not find matches for lines in runtime profile:
> E   EXPECTED LINES:
> E   row_regex:.*ParquetRowGroupIdealReservation.*Avg: 3.50 MB.*
> E   
> E   ACTUAL PROFILE:
> E   Query (id=3b48738ce971e36b:b6f52bf5):
> E DEBUG MODE WARNING: Query profile created while running a DEBUG build 
> of Impala. Use RELEASE builds to measure query performance.
> E Summary:
> E   Session ID: 2e4c96b22f2ac6e3:88afa967b63e7983
> E   Session Type: BEESWAX
> E   Start Time: 2019-10-24 21:22:06.311001000
> E   End Time: 2019-10-24 21:22:06.520778000
> E   Query Type: QUERY
> E   Query State: FINISHED
> E   Query Status: OK
> E   Impala Version: impalad version 3.4.0-SNAPSHOT DEBUG (build 
> 8c60e91f7c3812aca14739535a994d21c51fc0b0)
> E   User: ubuntu
> E   Connected User: ubuntu
> E   Delegated User: 
> E   Network Address: :::127.0.0.1:37312
> E   Default Db: functional
> E   Sql Statement: select * from tpch_parquet.lineitem
> E   where l_orderkey < 10
> E   Coordinator: ip-172-31-20-105:22000
> E   Query Options (set by configuration): 
> ABORT_ON_ERROR=1,EXEC_SINGLE_NODE_ROWS_THRESHOLD=0,DISABLE_CODEGEN_ROWS_THRESHOLD=0,TIMEZONE=Universal,CLIENT_IDENTIFIER=query_test/test_scanners.py::TestScannerReservation::()::test_scanners[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'exec_single_node_rows_threshold':0}|table_form
> E   Query Options (set by configuration and planner): 
> ABORT_ON_ERROR=1,EXEC_SINGLE_NODE_ROWS_THRESHOLD=0,MT_DOP=0,DISABLE_CODEGEN_ROWS_THRESHOLD=0,TIMEZONE=Universal,CLIENT_IDENTIFIER=query_test/test_scanners.py::TestScannerReservation::()::test_scanners[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'exec_single_node_rows_threshold':0}|table_form
> E   Plan: 
> E   
> E   Max Per-Host Resource Reservation: Memory=40.00MB Threads=3
> E   Per-Host Resource Estimates: Memory=1.26GB
> E   Analyzed query: SELECT * FROM tpch_parquet.lineitem WHERE l_orderkey < 
> CAST(10
> E   AS BIGINT)
> E   
> E   F01:PLAN FRAGMENT [UNPARTITIONED] hosts=1 instances=1
> E   |  Per-Host Resources: mem-estimate=10.69MB mem-reservation=0B 
> thread-reservation=1
> E   PLAN-ROOT SINK
> E   |  output exprs: tpch_parquet.lineitem.l_orderkey, 
> tpch_parquet.lineitem.l_partkey, tpch_parquet.lineitem.l_suppkey, 
> tpch_parquet.lineitem.l_linenumber, tpch_parquet.lineitem.l_quantity, 
> tpch_parquet.lineitem.l_extendedprice, tpch_parquet.lineitem.l_discount, 
> tpch_parquet.lineitem.l_tax, tpch_parquet.lineitem.l_returnflag, 
> tpch_parquet.lineitem.l_linestatus, tpch_parquet.lineitem.l_shipdate, 
> tpch_parquet.lineitem.l_commitdate, tpch_parquet.lineitem.l_receiptdate, 
> tpch_parquet.lineitem.l_shipinstruct, tpch_parquet.lineitem.l_shipmode, 
> tpch_parquet.lineitem.l_comment
> E   |  mem-estimate=0B mem-reservation=0B thread-reservation=0
> E   |
> E   01:EXCHANGE [UNPARTITIONED]
> E   |  mem-estimate=10.69MB mem-reservation=0B thread-reservation=0
> E   |  tuple-ids=0 row-size=231B cardinality=600.12K
> E   |  in pipelines: 00(GETNEXT)
> E   |
> E   F00:PLAN FRAGMENT [RANDOM] hosts=3 instances=3
> E   Per-Host Resources: mem-estimate=1.25GB mem-reservation=40.00MB 
> thread-reservation=2
> E   00:SCAN HDFS [tpch_parquet.lineitem, RANDOM]
> E  HDFS partitions=1/1 files=3 size=193.97MB
> E  predicates: l_orderkey < CAST(10 AS BIGINT)
> E  stored statistics:
> Etable: rows=6.00M size=193.97MB
> Ecolumns: all
> E  extrapolated-rows=disabled max-scan-range-rows=2.14M
> E  parquet statistics predicates: l_orderkey < CAST(10 AS 

[jira] [Commented] (IMPALA-9148) AuthorizationStmtTest.testColumnMaskEnabled seems flaky

2019-11-19 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16977794#comment-16977794
 ] 

Lars Volker commented on IMPALA-9148:
-

His this again here: 
https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/8877/

> AuthorizationStmtTest.testColumnMaskEnabled seems flaky
> ---
>
> Key: IMPALA-9148
> URL: https://issues.apache.org/jira/browse/IMPALA-9148
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Reporter: Fang-Yu Rao
>Assignee: Fang-Yu Rao
>Priority: Major
>
> Recently we have seen failed {{AuthorizationStmtTest.testColumnMaskEnabled}} 
> occasionally with the following error messages (e.g., 
> [https://jenkins.impala.io/job/gerrit-verify-dryrun/5194/consoleFull]).
> {code:java}
> Impala does not support row filtering yet. Row filtering is enabled on table: 
> functional.alltypes_view
> expected:
> Impala does not support column masking yet. Column masking is enabled on 
> column: functional.alltypes_view.string_col
> {code}
> Taking a look at the {{testColumnMaskEnabled()}}, we can see the related SQL 
> statement is
> {code:java}
> select string_col from functional.alltypes_view;
> {code}
> I found that for this SQL statement, {{authorizeRowFilterAndColumnMask()}} in 
> {{RangerAuthorizationCheker.java}} will be called first 
> ([https://github.com/apache/impala/blob/master/fe/src/main/java/org/apache/impala/authorization/ranger/RangerAuthorizationChecker.java#L183-L200]).
>  There will be two privilege requests, one request for column, and the other 
> for table. The function {{authorizeRowFilter()}} is the only function that 
> could produce the error message above 
> ([https://github.com/apache/impala/blob/master/fe/src/main/java/org/apache/impala/authorization/ranger/RangerAuthorizationChecker.java#L295-L308]).
>  Specifically, this error would be generated if 
> {{plugin_.evalRowFilterPolicies(req, null).isRowFilterEnabled()}} returns 
> true 
> ([https://github.com/apache/impala/blob/master/fe/src/main/java/org/apache/impala/authorization/ranger/RangerAuthorizationChecker.java#L303]).
> I have taken a brief look at {{isRowFilterEnabled()}}, and found that it will 
> return true only if there is some policy on the Ranger server that specifies 
> the policy of row filtering (according to my current understanding). However, 
> in {{testColumnMaskEnabled()}} 
> ([https://github.com/apache/impala/blob/master/fe/src/test/java/org/apache/impala/authorization/AuthorizationStmtTest.java#L2836]),
>  we only add a policy for column masking. Therefore, I suspect it may be 
> possible that some other tests added to the Ranger server some policy for row 
> filtering but did not properly do the cleanup of this row filtering policy 
> afterwards.
> To address this issue, we should add some logic to clean up the policies 
> stored on the Ranger server before running this JUnit test. This JUnit test 
> assumes that the Ranger server does not store any policies related to column 
> masking and row filtering before the testing.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-7550) Create documentation for all profile fields

2019-11-18 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7550?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker reassigned IMPALA-7550:
---

Assignee: Jiawei Wang  (was: Lars Volker)

> Create documentation for all profile fields
> ---
>
> Key: IMPALA-7550
> URL: https://issues.apache.org/jira/browse/IMPALA-7550
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Docs
>Affects Versions: Impala 3.1.0
>Reporter: Lars Volker
>Assignee: Jiawei Wang
>Priority: Major
>  Labels: supportability
>
> It would be great to have some documentation for all profile fields. 
> Currently users often have to look at the source code to figure out the exact 
> meaning of a field. Instead we should investigate ways to generate 
> documentation on the fields, e.g. by cleaning and automatically extracting 
> relevant comments from the source code.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-9155) Single Admission Controller per Cluster

2019-11-13 Thread Lars Volker (Jira)
Lars Volker created IMPALA-9155:
---

 Summary: Single Admission Controller per Cluster
 Key: IMPALA-9155
 URL: https://issues.apache.org/jira/browse/IMPALA-9155
 Project: IMPALA
  Issue Type: New Feature
  Components: Backend
Affects Versions: Impala 3.4.0
Reporter: Lars Volker
Assignee: Lars Volker


We should consider building a single admission control service that can admit 
queries for multiple coordinators. This would simplify the current approach 
where coordinators make distributed decisions, remove the risk of overadmission 
by concurrent decisions, and allow us to implement more sophisticated admission 
control strategies.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-9151) Number of executors during planning needs to account for suspended executor groupa

2019-11-12 Thread Lars Volker (Jira)
Lars Volker created IMPALA-9151:
---

 Summary: Number of executors during planning needs to account for 
suspended executor groupa
 Key: IMPALA-9151
 URL: https://issues.apache.org/jira/browse/IMPALA-9151
 Project: IMPALA
  Issue Type: Bug
  Components: Frontend
Reporter: Lars Volker
Assignee: Lars Volker


When configuring Impala with executor groups, the planner might see a 
{{ExecutorMembershipSnapshot}} that has no executors in it. This can happen if 
the first executor group has not started up yet, or if all executor groups have 
been shutdown. If this happens, the planner will make sub-optimal decisions, 
e.g. decide on a broadcast join vs a PHJ. In the former case, we should have a 
configurable fallback cluster size to use during planning. In the latter case, 
we should hang on to the last executor group size that we had observed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9032) Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms timeout

2019-10-09 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16948030#comment-16948030
 ] 

Lars Volker commented on IMPALA-9032:
-

Thanks, I'll take a look whether it still happens with that change. Can you 
elaborate what the problem with the conversion was? I had taken a look at the 
original code and it seems ok to me.

> Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms 
> timeout
> -
>
> Key: IMPALA-9032
> URL: https://issues.apache.org/jira/browse/IMPALA-9032
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Lars Volker
>Priority: Major
>
> This looks like a bug to me but I'm not entirely sure. I'm trying to run our 
> tests over hs2-http (IMPALA-8863) and after the change for IMPALA-7312 to 
> introduce a non-blocking mode for FetchResults() it looks like we sometimes 
> return an empty result way before {{fetch_rows_timeout_ms}} has elapsed. This 
> triggers a bug in Impyla 
> ([#369|https://github.com/cloudera/impyla/issues/369]), but it also seems 
> like something we should investigate and fix in Impala.
> {noformat}
> I1007 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697988 56527 scheduler.cc:468] 6d4cba4d2e8ccc42:66ce26a8] 
> Exec at coord is falseI1007 22:10:10.698014 54090 impala-hs2-server.cc:663] 
> GetOperationStatus(): query_id=0d43fd73ce4403fd:da25dde9I1007 
> 22:10:10.698173   127 control-service.cc:142] 
> 0646e91fd6a0a953:02949ff3] ExecQueryFInstances(): 
> query_id=0646e91fd6a0a953:02949ff3 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.698356 56527 admission-controller.cc:1270] 
> 6d4cba4d2e8ccc42:66ce26a8] Trying to admit 
> id=6d4cba4d2e8ccc42:66ce26a8 in pool_name=root.default 
> executor_group_name=default per_host_mem_estimate=52.02 MB 
> dedicated_coord_mem_estimate=110.02 MB max_requests=-1 (configured 
> statically) max_queued=200 (configured statically) max_mem=29.30 GB 
> (configured statically)I1007 22:10:10.698386 56527 
> admission-controller.cc:1282] 6d4cba4d2e8ccc42:66ce26a8] Stats: 
> agg_num_running=9, agg_num_queued=0, agg_mem_reserved=8.34 GB,  
> local_host(local_mem_admitted=9.09 GB, num_admitted_running=9, num_queued=0, 
> backend_mem_reserved=6.70 GB)I1007 22:10:10.698415 56527 
> admission-controller.cc:871] 6d4cba4d2e8ccc42:66ce26a8] Admitting 
> query id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.698479 56527 
> impala-server.cc:1713] 6d4cba4d2e8ccc42:66ce26a8] Registering query 
> locationsI1007 22:10:10.698529 56527 coordinator.cc:97] 
> 6d4cba4d2e8ccc42:66ce26a8] Exec() 
> query_id=6d4cba4d2e8ccc42:66ce26a8 stmt=select count(*) from alltypes 
> where month=1I1007 22:10:10.698992 56527 coordinator.cc:361] 
> 6d4cba4d2e8ccc42:66ce26a8] starting execution on 3 backends for 
> query_id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.699383 56523 
> coordinator.cc:375] 0646e91fd6a0a953:02949ff3] started execution on 1 
> backends for query_id=0646e91fd6a0a953:02949ff3I1007 22:10:10.699409 
> 56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec at coord is 
> falseI1007 22:10:10.700017   127 control-service.cc:142] 
> 6d4cba4d2e8ccc42:66ce26a8] ExecQueryFInstances(): 
> query_id=6d4cba4d2e8ccc42:66ce26a8 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.700147 56534 scheduler.cc:468] 
> e1495f928c2cd4f6:eeda82aa] Exec at coord is falseI1007 
> 22:10:10.700234   325 TAcceptQueueServer.cpp:340] New connection to server 
> hiveserver2-http-frontend from client I1007 
> 22:10:10.700286   329 TAcceptQueueServer.cpp:227] TAcceptQueueServer: 
> hiveserver2-http-frontend started connection setup for client  172.18.0.1 Port: 51580>I1007 22:10:10.700314   329 
> TAcceptQueueServer.cpp:245] TAcceptQueueServer: hiveserver2-http-frontend 
> finished connection setup for client I1007 
> 22:10:10.700371 56550 impala-hs2-server.cc:844] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c #results=1 has_more=trueI1007 
> 22:10:10.700508 56551 impala-server.cc:1969] Connection 
> 8249c7defcb10124:1bc65ed9ea562aab from client 172.18.0.1:51576 to server 
> hiveserver2-http-frontend closed. The connection had 1 associated 
> session(s).I1007 22:10:10.700688 53748 impala-beeswax-server.cc:260] close(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700711 53748 
> impala-server.cc:1129] UnregisterQuery(): 
> 

[jira] [Commented] (IMPALA-9032) Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms timeout

2019-10-09 Thread Lars Volker (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16947845#comment-16947845
 ] 

Lars Volker commented on IMPALA-9032:
-

CC [~stakiar], [~tarmstrong]

> Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms 
> timeout
> -
>
> Key: IMPALA-9032
> URL: https://issues.apache.org/jira/browse/IMPALA-9032
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Lars Volker
>Priority: Major
>
> This looks like a bug to me but I'm not entirely sure. I'm trying to run our 
> tests over hs2-http (IMPALA-8863) and after the change for IMPALA-7312 to 
> introduce a non-blocking mode for FetchResults() it looks like we sometimes 
> return an empty result way before {{fetch_rows_timeout_ms}} has elapsed. This 
> triggers a bug in Impyla 
> ([#369|https://github.com/cloudera/impyla/issues/369]), but it also seems 
> like something we should investigate and fix in Impala.
> {noformat}
> I1007 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697988 56527 scheduler.cc:468] 6d4cba4d2e8ccc42:66ce26a8] 
> Exec at coord is falseI1007 22:10:10.698014 54090 impala-hs2-server.cc:663] 
> GetOperationStatus(): query_id=0d43fd73ce4403fd:da25dde9I1007 
> 22:10:10.698173   127 control-service.cc:142] 
> 0646e91fd6a0a953:02949ff3] ExecQueryFInstances(): 
> query_id=0646e91fd6a0a953:02949ff3 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.698356 56527 admission-controller.cc:1270] 
> 6d4cba4d2e8ccc42:66ce26a8] Trying to admit 
> id=6d4cba4d2e8ccc42:66ce26a8 in pool_name=root.default 
> executor_group_name=default per_host_mem_estimate=52.02 MB 
> dedicated_coord_mem_estimate=110.02 MB max_requests=-1 (configured 
> statically) max_queued=200 (configured statically) max_mem=29.30 GB 
> (configured statically)I1007 22:10:10.698386 56527 
> admission-controller.cc:1282] 6d4cba4d2e8ccc42:66ce26a8] Stats: 
> agg_num_running=9, agg_num_queued=0, agg_mem_reserved=8.34 GB,  
> local_host(local_mem_admitted=9.09 GB, num_admitted_running=9, num_queued=0, 
> backend_mem_reserved=6.70 GB)I1007 22:10:10.698415 56527 
> admission-controller.cc:871] 6d4cba4d2e8ccc42:66ce26a8] Admitting 
> query id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.698479 56527 
> impala-server.cc:1713] 6d4cba4d2e8ccc42:66ce26a8] Registering query 
> locationsI1007 22:10:10.698529 56527 coordinator.cc:97] 
> 6d4cba4d2e8ccc42:66ce26a8] Exec() 
> query_id=6d4cba4d2e8ccc42:66ce26a8 stmt=select count(*) from alltypes 
> where month=1I1007 22:10:10.698992 56527 coordinator.cc:361] 
> 6d4cba4d2e8ccc42:66ce26a8] starting execution on 3 backends for 
> query_id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.699383 56523 
> coordinator.cc:375] 0646e91fd6a0a953:02949ff3] started execution on 1 
> backends for query_id=0646e91fd6a0a953:02949ff3I1007 22:10:10.699409 
> 56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec at coord is 
> falseI1007 22:10:10.700017   127 control-service.cc:142] 
> 6d4cba4d2e8ccc42:66ce26a8] ExecQueryFInstances(): 
> query_id=6d4cba4d2e8ccc42:66ce26a8 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.700147 56534 scheduler.cc:468] 
> e1495f928c2cd4f6:eeda82aa] Exec at coord is falseI1007 
> 22:10:10.700234   325 TAcceptQueueServer.cpp:340] New connection to server 
> hiveserver2-http-frontend from client I1007 
> 22:10:10.700286   329 TAcceptQueueServer.cpp:227] TAcceptQueueServer: 
> hiveserver2-http-frontend started connection setup for client  172.18.0.1 Port: 51580>I1007 22:10:10.700314   329 
> TAcceptQueueServer.cpp:245] TAcceptQueueServer: hiveserver2-http-frontend 
> finished connection setup for client I1007 
> 22:10:10.700371 56550 impala-hs2-server.cc:844] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c #results=1 has_more=trueI1007 
> 22:10:10.700508 56551 impala-server.cc:1969] Connection 
> 8249c7defcb10124:1bc65ed9ea562aab from client 172.18.0.1:51576 to server 
> hiveserver2-http-frontend closed. The connection had 1 associated 
> session(s).I1007 22:10:10.700688 53748 impala-beeswax-server.cc:260] close(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700711 53748 
> impala-server.cc:1129] UnregisterQuery(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700721 53748 
> impala-server.cc:1234] Cancel(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 

[jira] [Updated] (IMPALA-9032) Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms timeout

2019-10-09 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-9032:

Summary: Impala returns 0 rows over hs2-http without waiting for 
fetch_rows_timeout_ms timeout  (was: Impala returns 0 rows over hs2-http 
without waiting for TODO timeout)

> Impala returns 0 rows over hs2-http without waiting for fetch_rows_timeout_ms 
> timeout
> -
>
> Key: IMPALA-9032
> URL: https://issues.apache.org/jira/browse/IMPALA-9032
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Lars Volker
>Priority: Major
>
> This looks like a bug to me but I'm not entirely sure. I'm trying to run our 
> tests over hs2-http (IMPALA-8863) and after the change for IMPALA-7312 to 
> introduce a non-blocking mode for FetchResults() it looks like we sometimes 
> return an empty result way before {{fetch_rows_timeout_ms}} has elapsed. This 
> triggers a bug in Impyla 
> ([#369|https://github.com/cloudera/impyla/issues/369]), but it also seems 
> like something we should investigate and fix in Impala.
> {noformat}
> I1007 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 
> 22:10:10.697988 56527 scheduler.cc:468] 6d4cba4d2e8ccc42:66ce26a8] 
> Exec at coord is falseI1007 22:10:10.698014 54090 impala-hs2-server.cc:663] 
> GetOperationStatus(): query_id=0d43fd73ce4403fd:da25dde9I1007 
> 22:10:10.698173   127 control-service.cc:142] 
> 0646e91fd6a0a953:02949ff3] ExecQueryFInstances(): 
> query_id=0646e91fd6a0a953:02949ff3 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.698356 56527 admission-controller.cc:1270] 
> 6d4cba4d2e8ccc42:66ce26a8] Trying to admit 
> id=6d4cba4d2e8ccc42:66ce26a8 in pool_name=root.default 
> executor_group_name=default per_host_mem_estimate=52.02 MB 
> dedicated_coord_mem_estimate=110.02 MB max_requests=-1 (configured 
> statically) max_queued=200 (configured statically) max_mem=29.30 GB 
> (configured statically)I1007 22:10:10.698386 56527 
> admission-controller.cc:1282] 6d4cba4d2e8ccc42:66ce26a8] Stats: 
> agg_num_running=9, agg_num_queued=0, agg_mem_reserved=8.34 GB,  
> local_host(local_mem_admitted=9.09 GB, num_admitted_running=9, num_queued=0, 
> backend_mem_reserved=6.70 GB)I1007 22:10:10.698415 56527 
> admission-controller.cc:871] 6d4cba4d2e8ccc42:66ce26a8] Admitting 
> query id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.698479 56527 
> impala-server.cc:1713] 6d4cba4d2e8ccc42:66ce26a8] Registering query 
> locationsI1007 22:10:10.698529 56527 coordinator.cc:97] 
> 6d4cba4d2e8ccc42:66ce26a8] Exec() 
> query_id=6d4cba4d2e8ccc42:66ce26a8 stmt=select count(*) from alltypes 
> where month=1I1007 22:10:10.698992 56527 coordinator.cc:361] 
> 6d4cba4d2e8ccc42:66ce26a8] starting execution on 3 backends for 
> query_id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.699383 56523 
> coordinator.cc:375] 0646e91fd6a0a953:02949ff3] started execution on 1 
> backends for query_id=0646e91fd6a0a953:02949ff3I1007 22:10:10.699409 
> 56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec at coord is 
> falseI1007 22:10:10.700017   127 control-service.cc:142] 
> 6d4cba4d2e8ccc42:66ce26a8] ExecQueryFInstances(): 
> query_id=6d4cba4d2e8ccc42:66ce26a8 coord=b04a12d76e27:22000 
> #instances=1I1007 22:10:10.700147 56534 scheduler.cc:468] 
> e1495f928c2cd4f6:eeda82aa] Exec at coord is falseI1007 
> 22:10:10.700234   325 TAcceptQueueServer.cpp:340] New connection to server 
> hiveserver2-http-frontend from client I1007 
> 22:10:10.700286   329 TAcceptQueueServer.cpp:227] TAcceptQueueServer: 
> hiveserver2-http-frontend started connection setup for client  172.18.0.1 Port: 51580>I1007 22:10:10.700314   329 
> TAcceptQueueServer.cpp:245] TAcceptQueueServer: hiveserver2-http-frontend 
> finished connection setup for client I1007 
> 22:10:10.700371 56550 impala-hs2-server.cc:844] FetchResults(): 
> query_id=764d4313dbc64e20:2831560c #results=1 has_more=trueI1007 
> 22:10:10.700508 56551 impala-server.cc:1969] Connection 
> 8249c7defcb10124:1bc65ed9ea562aab from client 172.18.0.1:51576 to server 
> hiveserver2-http-frontend closed. The connection had 1 associated 
> session(s).I1007 22:10:10.700688 53748 impala-beeswax-server.cc:260] close(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700711 53748 
> impala-server.cc:1129] UnregisterQuery(): 
> query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700721 

[jira] [Created] (IMPALA-9032) Impala returns 0 rows over hs2-http without waiting for TODO timeout

2019-10-09 Thread Lars Volker (Jira)
Lars Volker created IMPALA-9032:
---

 Summary: Impala returns 0 rows over hs2-http without waiting for 
TODO timeout
 Key: IMPALA-9032
 URL: https://issues.apache.org/jira/browse/IMPALA-9032
 Project: IMPALA
  Issue Type: Bug
  Components: Backend
Affects Versions: Impala 3.4.0
Reporter: Lars Volker


This looks like a bug to me but I'm not entirely sure. I'm trying to run our 
tests over hs2-http (IMPALA-8863) and after the change for IMPALA-7312 to 
introduce a non-blocking mode for FetchResults() it looks like we sometimes 
return an empty result way before {{fetch_rows_timeout_ms}} has elapsed. This 
triggers a bug in Impyla 
([#369|https://github.com/cloudera/impyla/issues/369]), but it also seems like 
something we should investigate and fix in Impala.

{noformat}
I1007 22:10:10.697760 56550 impala-hs2-server.cc:821] FetchResults(): 
query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 22:10:10.697760 
56550 impala-hs2-server.cc:821] FetchResults(): 
query_id=764d4313dbc64e20:2831560c fetch_size=1024I1007 22:10:10.697988 
56527 scheduler.cc:468] 6d4cba4d2e8ccc42:66ce26a8] Exec at coord is 
falseI1007 22:10:10.698014 54090 impala-hs2-server.cc:663] 
GetOperationStatus(): query_id=0d43fd73ce4403fd:da25dde9I1007 
22:10:10.698173   127 control-service.cc:142] 
0646e91fd6a0a953:02949ff3] ExecQueryFInstances(): 
query_id=0646e91fd6a0a953:02949ff3 coord=b04a12d76e27:22000 
#instances=1I1007 22:10:10.698356 56527 admission-controller.cc:1270] 
6d4cba4d2e8ccc42:66ce26a8] Trying to admit 
id=6d4cba4d2e8ccc42:66ce26a8 in pool_name=root.default 
executor_group_name=default per_host_mem_estimate=52.02 MB 
dedicated_coord_mem_estimate=110.02 MB max_requests=-1 (configured statically) 
max_queued=200 (configured statically) max_mem=29.30 GB (configured 
statically)I1007 22:10:10.698386 56527 admission-controller.cc:1282] 
6d4cba4d2e8ccc42:66ce26a8] Stats: agg_num_running=9, agg_num_queued=0, 
agg_mem_reserved=8.34 GB,  local_host(local_mem_admitted=9.09 GB, 
num_admitted_running=9, num_queued=0, backend_mem_reserved=6.70 GB)I1007 
22:10:10.698415 56527 admission-controller.cc:871] 
6d4cba4d2e8ccc42:66ce26a8] Admitting query 
id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.698479 56527 
impala-server.cc:1713] 6d4cba4d2e8ccc42:66ce26a8] Registering query 
locationsI1007 22:10:10.698529 56527 coordinator.cc:97] 
6d4cba4d2e8ccc42:66ce26a8] Exec() 
query_id=6d4cba4d2e8ccc42:66ce26a8 stmt=select count(*) from alltypes 
where month=1I1007 22:10:10.698992 56527 coordinator.cc:361] 
6d4cba4d2e8ccc42:66ce26a8] starting execution on 3 backends for 
query_id=6d4cba4d2e8ccc42:66ce26a8I1007 22:10:10.699383 56523 
coordinator.cc:375] 0646e91fd6a0a953:02949ff3] started execution on 1 
backends for query_id=0646e91fd6a0a953:02949ff3I1007 22:10:10.699409 
56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec at coord is 
falseI1007 22:10:10.700017   127 control-service.cc:142] 
6d4cba4d2e8ccc42:66ce26a8] ExecQueryFInstances(): 
query_id=6d4cba4d2e8ccc42:66ce26a8 coord=b04a12d76e27:22000 
#instances=1I1007 22:10:10.700147 56534 scheduler.cc:468] 
e1495f928c2cd4f6:eeda82aa] Exec at coord is falseI1007 22:10:10.700234  
 325 TAcceptQueueServer.cpp:340] New connection to server 
hiveserver2-http-frontend from client I1007 
22:10:10.700286   329 TAcceptQueueServer.cpp:227] TAcceptQueueServer: 
hiveserver2-http-frontend started connection setup for client I1007 22:10:10.700314   329 TAcceptQueueServer.cpp:245] 
TAcceptQueueServer: hiveserver2-http-frontend finished connection setup for 
client I1007 22:10:10.700371 56550 
impala-hs2-server.cc:844] FetchResults(): 
query_id=764d4313dbc64e20:2831560c #results=1 has_more=trueI1007 
22:10:10.700508 56551 impala-server.cc:1969] Connection 
8249c7defcb10124:1bc65ed9ea562aab from client 172.18.0.1:51576 to server 
hiveserver2-http-frontend closed. The connection had 1 associated 
session(s).I1007 22:10:10.700688 53748 impala-beeswax-server.cc:260] close(): 
query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700711 53748 
impala-server.cc:1129] UnregisterQuery(): 
query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700721 53748 
impala-server.cc:1234] Cancel(): 
query_id=e9473ff80c5d4afe:733cefe0I1007 22:10:10.700742 53748 
coordinator.cc:716] CancelBackends() 
query_id=e9473ff80c5d4afe:733cefe0, tried to cancel 0 backendsI1007 
22:10:10.700690 56534 scheduler.cc:468] e1495f928c2cd4f6:eeda82aa] Exec 
at coord is falseI1007 22:10:10.701387 56534 admission-controller.cc:1270] 
e1495f928c2cd4f6:eeda82aa] Trying to admit 
id=e1495f928c2cd4f6:eeda82aa in pool_name=root.default 
executor_group_name=default per_host_mem_estimate=231.95 MB 

[jira] [Resolved] (IMPALA-8973) Update Kudu version to fix openssl1.1.1 compatibility issue

2019-10-02 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8973?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8973.
-
Resolution: Fixed

> Update Kudu version to fix openssl1.1.1 compatibility issue
> ---
>
> Key: IMPALA-8973
> URL: https://issues.apache.org/jira/browse/IMPALA-8973
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.4.0
>Reporter: Kurt Deschler
>Assignee: Kurt Deschler
>Priority: Major
>  Labels: kudu
> Fix For: Impala 3.4.0
>
>
> openssl1.1.1/TLS1.3 exposed an issue with Kudu connectivity that was resolved 
> https://issues.apache.org/jira/browse/KUDU-2871.
> This issue was observed causing test failures in kudu tests:
> MESSAGE: ImpalaRuntimeException: Error creating Kudu table 
> 'impala::tpch_kudu.lineitem'
> CAUSED BY: NonRecoverableException: not enough live tablet servers to create 
> a table with the requested replication factor 3; 0 tablet servers are alive
> 16:37:21.564919  7150 [heartbeater.cc:566]|http://heartbeater.cc:566]/] 
> Failed to heartbeat to 127.0.0.1:7051 (0 consecutive failures): Network 
> error: Failed to ping master at 127.0.0.1:7051: Client connection negotiation 
> failed: client connection to 127.0.0.1:7051: connect: Connection refused 
> (error 111)
> logs/data_loading/impalad.kurt-cldr.kdeschle.log.INFO.XXX
> "not enough live tablet servers"
>  
> impala-config.sh needs to be updated to pull in a newer version of Kudu that 
> has this fix.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8967) impala-shell does not trim ldap password whitespace

2019-09-23 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8967:

Labels: newbie ramp-up usability  (was: usability)

> impala-shell does not trim ldap password whitespace
> ---
>
> Key: IMPALA-8967
> URL: https://issues.apache.org/jira/browse/IMPALA-8967
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Priority: Major
>  Labels: newbie, ramp-up, usability
>
> We should trim whitespace off the LDAP password instead of issuing a warning 
> (_Warning: LDAP password contains a trailing newline. Did you use ‘echo’ 
> instead of ‘echo -n’?`_).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8967) impala-shell does not trim ldap password whitespace

2019-09-23 Thread Lars Volker (Jira)
Lars Volker created IMPALA-8967:
---

 Summary: impala-shell does not trim ldap password whitespace
 Key: IMPALA-8967
 URL: https://issues.apache.org/jira/browse/IMPALA-8967
 Project: IMPALA
  Issue Type: Bug
  Components: Infrastructure
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


We should trim whitespace off the LDAP password instead of issuing a warning 
(_Warning: LDAP password contains a trailing newline. Did you use ‘echo’ 
instead of ‘echo -n’?`_).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8936) Make queuing reason for unhealthy executor groups more generic

2019-09-15 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8936?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8936.
-
Fix Version/s: Impala 3.4.0
   Resolution: Fixed

> Make queuing reason for unhealthy executor groups more generic
> --
>
> Key: IMPALA-8936
> URL: https://issues.apache.org/jira/browse/IMPALA-8936
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: user-experience
> Fix For: Impala 3.4.0
>
>
> In some situations, users might actually expect not having a healthy executor 
> group around, e.g. when they're starting one and it takes a while to come 
> online. We should make the queuing reason more generic and drop the 
> "unhealthy" concept from it to reduce confusion.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8936) Make queuing reason for unhealthy executor groups more generic

2019-09-10 Thread Lars Volker (Jira)
Lars Volker created IMPALA-8936:
---

 Summary: Make queuing reason for unhealthy executor groups more 
generic
 Key: IMPALA-8936
 URL: https://issues.apache.org/jira/browse/IMPALA-8936
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Reporter: Lars Volker
Assignee: Lars Volker


In some situations, users might actually expect not having a healthy executor 
group around, e.g. when they're starting one and it takes a while to come 
online. We should make the queuing reason more generic and drop the "unhealthy" 
concept from it to reduce confusion.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8802) Switch to pgrep for graceful_shutdown_backends.sh

2019-08-29 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8802.
-
Fix Version/s: Impala 3.3.0
   Resolution: Fixed

> Switch to pgrep for graceful_shutdown_backends.sh
> -
>
> Key: IMPALA-8802
> URL: https://issues.apache.org/jira/browse/IMPALA-8802
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.3.0
>
>
> IMPALA-8798 added a script with a call to {{pidof}}. However, {{pgrep}} seems 
> generally preferred (https://mywiki.wooledge.org/BadUtils) and we should 
> switch to it.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8892) Add tools to Docker images

2019-08-29 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8892.
-
Fix Version/s: Impala 3.4.0
   Resolution: Fixed

> Add tools to Docker images
> --
>
> Key: IMPALA-8892
> URL: https://issues.apache.org/jira/browse/IMPALA-8892
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: supportability
> Fix For: Impala 3.4.0
>
>
> Our docker images lack a bunch of tools that help to implement health checks 
> and debugging issues.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8895) Expose daemon health on /healthz endpoint

2019-08-27 Thread Lars Volker (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8895.
-
Fix Version/s: Impala 3.4.0
   Resolution: Fixed

> Expose daemon health on /healthz endpoint
> -
>
> Key: IMPALA-8895
> URL: https://issues.apache.org/jira/browse/IMPALA-8895
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: observability
> Fix For: Impala 3.4.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8900) Allow /healthz access without authentication

2019-08-27 Thread Lars Volker (Jira)
Lars Volker created IMPALA-8900:
---

 Summary: Allow /healthz access without authentication
 Key: IMPALA-8900
 URL: https://issues.apache.org/jira/browse/IMPALA-8900
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.4.0
Reporter: Lars Volker


When enabling SPNEGO authentication for the debug webpages, /healthz becomes 
unavailable. Some tooling might rely on the endpoint being accessible without 
authentication and it does not pose a security risk to make it available.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8895) Expose daemon health on /healthz endpoint

2019-08-26 Thread Lars Volker (Jira)
Lars Volker created IMPALA-8895:
---

 Summary: Expose daemon health on /healthz endpoint
 Key: IMPALA-8895
 URL: https://issues.apache.org/jira/browse/IMPALA-8895
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.4.0
Reporter: Lars Volker
Assignee: Lars Volker






--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8892) Add tools to Docker images

2019-08-26 Thread Lars Volker (Jira)
Lars Volker created IMPALA-8892:
---

 Summary: Add tools to Docker images
 Key: IMPALA-8892
 URL: https://issues.apache.org/jira/browse/IMPALA-8892
 Project: IMPALA
  Issue Type: Improvement
Reporter: Lars Volker
Assignee: Lars Volker


Our docker images lack a bunch of tools that help to implement health checks 
and debugging issues.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-7070) Failed test: query_test.test_nested_types.TestParquetArrayEncodings.test_thrift_array_of_arrays on S3

2019-08-18 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-7070.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Failed test: 
> query_test.test_nested_types.TestParquetArrayEncodings.test_thrift_array_of_arrays
>  on S3
> -
>
> Key: IMPALA-7070
> URL: https://issues.apache.org/jira/browse/IMPALA-7070
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.0
>Reporter: Dimitris Tsirogiannis
>Priority: Critical
>  Labels: broken-build, flaky, s3, test-failure
> Fix For: Impala 3.3.0
>
>
>  
> {code:java}
> Error Message
> query_test/test_nested_types.py:406: in test_thrift_array_of_arrays "col1 
> array>") query_test/test_nested_types.py:579: in 
> _create_test_table check_call(["hadoop", "fs", "-put", local_path, 
> location], shell=False) /usr/lib64/python2.6/subprocess.py:505: in check_call 
> raise CalledProcessError(retcode, cmd) E   CalledProcessError: Command 
> '['hadoop', 'fs', '-put', 
> '/data/jenkins/workspace/impala-asf-2.x-core-s3/repos/Impala/testdata/parquet_nested_types_encodings/bad-thrift.parquet',
>  
> 's3a://impala-cdh5-s3-test/test-warehouse/test_thrift_array_of_arrays_11da5fde.db/ThriftArrayOfArrays']'
>  returned non-zero exit status 1
> Stacktrace
> query_test/test_nested_types.py:406: in test_thrift_array_of_arrays
> "col1 array>")
> query_test/test_nested_types.py:579: in _create_test_table
> check_call(["hadoop", "fs", "-put", local_path, location], shell=False)
> /usr/lib64/python2.6/subprocess.py:505: in check_call
> raise CalledProcessError(retcode, cmd)
> E   CalledProcessError: Command '['hadoop', 'fs', '-put', 
> '/data/jenkins/workspace/impala-asf-2.x-core-s3/repos/Impala/testdata/parquet_nested_types_encodings/bad-thrift.parquet',
>  
> 's3a://impala-cdh5-s3-test/test-warehouse/test_thrift_array_of_arrays_11da5fde.db/ThriftArrayOfArrays']'
>  returned non-zero exit status 1
> Standard Error
> SET sync_ddl=False;
> -- executing against localhost:21000
> DROP DATABASE IF EXISTS `test_thrift_array_of_arrays_11da5fde` CASCADE;
> SET sync_ddl=False;
> -- executing against localhost:21000
> CREATE DATABASE `test_thrift_array_of_arrays_11da5fde`;
> MainThread: Created database "test_thrift_array_of_arrays_11da5fde" for test 
> ID 
> "query_test/test_nested_types.py::TestParquetArrayEncodings::()::test_thrift_array_of_arrays[exec_option:
>  {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 
> 'disable_codegen': False, 'abort_on_error': 1, 'debug_action': None, 
> 'exec_single_node_rows_threshold': 0} | table_format: parquet/none]"
> -- executing against localhost:21000
> create table test_thrift_array_of_arrays_11da5fde.ThriftArrayOfArrays (col1 
> array>) stored as parquet location 
> 's3a://impala-cdh5-s3-test/test-warehouse/test_thrift_array_of_arrays_11da5fde.db/ThriftArrayOfArrays';
> 18/05/20 18:31:03 WARN impl.MetricsConfig: Cannot locate configuration: tried 
> hadoop-metrics2-s3a-file-system.properties,hadoop-metrics2.properties
> 18/05/20 18:31:03 INFO impl.MetricsSystemImpl: Scheduled snapshot period at 
> 10 second(s).
> 18/05/20 18:31:03 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
> started
> 18/05/20 18:31:06 INFO Configuration.deprecation: 
> fs.s3a.server-side-encryption-key is deprecated. Instead, use 
> fs.s3a.server-side-encryption.key
> put: rename 
> `s3a://impala-cdh5-s3-test/test-warehouse/test_thrift_array_of_arrays_11da5fde.db/ThriftArrayOfArrays/bad-thrift.parquet._COPYING_'
>  to 
> `s3a://impala-cdh5-s3-test/test-warehouse/test_thrift_array_of_arrays_11da5fde.db/ThriftArrayOfArrays/bad-thrift.parquet':
>  Input/output error
> 18/05/20 18:31:08 INFO impl.MetricsSystemImpl: Stopping s3a-file-system 
> metrics system...
> 18/05/20 18:31:08 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
> stopped.
> 18/05/20 18:31:08 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
> shutdown complete.{code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7117) Lower debug level for HDFS S3 connector back to INFO

2019-08-18 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16910063#comment-16910063
 ] 

Lars Volker commented on IMPALA-7117:
-

I think we can just keep it at Debug since it only affects our own tests and 
no-one objected so far. Closing this one for now.

> Lower debug level for HDFS S3 connector back to INFO
> 
>
> Key: IMPALA-7117
> URL: https://issues.apache.org/jira/browse/IMPALA-7117
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 2.13.0, Impala 3.1.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Blocker
>  Labels: s3
>
> This change will increase the log level for the HDFS S3 connector to DEBUG to 
> help with IMPALA-6910 and IMPALA-7070. Before the next release we need to 
> lower it again.
> https://gerrit.cloudera.org/#/c/10596/
> I'm making this a P1 to remind us that we must do this before cutting a 
> release.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Closed] (IMPALA-7117) Lower debug level for HDFS S3 connector back to INFO

2019-08-18 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7117?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker closed IMPALA-7117.
---
Resolution: Won't Fix

> Lower debug level for HDFS S3 connector back to INFO
> 
>
> Key: IMPALA-7117
> URL: https://issues.apache.org/jira/browse/IMPALA-7117
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 2.13.0, Impala 3.1.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Blocker
>  Labels: s3
>
> This change will increase the log level for the HDFS S3 connector to DEBUG to 
> help with IMPALA-6910 and IMPALA-7070. Before the next release we need to 
> lower it again.
> https://gerrit.cloudera.org/#/c/10596/
> I'm making this a P1 to remind us that we must do this before cutting a 
> release.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8863) Add support to run tests over HS2-HTTP

2019-08-14 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8863:
---

 Summary: Add support to run tests over HS2-HTTP
 Key: IMPALA-8863
 URL: https://issues.apache.org/jira/browse/IMPALA-8863
 Project: IMPALA
  Issue Type: Improvement
  Components: Infrastructure
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Lars Volker


Once https://github.com/cloudera/impyla/issues/357 gets addressed, we should 
run at least some of our tests over hs2-http using Impyla to better test 
Impyla's HTTP endpoint support.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8852) ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads configuration"

2019-08-12 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8852:

Affects Version/s: Impala 3.3.0

> ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads 
> configuration"
> 
>
> Key: IMPALA-8852
> URL: https://issues.apache.org/jira/browse/IMPALA-8852
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0, Impala 3.3.0
>Reporter: Adriano
>Priority: Major
>  Labels: ramp-up
>
> On coordinator only nodes ([typically the edge 
> nodes|https://www.cloudera.com/documentation/enterprise/5-15-x/topics/impala_dedicated_coordinator.html#concept_omm_gf1_n2b]):
> {code:java}
> --is_coordinator=true
> --is_executor=false
> {code}
> the *dfs.domain.socket.path* (can be nonexistent on the local FS as the 
> Datanode role eventually is not installed on the edge node).
> The non existing path prevent the ImpalaD to start with the message:
> {code:java}
> I0809 04:15:53.899714 25364 status.cc:124] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> @   0xb35f19
> @  0x100e2fe
> @  0x103f274
> @  0x102836f
> @   0xa9f573
> @ 0x7f97807e93d4
> @   0xafb3b8
> E0809 04:15:53.899749 25364 impala-server.cc:278] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> {code}
> despite a coordinator-only ImpalaD does not do short circuit reads.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8852) ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads configuration"

2019-08-12 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8852:

Priority: Major  (was: Minor)

> ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads 
> configuration"
> 
>
> Key: IMPALA-8852
> URL: https://issues.apache.org/jira/browse/IMPALA-8852
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Adriano
>Priority: Major
>  Labels: ramp-up
>
> On coordinator only nodes ([typically the edge 
> nodes|https://www.cloudera.com/documentation/enterprise/5-15-x/topics/impala_dedicated_coordinator.html#concept_omm_gf1_n2b]):
> {code:java}
> --is_coordinator=true
> --is_executor=false
> {code}
> the *dfs.domain.socket.path* (can be nonexistent on the local FS as the 
> Datanode role eventually is not installed on the edge node).
> The non existing path prevent the ImpalaD to start with the message:
> {code:java}
> I0809 04:15:53.899714 25364 status.cc:124] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> @   0xb35f19
> @  0x100e2fe
> @  0x103f274
> @  0x102836f
> @   0xa9f573
> @ 0x7f97807e93d4
> @   0xafb3b8
> E0809 04:15:53.899749 25364 impala-server.cc:278] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> {code}
> despite a coordinator-only ImpalaD does not do short circuit reads.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8852) ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads configuration"

2019-08-12 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8852?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16905325#comment-16905325
 ] 

Lars Volker commented on IMPALA-8852:
-

Thanks for filing this issue. As a permanent solution we should only emit a 
warning when the socket cannot be found and {{-is_executor=false}}.

> ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads 
> configuration"
> 
>
> Key: IMPALA-8852
> URL: https://issues.apache.org/jira/browse/IMPALA-8852
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Adriano
>Priority: Minor
>
> On coordinator only nodes ([typically the edge 
> nodes|https://www.cloudera.com/documentation/enterprise/5-15-x/topics/impala_dedicated_coordinator.html#concept_omm_gf1_n2b]):
> {code:java}
> --is_coordinator=true
> --is_executor=false
> {code}
> the *dfs.domain.socket.path* (can be nonexistent on the local FS as the 
> Datanode role eventually is not installed on the edge node).
> The non existing path prevent the ImpalaD to start with the message:
> {code:java}
> I0809 04:15:53.899714 25364 status.cc:124] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> @   0xb35f19
> @  0x100e2fe
> @  0x103f274
> @  0x102836f
> @   0xa9f573
> @ 0x7f97807e93d4
> @   0xafb3b8
> E0809 04:15:53.899749 25364 impala-server.cc:278] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> {code}
> despite a coordinator-only ImpalaD does not do short circuit reads.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8852) ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads configuration"

2019-08-12 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8852:

Labels: ramp-up  (was: )

> ImpalaD fail to start on a non-datanode with "Invalid short-circuit reads 
> configuration"
> 
>
> Key: IMPALA-8852
> URL: https://issues.apache.org/jira/browse/IMPALA-8852
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Adriano
>Priority: Minor
>  Labels: ramp-up
>
> On coordinator only nodes ([typically the edge 
> nodes|https://www.cloudera.com/documentation/enterprise/5-15-x/topics/impala_dedicated_coordinator.html#concept_omm_gf1_n2b]):
> {code:java}
> --is_coordinator=true
> --is_executor=false
> {code}
> the *dfs.domain.socket.path* (can be nonexistent on the local FS as the 
> Datanode role eventually is not installed on the edge node).
> The non existing path prevent the ImpalaD to start with the message:
> {code:java}
> I0809 04:15:53.899714 25364 status.cc:124] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> @   0xb35f19
> @  0x100e2fe
> @  0x103f274
> @  0x102836f
> @   0xa9f573
> @ 0x7f97807e93d4
> @   0xafb3b8
> E0809 04:15:53.899749 25364 impala-server.cc:278] Invalid short-circuit reads 
> configuration:
>   - Impala cannot read or execute the parent directory of 
> dfs.domain.socket.path
> {code}
> despite a coordinator-only ImpalaD does not do short circuit reads.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8798) TestAutoScaling does not work on erasure-coded files

2019-07-31 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8798?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8798.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> TestAutoScaling does not work on erasure-coded files
> 
>
> Key: IMPALA-8798
> URL: https://issues.apache.org/jira/browse/IMPALA-8798
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
>  Labels: scalability
> Fix For: Impala 3.3.0
>
>
> TestAutoScaling uses the ConcurrentWorkload class, which does not set the 
> required query option to support scanning erasure-coded files. We should 
> disable the test for those cases.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8802) Switch to pgrep for graceful_shutdown_backends.sh

2019-07-29 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8802:
---

 Summary: Switch to pgrep for graceful_shutdown_backends.sh
 Key: IMPALA-8802
 URL: https://issues.apache.org/jira/browse/IMPALA-8802
 Project: IMPALA
  Issue Type: Bug
  Components: Infrastructure
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Lars Volker


IMPALA-8798 added a script with a call to {{pidof}}. However, {{pgrep}} seems 
generally preferred (https://mywiki.wooledge.org/BadUtils) and we should switch 
to it.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8798) TestAutoScaling does not work on erasure-coded files

2019-07-26 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8798:
---

 Summary: TestAutoScaling does not work on erasure-coded files
 Key: IMPALA-8798
 URL: https://issues.apache.org/jira/browse/IMPALA-8798
 Project: IMPALA
  Issue Type: Bug
  Components: Infrastructure
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Lars Volker


TestAutoScaling uses the ConcurrentWorkload class, which does not set the 
required query option to support scanning erasure-coded files. We should 
disable the test for those cases.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8789) Include script to trigger graceful shutdown in docker containers

2019-07-24 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8789:
---

 Summary: Include script to trigger graceful shutdown in docker 
containers
 Key: IMPALA-8789
 URL: https://issues.apache.org/jira/browse/IMPALA-8789
 Project: IMPALA
  Issue Type: Improvement
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Lars Volker


We should include a utility script in the docker containers to trigger a 
graceful shutdown by sending SIGRTMIN to all impalads.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8610) Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8610.
-
   Resolution: Fixed
Fix Version/s: Not Applicable

It seems that my change to the configuration was sufficient, at least we 
haven't seen this again since then.

> Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs
> -
>
> Key: IMPALA-8610
> URL: https://issues.apache.org/jira/browse/IMPALA-8610
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure, Jenkins
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
> Fix For: Not Applicable
>
>
> It seems that otherwise successful runs of {{ubuntu-16.04-from-scratch}} can 
> fail at the end when copying log files in the end. If the impalads are still 
> running, they can rotate a logfile while {{cp}} tries to copy it, which 
> results in output like this ([affected 
> job|https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/console]):
> {noformat}
> 00:24:28 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102026':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072304.100810':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102023':
>  No such file or directory
> {noformat}
> Note that the script is currently configured in Jenkins to run in a shell 
> with {{-e}} and that a successful run contains additional lines after the 
> {{cp}} command:
> {noformat}
> 04:16:18 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 04:16:21 + rm -f /tmp/git-err-ksiZp.log
> {noformat}
> As a fix we should kill all daemons before copying the files.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-8610) Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-8610 started by Lars Volker.
---
> Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs
> -
>
> Key: IMPALA-8610
> URL: https://issues.apache.org/jira/browse/IMPALA-8610
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure, Jenkins
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
>
> It seems that otherwise successful runs of {{ubuntu-16.04-from-scratch}} can 
> fail at the end when copying log files in the end. If the impalads are still 
> running, they can rotate a logfile while {{cp}} tries to copy it, which 
> results in output like this ([affected 
> job|https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/console]):
> {noformat}
> 00:24:28 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102026':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072304.100810':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102023':
>  No such file or directory
> {noformat}
> Note that the script is currently configured in Jenkins to run in a shell 
> with {{-e}} and that a successful run contains additional lines after the 
> {{cp}} command:
> {noformat}
> 04:16:18 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 04:16:21 + rm -f /tmp/git-err-ksiZp.log
> {noformat}
> As a fix we should kill all daemons before copying the files.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Closed] (IMPALA-8596) TestObservability.test_global_exchange_counters failed in ASAN

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker closed IMPALA-8596.
---
Resolution: Cannot Reproduce

This hasn't shown up again since May. Closing it for now; we can always re-open 
if we see if again

> TestObservability.test_global_exchange_counters failed in ASAN
> --
>
> Key: IMPALA-8596
> URL: https://issues.apache.org/jira/browse/IMPALA-8596
> Project: IMPALA
>  Issue Type: Bug
>  Components: Distributed Exec
>Affects Versions: Impala 3.3.0
>Reporter: Zoltán Borók-Nagy
>Assignee: Lars Volker
>Priority: Blocker
>  Labels: broken-build, flaky
>
> Seen in an ASAN build: 
> {noformat}
> Error Message
> query_test/test_observability.py:415: in test_global_exchange_counters assert 
> m E assert None
> Stacktrace
> query_test/test_observability.py:415: in test_global_exchange_counters assert 
> m E assert None
> Standard Error
> -- executing against localhost:21000 select count(*) from tpch_parquet.orders 
> o inner join tpch_parquet.lineitem l on o.o_orderkey = l.l_orderkey group by 
> o.o_clerk limit 10; -- 2019-05-28 05:24:17,072 INFO MainThread: Started query 
> 664116fde66bdd8c:4ca951da
> {noformat}
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8158) Bump Impyla version, use HS2 service to retrieve thrift profiles

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8158.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Bump Impyla version, use HS2 service to retrieve thrift profiles
> 
>
> Key: IMPALA-8158
> URL: https://issues.apache.org/jira/browse/IMPALA-8158
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Infrastructure
>Affects Versions: Impala 3.2.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.3.0
>
>
> Once Impyla has been updated, we should retrieve Thrift profiles through HS2 
> synchronously instead of scraping the debug web pages.
> https://github.com/cloudera/impyla/issues/332



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8461) Re-schedule queries if the executor configuration has changed while queued in AC

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8461.
-
   Resolution: Duplicate
Fix Version/s: Impala 3.3.0

This has been folded into IMPALA-8484

> Re-schedule queries if the executor configuration has changed while queued in 
> AC
> 
>
> Key: IMPALA-8461
> URL: https://issues.apache.org/jira/browse/IMPALA-8461
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.3.0
>
>
> If the executor configuration changes while a query is waiting to be 
> admitted, we need to reschedule it. The current behavior tries to run it as 
> is which will then fail. To achieve this, we should call 
> Scheduler::Schedule() from the AdmissionController and then re-schedule if 
> necessary. We need to think about ways to detect changes to the executor 
> configuration, but a simple hash might be good enough.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8484) Add support to run queries on disjoint executor groups

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8484.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Add support to run queries on disjoint executor groups
> --
>
> Key: IMPALA-8484
> URL: https://issues.apache.org/jira/browse/IMPALA-8484
> Project: IMPALA
>  Issue Type: New Feature
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: scalability
> Fix For: Impala 3.3.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8776) test_event_processing.TestEventProcessing.test_insert_events is flaky

2019-07-21 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8776?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8776.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> test_event_processing.TestEventProcessing.test_insert_events is flaky
> -
>
> Key: IMPALA-8776
> URL: https://issues.apache.org/jira/browse/IMPALA-8776
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: flaky, flaky-test
> Fix For: Impala 3.3.0
>
>
> It looks like test_event_processing.TestEventProcessing.test_insert_events 
> can sporadically fail when waiting for insert events to be processed. To make 
> it more robust, we should wait for longer than 10 seconds. We should also add 
> a small delay when looping inside {{wait_for_insert_event_processing()}} to 
> keep system load low and reduce the risk of starving the other processes of 
> resources.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-8776) test_event_processing.TestEventProcessing.test_insert_events is flaky

2019-07-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8776?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker reassigned IMPALA-8776:
---

Assignee: Lars Volker

> test_event_processing.TestEventProcessing.test_insert_events is flaky
> -
>
> Key: IMPALA-8776
> URL: https://issues.apache.org/jira/browse/IMPALA-8776
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: flaky, flaky-test
>
> It looks like test_event_processing.TestEventProcessing.test_insert_events 
> can sporadically fail when waiting for insert events to be processed. To make 
> it more robust, we should wait for longer than 10 seconds. We should also add 
> a small delay when looping inside {{wait_for_insert_event_processing()}} to 
> keep system load low and reduce the risk of starving the other processes of 
> resources.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8776) test_event_processing.TestEventProcessing.test_insert_events is flaky

2019-07-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8776?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8776:

Labels: flaky flaky-test  (was: )

> test_event_processing.TestEventProcessing.test_insert_events is flaky
> -
>
> Key: IMPALA-8776
> URL: https://issues.apache.org/jira/browse/IMPALA-8776
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Reporter: Lars Volker
>Priority: Major
>  Labels: flaky, flaky-test
>
> It looks like test_event_processing.TestEventProcessing.test_insert_events 
> can sporadically fail when waiting for insert events to be processed. To make 
> it more robust, we should wait for longer than 10 seconds. We should also add 
> a small delay when looping inside {{wait_for_insert_event_processing()}} to 
> keep system load low and reduce the risk of starving the other processes of 
> resources.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8776) test_event_processing.TestEventProcessing.test_insert_events is flaky

2019-07-20 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8776:
---

 Summary: 
test_event_processing.TestEventProcessing.test_insert_events is flaky
 Key: IMPALA-8776
 URL: https://issues.apache.org/jira/browse/IMPALA-8776
 Project: IMPALA
  Issue Type: Improvement
  Components: Infrastructure
Reporter: Lars Volker


It looks like test_event_processing.TestEventProcessing.test_insert_events can 
sporadically fail when waiting for insert events to be processed. To make it 
more robust, we should wait for longer than 10 seconds. We should also add a 
small delay when looping inside {{wait_for_insert_event_processing()}} to keep 
system load low and reduce the risk of starving the other processes of 
resources.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8758) Misleading error message "Unknown executor group" during cluster startup with dedicated coordinator

2019-07-16 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8758?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8758.
-
   Resolution: Fixed
 Assignee: Lars Volker
Fix Version/s: Impala 3.3.0

> Misleading error message "Unknown executor group" during cluster startup with 
> dedicated coordinator
> ---
>
> Key: IMPALA-8758
> URL: https://issues.apache.org/jira/browse/IMPALA-8758
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
>  Labels: cluster-membership, dedicated-coordinator, ramp-up, 
> scheduler
> Fix For: Impala 3.3.0
>
>
> During cluster startup the Scheduler will return an error until the local 
> backend has started up ("Local backend has not been registered in the cluster 
> membership"). Afterwards, it will assume that the default executor group 
> exists. However, if the coordinator is not also an executor (i.e. it is a 
> dedicated coordinator), then it will not actually create the default executor 
> group in cluster-membership-mgr.cc:256.
> Queries are expected to fail in this scenario, but the error message should 
> certainly be improved to indicate that no executors could be found. For this 
> purpose, we should make sure that the default executor group gets created as 
> soon as the local backend has started, but keep it empty if it is not an 
> executor. Then we can warn in the scheduler that no executors have been 
> registered so far.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Closed] (IMPALA-8724) Don't run queries on unhealthy executor groups

2019-07-14 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8724?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker closed IMPALA-8724.
---
Resolution: Duplicate

This will be folded into IMPALA-8484

> Don't run queries on unhealthy executor groups
> --
>
> Key: IMPALA-8724
> URL: https://issues.apache.org/jira/browse/IMPALA-8724
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
>  Labels: admission-control, fault-tolerance, scalability, 
> scheduler
>
> After IMPALA-8484 we need to add a way to exclude executor groups that are 
> only partially available. This will help to keep queries from running on 
> partially started groups and in cases where some nodes of an executor group 
> have failed.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8762) Track number of running queries on all backends in admission controller

2019-07-14 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8762:
---

 Summary: Track number of running queries on all backends in 
admission controller
 Key: IMPALA-8762
 URL: https://issues.apache.org/jira/browse/IMPALA-8762
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


To support running multiple coordinators with executor groups and slot based 
admission checks, all executors need to include the number of currently running 
queries in their statestore updates, similar to mem reserved.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8758) Misleading error message "Unknown executor group" during cluster startup with dedicated coordinator

2019-07-12 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8758:
---

 Summary: Misleading error message "Unknown executor group" during 
cluster startup with dedicated coordinator
 Key: IMPALA-8758
 URL: https://issues.apache.org/jira/browse/IMPALA-8758
 Project: IMPALA
  Issue Type: Bug
  Components: Backend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


During cluster startup the Scheduler will return an error until the local 
backend has started up ("Local backend has not been registered in the cluster 
membership"). Afterwards, it will assume that the default executor group 
exists. However, if the coordinator is not also an executor (i.e. it is a 
dedicated coordinator), then it will not actually create the default executor 
group in cluster-membership-mgr.cc:256.

Queries are expected to fail in this scenario, but the error message should 
certainly be improved to indicate that no executors could be found. For this 
purpose, we should make sure that the default executor group gets created as 
soon as the local backend has started, but keep it empty if it is not an 
executor. Then we can warn in the scheduler that no executors have been 
registered so far.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8757) Extend slot based admission to default executor group

2019-07-12 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8757:
---

 Summary: Extend slot based admission to default executor group
 Key: IMPALA-8757
 URL: https://issues.apache.org/jira/browse/IMPALA-8757
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


IMPALA-8484 adds support for multiple executor groups and uses a slot-based 
mechanism to admit queries to executors. In order to keep the existing behavior 
stable, this logic is not applied to the default executor group.

This Jira tracks work on doing that. We have to be careful not to break the 
existing behavior, and if we do, hold the change back until the next 
compatibility-breaking release.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8731) Balance queries between executor groups

2019-06-28 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8731:
---

 Summary: Balance queries between executor groups
 Key: IMPALA-8731
 URL: https://issues.apache.org/jira/browse/IMPALA-8731
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


After IMPALA-8484, we should revisit the assignment policy that we use to 
distribute queries to executor groups. In particular we should implement a 
policy that balances queries across executor groups instead of filling them up 
one by one.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8724) Don't run queries on unhealthy executor groups

2019-06-27 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8724:
---

 Summary: Don't run queries on unhealthy executor groups
 Key: IMPALA-8724
 URL: https://issues.apache.org/jira/browse/IMPALA-8724
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Lars Volker


After IMPALA-8484 we need to add a way to exclude executor groups that are only 
partially available. This will help to keep queries from running on partially 
started groups and in cases where some nodes of an executor group have failed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-2968) Improve admission control dequeuing policy

2019-06-25 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-2968?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16872494#comment-16872494
 ] 

Lars Volker commented on IMPALA-2968:
-

One easy way that I can think of to do this is to add a "attempts before 
blocking" counter to each queue node. Every time a node fails to get admitted, 
we increment the counter by 1, and when hitting a configurable limit (e.g. 5) 
we block.

[~bikramjeet.vig], [~tarmstrong], [~asherman] - Thoughts?

> Improve admission control dequeuing policy
> --
>
> Key: IMPALA-2968
> URL: https://issues.apache.org/jira/browse/IMPALA-2968
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 2.3.0
>Reporter: Matthew Jacobs
>Priority: Minor
>  Labels: admission-control, resource-management
>
> The current behavior only attempts to admit the head of the queue but the 
> head might need resources which are contended (e.g. on a hot node) while a 
> queued request behind the head might be admitted. We should consider a policy 
> which would not block the entire queue but yet is unlikely to starve the head 
> if other requests are continuously admitted from behind it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8700) test_hdfs_scan_node_errors DCHECK-failing on invalid mtime

2019-06-24 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8700:

Labels: broken-build  (was: )

> test_hdfs_scan_node_errors DCHECK-failing on invalid mtime
> --
>
> Key: IMPALA-8700
> URL: https://issues.apache.org/jira/browse/IMPALA-8700
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Jim Apple
>Assignee: Joe McDonnell
>Priority: Critical
>  Labels: broken-build
>
> The test 
> data_errors/test_data_errors.py::TestHdfsScanNodeErrors::test_hdfs_scan_node_errors
> is failing with
> {{F0623 scan-range.cc:480 Check failed: mtime > 0 (-2114487582 vs. 0)}}
> I've been able to reproduce it several times. Here is a saved nightly job on 
> master with the issue:
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6403/
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6403/artifact/Impala/logs_static/logs/ee_tests/impalad_node2.FATAL/*view*/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8685) Evaluate default configuration of NUM_REMOTE_EXECUTOR_CANDIDATES

2019-06-19 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8685?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16868142#comment-16868142
 ] 

Lars Volker commented on IMPALA-8685:
-

Another reason of considering more than one remote executor per scan range 
during scheduling is to prevent skew. The scheduler tries to balance the amount 
of data processes by each executor and having a single choice (1 candidate) can 
prevent it from doing so.

> Evaluate default configuration of NUM_REMOTE_EXECUTOR_CANDIDATES
> 
>
> Key: IMPALA-8685
> URL: https://issues.apache.org/jira/browse/IMPALA-8685
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Michael Ho
>Priority: Critical
>
> The query option {{NUM_REMOTE_EXECUTOR_CANDIDATES}} is set to 3 by default. 
> This means that there are potentially 3 different executors which can process 
> a remote scan range. Over time, the data of a given remote scan range will be 
> spread across these 3 executors. My understanding of why this is not set to 1 
> is to avoid hot spots in pathological cases. On the other hand, this may mean 
> that we may not maximize the utilization of the file handle cache and data 
> cache. Also, for small clusters (e.g. a 3 node cluster), the default value 
> may render deterministic remote scan range scheduling ineffective. We may 
> want to re-evaluate the default value of {{NUM_REMOTE_EXECUTOR_CANDIDATES}}. 
> One idea is to set it to min(3, half of cluster size) so it works okay with 
> small cluster, which may be rather common for demo purposes. There may also 
> be other criteria for evaluating the default value.
> cc'ing [~joemcdonnell], [~tlipcon] and [~drorke]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8673) Add query option to force plan hints for insert queries

2019-06-18 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8673:
---

 Summary: Add query option to force plan hints for insert queries
 Key: IMPALA-8673
 URL: https://issues.apache.org/jira/browse/IMPALA-8673
 Project: IMPALA
  Issue Type: Improvement
  Components: Backend, Frontend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker


In Impala 3.0 we turned on insert clustering by default (IMPALA-5293). This can 
lead to performance regressions from 2.x for highly skewed data sets. To help 
with those cases, we should add a way to force plan hints for insert queries 
through a query option.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-06-12 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Description: 
Add configuration parameters to Admission Controller that scale with
the number of hosts in the resource pool. The planned parameters are

* Max Running Queries Multiple - the Multiple of the current total number of 
running Impalads which gives the maximum number of concurrently running queries 
in this pool. 
* Max Queued Queries Multiple - the Multiple of the current total number of 
running Impalads which gives the maximum number of queries that can be queued 
in this pool. 
* Max Memory Multiple - the Multiple of the current total number of running 
Impalads which gives the maximum memory available across the cluster for the 
pool.


  was:
Add configuration parameters to Admission Controller that scale with
the number of hosts in the resource pool. The planned parameters are Max 
Running Queries * * Multiple - the Multiple of the current total number of 
running Impalads which gives the maximum number of concurrently running queries 
in this pool. 
* Max Queued Queries Multiple - the Multiple of the current total number of 
running Impalads which gives the maximum number of queries that can be queued 
in this pool. 
* Max Memory Multiple - the Multiple of the current total number of running 
Impalads which gives the maximum memory available across the cluster for the 
pool.



> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>  Labels: admission-control, resource-management, scalability
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are
> * Max Running Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8662) Timeout in test test_owner_privileges_without_grant

2019-06-12 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8662:
---

 Summary: Timeout in test test_owner_privileges_without_grant
 Key: IMPALA-8662
 URL: https://issues.apache.org/jira/browse/IMPALA-8662
 Project: IMPALA
  Issue Type: Bug
  Components: Frontend
Affects Versions: Impala 3.3.0
Reporter: Lars Volker
Assignee: Fredy Wijaya


test_owner_privileges_without_grant seems to time out:L

{noformat}
11:39:25 
authorization/test_owner_privileges.py::TestOwnerPrivileges::test_owner_privileges_disabled[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
text/none] PASSED
11:39:45 
authorization/test_owner_privileges.py::TestOwnerPrivileges::test_owner_privileges_without_grant[protocol:
 beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
text/none] 
11:39:45 
11:39:45  Tests TIMED OUT! 
11:39:45 
11:39:45 
11:39:45  Generating stacktrace of impalad with process id: 843 
11:40:09 warning: File 
"/data/jenkins/workspace/impala-asf-master-exhaustive-centos6/Impala-Toolchain/gcc-4.9.2/lib64/libstdc++.so.6.0.20-gdb.py"
 auto-loading has been declined by your `auto-load safe-path' set to 
"$debugdir:$datadir/auto-load".
11:40:27 Couldn't write debug register: No such process.
11:40:28  Generating stacktrace of impalad with process id: 847 
11:40:28 ptrace: No such process.
11:40:28  Generating stacktrace of impalad with process id: 855 
11:40:28 ptrace: No such process.
11:40:28 
/data/jenkins/workspace/impala-asf-master-exhaustive-centos6/repos/Impala/buildall.sh:
 line 530: 12310 Terminated  "${IMPALA_HOME}/bin/run-all-tests.sh" 
-e $EXPLORATION_STRATEGY $RUN_ALL_TESTS_ARGS
11:40:28 + echo 'buildall.sh ' -format '-testdata failed.'
11:40:28 buildall.sh  -format -testdata failed.
{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-8554) The timeout in run-all-tests-timeout-check.sh is triggered in exhaustive test runs on CentOS 6

2019-06-10 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker reassigned IMPALA-8554:
---

Assignee: Laszlo Gaal

> The timeout in run-all-tests-timeout-check.sh is triggered in exhaustive test 
> runs on CentOS 6
> --
>
> Key: IMPALA-8554
> URL: https://issues.apache.org/jira/browse/IMPALA-8554
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.3.0
> Environment: CentOS 6, exhaustive mode
>Reporter: Laszlo Gaal
>Assignee: Laszlo Gaal
>Priority: Blocker
>  Labels: broken-build
>
> Exhaustive test runs on CentOS 6 can take around 20-22 hours even in the 
> normal case.
> run-all-tests-timeout-check.sh installs a timeout guard of 72000 s, i.e. 20 
> hours, which includes only the tests, but excludes build and dataload time.
> On recent runs some of the spilling tests ran slowly but successfully, which 
> made the timeout fire during the last items of custom_cluster_tests, cutting 
> the test run short.
> I think the timeout should be made longer for CentOS 6 based tests until we 
> find ways to run those faster.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker reassigned IMPALA-7154:
---

Assignee: Vihang Karajgaonkar  (was: Gabor Kaszab)

> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Vihang Karajgaonkar
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropDatabase(HiveMetaStoreClient.java:791)
> at sun.reflect.GeneratedMethodAccessor26.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:101)
> at com.sun.proxy.$Proxy5.dropDatabase(Unknown Source)
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1305)
> ... 2 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16860538#comment-16860538
 ] 

Lars Volker commented on IMPALA-7154:
-

This has become more frequent, making it a P1. [~vihangk1] - I’m assigning this 
to you thinking you might have an idea what’s going on here; feel free to find 
another person or assign back to me if you're swamped.


> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Gabor Kaszab
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropDatabase(HiveMetaStoreClient.java:791)
> at sun.reflect.GeneratedMethodAccessor26.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:101)
> at com.sun.proxy.$Proxy5.dropDatabase(Unknown Source)
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1305)
> ... 2 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8554) The timeout in run-all-tests-timeout-check.sh is triggered in exhaustive test runs on CentOS 6

2019-06-10 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16860523#comment-16860523
 ] 

Lars Volker commented on IMPALA-8554:
-

[~laszlog] - Do you have time to make this change? If not I'll try to find 
someone else.

> The timeout in run-all-tests-timeout-check.sh is triggered in exhaustive test 
> runs on CentOS 6
> --
>
> Key: IMPALA-8554
> URL: https://issues.apache.org/jira/browse/IMPALA-8554
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.3.0
> Environment: CentOS 6, exhaustive mode
>Reporter: Laszlo Gaal
>Priority: Blocker
>  Labels: broken-build
>
> Exhaustive test runs on CentOS 6 can take around 20-22 hours even in the 
> normal case.
> run-all-tests-timeout-check.sh installs a timeout guard of 72000 s, i.e. 20 
> hours, which includes only the tests, but excludes build and dataload time.
> On recent runs some of the spilling tests ran slowly but successfully, which 
> made the timeout fire during the last items of custom_cluster_tests, cutting 
> the test run short.
> I think the timeout should be made longer for CentOS 6 based tests until we 
> find ways to run those faster.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-7154:

Priority: Blocker  (was: Critical)

> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Gabor Kaszab
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropDatabase(HiveMetaStoreClient.java:791)
> at sun.reflect.GeneratedMethodAccessor26.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:101)
> at com.sun.proxy.$Proxy5.dropDatabase(Unknown Source)
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1305)
> ... 2 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker reassigned IMPALA-7154:
---

Assignee: Gabor Kaszab  (was: Bharathkrishna Guruvayoor Murali)

> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0
>Reporter: Tim Armstrong
>Assignee: Gabor Kaszab
>Priority: Critical
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropDatabase(HiveMetaStoreClient.java:791)
> at sun.reflect.GeneratedMethodAccessor26.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:101)
> at com.sun.proxy.$Proxy5.dropDatabase(Unknown Source)
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1305)
> ... 2 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-7154:

Affects Version/s: Impala 3.3.0

> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0, Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Gabor Kaszab
>Priority: Critical
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropDatabase(HiveMetaStoreClient.java:791)
> at sun.reflect.GeneratedMethodAccessor26.invoke(Unknown Source)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:101)
> at com.sun.proxy.$Proxy5.dropDatabase(Unknown Source)
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1305)
> ... 2 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7154) Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException thrown

2019-06-10 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7154?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16860513#comment-16860513
 ] 

Lars Volker commented on IMPALA-7154:
-

Another occurrence:

{noformat}
conftest.py:319: in cleanup
{'sync_ddl': sync_ddl})
common/impala_test_suite.py:691: in wrapper
return function(*args, **kwargs)
common/impala_test_suite.py:699: in execute_query_expect_success
result = cls.__execute_query(impalad_client, query, query_options, user)
common/impala_test_suite.py:793: in __execute_query
return impalad_client.execute(query, user=user)
common/impala_connection.py:180: in execute
return self.__beeswax_client.execute(sql_stmt, user=user)
beeswax/impala_beeswax.py:187: in execute
handle = self.__execute_query(query_string.strip(), user=user)
beeswax/impala_beeswax.py:362: in __execute_query
handle = self.execute_query_async(query_string, user=user)
beeswax/impala_beeswax.py:356: in execute_query_async
handle = self.__do_rpc(lambda: self.imp_service.query(query,))
beeswax/impala_beeswax.py:516: in __do_rpc
raise ImpalaBeeswaxException(self.__build_error_message(b), b)
E   ImpalaBeeswaxException: ImpalaBeeswaxException:
EINNER EXCEPTION: 
EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
Metastore: 
E   CAUSED BY: NoSuchObjectException: test_fuzz_uncompressed_parquet_e0e417c9
{noformat}

> Error making 'dropDatabase' RPC to Hive Metastore. NoSuchObjectException 
> thrown
> ---
>
> Key: IMPALA-7154
> URL: https://issues.apache.org/jira/browse/IMPALA-7154
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Affects Versions: Impala 2.13.0
>Reporter: Tim Armstrong
>Assignee: Bharathkrishna Guruvayoor Murali
>Priority: Critical
>  Labels: broken-build, flaky
> Attachments: TEST-impala-parallel.log.gz, 
> TEST-impala-parallel.xml.gz, 
> catalogd.ec2-m2-4xlarge-centos-6-4-0f46.vpc.cloudera.com.jenkins.log.INFO.20180608-024815.32143.gz,
>  hive.log.gz
>
>
> {noformat}
> conftest.py:293: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:528: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:535: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options)
> common/impala_test_suite.py:620: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:160: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:173: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:339: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:335: in execute_query_async
> return self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:460: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_resolution_by_name_56b45511
> {noformat}
> The backtrace in the catalogd log is:
> {noformat}
> I0608 05:49:26.111824 24195 jni-util.cc:230] 
> org.apache.impala.common.ImpalaRuntimeException: Error making 'dropDatabase' 
> RPC to Hive Metastore: 
> at 
> org.apache.impala.service.CatalogOpExecutor.dropDatabase(CatalogOpExecutor.java:1309)
> at 
> org.apache.impala.service.CatalogOpExecutor.execDdlRequest(CatalogOpExecutor.java:300)
> at org.apache.impala.service.JniCatalog.execDdl(JniCatalog.java:146)
> Caused by: NoSuchObjectException(message:test_resolution_by_name_56b45511)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16387)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result$get_database_resultStandardScheme.read(ThriftHiveMetastore.java:16364)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_database_result.read(ThriftHiveMetastore.java:16295)
> at 
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:702)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:689)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1232)
> at 
> 

[jira] [Updated] (IMPALA-8638) Test failure: test_create_table_timestamp

2019-06-07 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8638?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8638:

Labels: broken-build  (was: )

> Test failure: test_create_table_timestamp
> -
>
> Key: IMPALA-8638
> URL: https://issues.apache.org/jira/browse/IMPALA-8638
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.3.0
>Reporter: Csaba Ringhofer
>Assignee: Fredy Wijaya
>Priority: Critical
>  Labels: broken-build
>
> I saw two recent failures of this test, on without proper explanation for the 
> error, and one with the following message:
> 04:41:33.324 FAIL 
> custom_cluster/test_lineage.py::TestLineage::()::test_create_table_timestamp[protocol:
>  beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none]
> 04:41:33.324 === FAILURES 
> ===
> 04:41:33.341  TestLineage.test_create_table_timestamp[protocol: beeswax | 
> exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none] 
> 04:41:33.341 custom_cluster/test_lineage.py:98: in test_create_table_timestamp
> 04:41:33.342 assert lineage_json["queryId"] == profile_query_id
> 04:41:33.342 E   assert '75440fbf4d05...d1bd9' == 
> 'fd4d30a1d0696...2b93c'
> 04:41:33.342 E - 75440fbf4d05a7a7:832d1bd9
> 04:41:33.342 E + fd4d30a1d06964f4:99e2b93c



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8638) Test failure: test_create_table_timestamp

2019-06-07 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8638?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8638:

Priority: Blocker  (was: Critical)

> Test failure: test_create_table_timestamp
> -
>
> Key: IMPALA-8638
> URL: https://issues.apache.org/jira/browse/IMPALA-8638
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.3.0
>Reporter: Csaba Ringhofer
>Assignee: Fredy Wijaya
>Priority: Blocker
>  Labels: broken-build
>
> I saw two recent failures of this test, on without proper explanation for the 
> error, and one with the following message:
> 04:41:33.324 FAIL 
> custom_cluster/test_lineage.py::TestLineage::()::test_create_table_timestamp[protocol:
>  beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none]
> 04:41:33.324 === FAILURES 
> ===
> 04:41:33.341  TestLineage.test_create_table_timestamp[protocol: beeswax | 
> exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none] 
> 04:41:33.341 custom_cluster/test_lineage.py:98: in test_create_table_timestamp
> 04:41:33.342 assert lineage_json["queryId"] == profile_query_id
> 04:41:33.342 E   assert '75440fbf4d05...d1bd9' == 
> 'fd4d30a1d0696...2b93c'
> 04:41:33.342 E - 75440fbf4d05a7a7:832d1bd9
> 04:41:33.342 E + fd4d30a1d06964f4:99e2b93c



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-4047) Remove "CDH" from Impala

2019-06-04 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-4047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-4047.
-
Resolution: Fixed

> Remove "CDH" from Impala
> 
>
> Key: IMPALA-4047
> URL: https://issues.apache.org/jira/browse/IMPALA-4047
> Project: IMPALA
>  Issue Type: Task
>  Components: Infrastructure
>Affects Versions: Impala 2.7.0
>Reporter: Jim Apple
>Assignee: Lars Volker
>Priority: Major
>  Labels: asf
>
> Impala is an Apache (incubating) project, and "CDH"/"cdh" should be removed 
> from shell scripts and config files and so on.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-7117) Lower debug level for HDFS S3 connector back to INFO

2019-06-04 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-7117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16855849#comment-16855849
 ] 

Lars Volker commented on IMPALA-7117:
-

[[~joemcdonnell], [~stakiar] - Looks like this one fell through the cracks more 
than once but the log level increase only had an effect during tests. Should we 
keep it at "debug"?

> Lower debug level for HDFS S3 connector back to INFO
> 
>
> Key: IMPALA-7117
> URL: https://issues.apache.org/jira/browse/IMPALA-7117
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 2.13.0, Impala 3.1.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Blocker
>  Labels: s3
>
> This change will increase the log level for the HDFS S3 connector to DEBUG to 
> help with IMPALA-6910 and IMPALA-7070. Before the next release we need to 
> lower it again.
> https://gerrit.cloudera.org/#/c/10596/
> I'm making this a P1 to remind us that we must do this before cutting a 
> release.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-6951) java7 vs. java8 warning prevents installation of impala-python on test machines

2019-06-04 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-6951.
-
Resolution: Fixed

> java7 vs. java8 warning prevents installation of impala-python on test 
> machines
> ---
>
> Key: IMPALA-6951
> URL: https://issues.apache.org/jira/browse/IMPALA-6951
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Michael Brown
>Assignee: Lars Volker
>Priority: Major
>
> Commit {{22c7ded07eb2710aba3e1aa07ed7ec1a448f7c61}} now prevents the 
> installation of impala-python on a test machine wanting to run tests against 
> an Impala cluster. It was very convenient to be able to install impala-python 
> and run tests like the stress test regardless of the Java version on the host.
> In order to install impala-python, one must source impala-config.sh. When the 
> failpoint is there, then impala-python now MUST have Java 8. This is a burden 
> to update.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-5967) Impala process hangs in breakpad exception handler

2019-06-04 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-5967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16855785#comment-16855785
 ] 

Lars Volker commented on IMPALA-5967:
-

[~tarmstrong] - A lot has changed in the past 18 months. Do you think this is 
still worth investigating?

> Impala process hangs in breakpad exception handler
> --
>
> Key: IMPALA-5967
> URL: https://issues.apache.org/jira/browse/IMPALA-5967
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 2.11.0
>Reporter: Tim Armstrong
>Assignee: Lars Volker
>Priority: Major
> Attachments: gdb.txt
>
>
> I managed to reproduce this issue by accident locally. I believe we've seen 
> it before but I'm not sure if we have an open JIRA for it. 
> With this branch I was able to trigger the bug:
> https://github.com/timarmstrong/incubator-impala/tree/buggy-avro-trigger-breakpad-hang
> By running this from impala-shell.sh
> {code}
> set num_nodes=1;
> set disable_codegen=0;
> set disable_codegen_rows_threshold=0;
> use functional_avro;
> select count(*),
>   sum(id)
> from alltypesagg
> where id % 2 = 0 and day is not null;
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8460) Unify cluster membership management in the backend

2019-06-03 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16855051#comment-16855051
 ] 

Lars Volker commented on IMPALA-8460:
-

[~arodoni_cloudera] - Thanks for checking. Nothing changes from a user's 
perspective, this was purely a refactoring of existing behavior.

> Unify cluster membership management in the backend
> --
>
> Key: IMPALA-8460
> URL: https://issues.apache.org/jira/browse/IMPALA-8460
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.3.0
>
>
> We currently subscribe to the cluster-membership statestore topic in both the 
> ImpalaServer and the Scheduler. Updates are then also propagated to the 
> Frontend. Additionally the Scheduler has its own representation of the local 
> backend descriptor, which is not kept in sync with the one in the Impala 
> server.
> To make cluster-membership and the interactions with the statestore easier to 
> reason about, we should pull all membership handling in to a single place.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8460) Unify cluster membership management in the backend

2019-06-03 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker resolved IMPALA-8460.
-
   Resolution: Fixed
Fix Version/s: Impala 3.3.0

> Unify cluster membership management in the backend
> --
>
> Key: IMPALA-8460
> URL: https://issues.apache.org/jira/browse/IMPALA-8460
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Major
> Fix For: Impala 3.3.0
>
>
> We currently subscribe to the cluster-membership statestore topic in both the 
> ImpalaServer and the Scheduler. Updates are then also propagated to the 
> Frontend. Additionally the Scheduler has its own representation of the local 
> backend descriptor, which is not kept in sync with the one in the Impala 
> server.
> To make cluster-membership and the interactions with the statestore easier to 
> reason about, we should pull all membership handling in to a single place.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8567) Many random catalog consistency issues with catalog v2/event processor

2019-06-03 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8567?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16854747#comment-16854747
 ] 

Lars Volker commented on IMPALA-8567:
-

You can access the Jenkins logs and a zip archive with all artifacts from the 
links that Tim and I posted throughout this Jira. E.g. for the latest one that 
Tim posted, you can download the archive from 
[here|https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/408/]. I 
marked some of the builds as "Keep forever" but some of them might rotate out 
(doesn't seem to hard to repro though).

> Many random catalog consistency issues with catalog v2/event processor
> --
>
> Key: IMPALA-8567
> URL: https://issues.apache.org/jira/browse/IMPALA-8567
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Vihang Karajgaonkar
>Priority: Blocker
>  Labels: broken-build, catalog, flaky
>
> [~tlipcon] [~vihangk1] FYI. I'm not sure whether the local catalog or the 
> event processor is likely to blame here so I'll let you look. The general 
> theme is tables and databases not existing when they should.
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/289/testReport/junit/metadata.test_refresh_partition/TestRefreshPartition/test_drop_hive_partition_and_refresh_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/267/testReport/junit/query_test.test_kudu/TestKuduOperations/test_kudu_insert_protocol__beeswax___exec_optionkudu_read_modeREAD_AT_SNAPSHOTbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_metadata_query_statements/TestMetadataQueryStatements/test_describe_db_protocol__beeswax___exec_optionsync_ddl___0___batch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_hms_integration/TestHmsIntegrationSanity/test_sanity_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/288/testReport/junit/query_test.test_insert_parquet/TestHdfsParquetTableStatsWriter/test_write_statistics_multiple_row_groups_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__parquet_none_/
> I'll include the output of each job in a follow-on comment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8567) Many random catalog consistency issues with catalog v2/event processor

2019-06-01 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8567?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16853819#comment-16853819
 ] 

Lars Volker commented on IMPALA-8567:
-

Is there a change we can back out to fix this temporarily? Should we consider 
disabling some tests in order to reduce the impact on our development workflows?

> Many random catalog consistency issues with catalog v2/event processor
> --
>
> Key: IMPALA-8567
> URL: https://issues.apache.org/jira/browse/IMPALA-8567
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Vihang Karajgaonkar
>Priority: Blocker
>  Labels: broken-build, catalog, flaky
>
> [~tlipcon] [~vihangk1] FYI. I'm not sure whether the local catalog or the 
> event processor is likely to blame here so I'll let you look. The general 
> theme is tables and databases not existing when they should.
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/289/testReport/junit/metadata.test_refresh_partition/TestRefreshPartition/test_drop_hive_partition_and_refresh_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/267/testReport/junit/query_test.test_kudu/TestKuduOperations/test_kudu_insert_protocol__beeswax___exec_optionkudu_read_modeREAD_AT_SNAPSHOTbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_metadata_query_statements/TestMetadataQueryStatements/test_describe_db_protocol__beeswax___exec_optionsync_ddl___0___batch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_hms_integration/TestHmsIntegrationSanity/test_sanity_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/288/testReport/junit/query_test.test_insert_parquet/TestHdfsParquetTableStatsWriter/test_write_statistics_multiple_row_groups_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__parquet_none_/
> I'll include the output of each job in a follow-on comment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8610) Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs

2019-06-01 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8610?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16853776#comment-16853776
 ] 

Lars Volker commented on IMPALA-8610:
-

I added a step to the clean up actions that kills any remaining daemons:

{code}
function onexit {
  df -m
  free -m
  uptime -p
  rm -rf "${IMPALA_HOME}"/logs_static
  mkdir -p "${IMPALA_HOME}"/logs_static
  killall impalad catalogd statestored || true   # <-- added this line
  cp -r -L "${IMPALA_HOME}"/logs "${IMPALA_HOME}"/logs_static
  rm -f "${GIT_STDERR}"
}
{code}

> Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs
> -
>
> Key: IMPALA-8610
> URL: https://issues.apache.org/jira/browse/IMPALA-8610
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure, jenkins
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
>
> It seems that otherwise successful runs of {{ubuntu-16.04-from-scratch}} can 
> fail at the end when copying log files in the end. If the impalads are still 
> running, they can rotate a logfile while {{cp}} tries to copy it, which 
> results in output like this ([affected 
> job|https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/console]):
> {noformat}
> 00:24:28 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102026':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072304.100810':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102023':
>  No such file or directory
> {noformat}
> Note that the script is currently configured in Jenkins to run in a shell 
> with {{-e}} and that a successful run contains additional lines after the 
> {{cp}} command:
> {noformat}
> 04:16:18 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 04:16:21 + rm -f /tmp/git-err-ksiZp.log
> {noformat}
> As a fix we should kill all daemons before copying the files.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8610) Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs

2019-06-01 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8610?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16853775#comment-16853775
 ] 

Lars Volker commented on IMPALA-8610:
-

[~tarmstr...@cloudera.com] - It seems your job here hit this as well: 
https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/

> Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs
> -
>
> Key: IMPALA-8610
> URL: https://issues.apache.org/jira/browse/IMPALA-8610
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure, jenkins
>Reporter: Lars Volker
>Assignee: Lars Volker
>Priority: Critical
>
> It seems that otherwise successful runs of {{ubuntu-16.04-from-scratch}} can 
> fail at the end when copying log files in the end. If the impalads are still 
> running, they can rotate a logfile while {{cp}} tries to copy it, which 
> results in output like this ([affected 
> job|https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/console]):
> {noformat}
> 00:24:28 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102026':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072304.100810':
>  No such file or directory
> 00:24:37 cp: cannot stat 
> '/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102023':
>  No such file or directory
> {noformat}
> Note that the script is currently configured in Jenkins to run in a shell 
> with {{-e}} and that a successful run contains additional lines after the 
> {{cp}} command:
> {noformat}
> 04:16:18 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
> 04:16:21 + rm -f /tmp/git-err-ksiZp.log
> {noformat}
> As a fix we should kill all daemons before copying the files.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8610) Log rotation can fail gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs

2019-06-01 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8610:
---

 Summary: Log rotation can fail 
gerrit-verify-dryrun/ubuntu-16.04-from-scratch jobs
 Key: IMPALA-8610
 URL: https://issues.apache.org/jira/browse/IMPALA-8610
 Project: IMPALA
  Issue Type: Bug
  Components: jenkins, Infrastructure
Reporter: Lars Volker
Assignee: Lars Volker


It seems that otherwise successful runs of {{ubuntu-16.04-from-scratch}} can 
fail at the end when copying log files in the end. If the impalads are still 
running, they can rotate a logfile while {{cp}} tries to copy it, which results 
in output like this ([affected 
job|https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/6027/console]):

{noformat}
00:24:28 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
00:24:37 cp: cannot stat 
'/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102026':
 No such file or directory
00:24:37 cp: cannot stat 
'/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072304.100810':
 No such file or directory
00:24:37 cp: cannot stat 
'/home/ubuntu/Impala/logs/cluster/impalad.ip-172-31-19-200.ubuntu.log.WARNING.20190601-072315.102023':
 No such file or directory
{noformat}

Note that the script is currently configured in Jenkins to run in a shell with 
{{-e}} and that a successful run contains additional lines after the {{cp}} 
command:

{noformat}
04:16:18 + cp -r -L /home/ubuntu/Impala/logs /home/ubuntu/Impala/logs_static
04:16:21 + rm -f /tmp/git-err-ksiZp.log
{noformat}

As a fix we should kill all daemons before copying the files.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8567) Many random catalog consistency issues with catalog v2/event processor

2019-05-31 Thread Lars Volker (JIRA)


[ 
https://issues.apache.org/jira/browse/IMPALA-8567?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16853537#comment-16853537
 ] 

Lars Volker commented on IMPALA-8567:
-

I think I've hit this in 
https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/398/

> Many random catalog consistency issues with catalog v2/event processor
> --
>
> Key: IMPALA-8567
> URL: https://issues.apache.org/jira/browse/IMPALA-8567
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend
>Affects Versions: Impala 3.3.0
>Reporter: Tim Armstrong
>Assignee: Vihang Karajgaonkar
>Priority: Blocker
>  Labels: broken-build, catalog, flaky
>
> [~tlipcon] [~vihangk1] FYI. I'm not sure whether the local catalog or the 
> event processor is likely to blame here so I'll let you look. The general 
> theme is tables and databases not existing when they should.
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/289/testReport/junit/metadata.test_refresh_partition/TestRefreshPartition/test_drop_hive_partition_and_refresh_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/267/testReport/junit/query_test.test_kudu/TestKuduOperations/test_kudu_insert_protocol__beeswax___exec_optionkudu_read_modeREAD_AT_SNAPSHOTbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_metadata_query_statements/TestMetadataQueryStatements/test_describe_db_protocol__beeswax___exec_optionsync_ddl___0___batch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/286/testReport/junit/metadata.test_hms_integration/TestHmsIntegrationSanity/test_sanity_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___5000___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_/
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/288/testReport/junit/query_test.test_insert_parquet/TestHdfsParquetTableStatsWriter/test_write_statistics_multiple_row_groups_protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__parquet_none_/
> I'll include the output of each job in a follow-on comment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-8608) Impala needs to support Python 3

2019-05-31 Thread Lars Volker (JIRA)
Lars Volker created IMPALA-8608:
---

 Summary: Impala needs to support Python 3
 Key: IMPALA-8608
 URL: https://issues.apache.org/jira/browse/IMPALA-8608
 Project: IMPALA
  Issue Type: Bug
  Components: Infrastructure
Reporter: Lars Volker


[The End of Python 2.7|https://pythonclock.org/] support is getting closer and 
as such we need to be able to move to Python 3.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8608) Impala needs to support Python 3

2019-05-31 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8608?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8608:

Affects Version/s: Impala 3.3.0

> Impala needs to support Python 3
> 
>
> Key: IMPALA-8608
> URL: https://issues.apache.org/jira/browse/IMPALA-8608
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Affects Versions: Impala 3.3.0
>Reporter: Lars Volker
>Priority: Critical
>
> [The End of Python 2.7|https://pythonclock.org/] support is getting closer 
> and as such we need to be able to move to Python 3.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-6596) Query failed with OOM on coordinator while remote fragments on other nodes continue to run

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6596?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-6596:

Labels: scalability  (was: )

> Query failed with OOM on coordinator while remote fragments on other nodes 
> continue to run
> --
>
> Key: IMPALA-6596
> URL: https://issues.apache.org/jira/browse/IMPALA-6596
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 2.11.0
>Reporter: Mostafa Mokhtar
>Priority: Critical
>  Labels: scalability
> Attachments: RowSizeFailureProfile.txt
>
>
> This is somewhat similar to IMPALA-2990.
> Query 
> {code:java}
> set NUM_SCANNER_THREADS=2;
> set MAX_ROW_SIZE=4194304;
> with cte as (select group_concat(cast(fnv_hash(concat(o_comment, 'd')) as 
> string)) as c1,group_concat(cast(fnv_hash(concat(o_comment, 'e')) as string)) 
> as c2 from orders where o_orderkey <12 and o_orderdate <"1993-01-01" 
> union all select group_concat(cast(fnv_hash(concat(o_comment, 'd')) as 
> string)) as c1,group_concat(cast(fnv_hash(concat(o_comment, 'e')) as string)) 
> as c2 from orders where o_orderkey <12 and o_orderdate 
> <"1993-01-01"), cte2 as (select c1,c2,s_suppkey from cte , supplier) select 
> count(*) from cte2 t1, cte2 t2 where t1.s_suppkey = t2.s_suppkey group by 
> t1.c1 , t1.c2 , t2.c1 , t2.c2 having count(*) = 1
> {code}
> Failed on coordinator node which is also an executor with 
> {code:java}
> h4. _Status:_ Row of size 1.82 GB could not be materialized in plan node with 
> id 14. Increase the max_row_size query option (currently 4.00 MB) to process 
> larger rows.
> {code}
> Log on the coordinator has lots of entries with 
> {code:java}
> I0227 19:20:58.057637 62974 impala-server.cc:1196] ReportExecStatus(): 
> Received report for unknown query ID (probably closed or cancelled): 
> 9b439fc1ee1addb7:82d41569 I0227 19:20:58.152979 63129 
> impala-server.cc:1196] ReportExecStatus(): Received report for unknown query 
> ID (probably closed or cancelled): 9b439fc1ee1addb7:82d41569 I0227 
> 19:20:58.714336 63930 impala-server.cc:1196] ReportExecStatus(): Received 
> report for unknown query ID (probably closed or cancelled): 
> 9b439fc1ee1addb7:82d41569 I0227 19:20:58.718415 63095 
> impala-server.cc:1196] ReportExecStatus(): Received report for unknown query 
> ID (probably closed or cancelled): 9b439fc1ee1addb7:82d41569 I0227 
> 19:20:58.757306 63339 impala-server.cc:1196] ReportExecStatus(): Received 
> report for unknown query ID (probably closed or cancelled): 
> 9b439fc1ee1addb7:82d41569 I0227 19:20:58.762310 63406 
> impala-server.cc:1196] ReportExecStatus(): Received report for unknown query 
> ID (probably closed or cancelled): 9b439fc1ee1addb7:82d41569
> {code}
> From the memz tab on a different node. 
> h2. Memory Usage
> Memory consumption / limit: *142.87 GB* / *100.00 GB*
> h3. Breakdown
> {code}
> Process: memory limit exceeded. Limit=100.00 GB Total=141.05 GB Peak=144.25 
> GB Buffer Pool: Free Buffers: Total=0 Buffer Pool: Clean Pages: Total=0 
> Buffer Pool: Unused Reservation: Total=-118.00 MB TCMalloc Overhead: 
> Total=1.68 GB RequestPool=root.default: Total=40.10 GB Peak=89.40 GB 
> Query(9b439fc1ee1addb7:82d41569): Reservation=122.00 MB 
> ReservationLimit=80.00 GB OtherMemory=39.98 GB Total=40.10 GB Peak=40.10 GB 
> Unclaimed reservations: Reservation=72.00 MB OtherMemory=0 Total=72.00 MB 
> Peak=226.00 MB Fragment 9b439fc1ee1addb7:82d415690059: Reservation=0 
> OtherMemory=0 Total=0 Peak=632.88 KB AGGREGATION_NODE (id=29): Total=0 
> Peak=76.12 KB EXCHANGE_NODE (id=28): Reservation=0 OtherMemory=0 Total=0 
> Peak=0 DataStreamRecvr: Total=0 Peak=0 DataStreamSender (dst_id=30): Total=0 
> Peak=1.75 KB CodeGen: Total=0 Peak=547.00 KB Fragment 
> 9b439fc1ee1addb7:82d415690050: Reservation=0 OtherMemory=0 Total=0 
> Peak=3.67 GB AGGREGATION_NODE (id=15): Total=0 Peak=76.12 KB HASH_JOIN_NODE 
> (id=14): Reservation=0 OtherMemory=0 Total=0 Peak=1.85 GB Hash Join Builder 
> (join_node_id=14): Total=0 Peak=13.12 KB EXCHANGE_NODE (id=26): Reservation=0 
> OtherMemory=0 Total=0 Peak=1.82 GB DataStreamRecvr: Total=0 Peak=1.82 GB 
> EXCHANGE_NODE (id=27): Reservation=0 OtherMemory=0 Total=0 Peak=1.82 GB 
> DataStreamRecvr: Total=0 Peak=1.82 GB DataStreamSender (dst_id=28): Total=0 
> Peak=15.75 KB CodeGen: Total=0 Peak=1.81 MB Fragment 
> 9b439fc1ee1addb7:82d415690023: Reservation=26.00 MB OtherMemory=19.99 GB 
> Total=20.02 GB Peak=20.02 GB Runtime Filter Bank: Reservation=2.00 MB 
> ReservationLimit=2.00 MB OtherMemory=0 Total=2.00 MB Peak=2.00 MB 
> NESTED_LOOP_JOIN_NODE (id=6): Total=3.63 GB Peak=3.63 GB Nested Loop Join 
> Builder: Total=3.63 GB Peak=3.63 GB 

[jira] [Updated] (IMPALA-6412) Memory issues with processing of incoming global runtime filters on coordinator

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6412?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-6412:

Labels: resource-management scalability  (was: resource-management)

> Memory issues with processing of incoming global runtime filters on 
> coordinator
> ---
>
> Key: IMPALA-6412
> URL: https://issues.apache.org/jira/browse/IMPALA-6412
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend, Distributed Exec
>Affects Versions: Impala 2.11.0
>Reporter: Tim Armstrong
>Priority: Major
>  Labels: resource-management, scalability
>
> This is a tracking JIRA for runtime filter memory issues that occur when 
> processing global runtime filters, which is currently done entirely on the 
> coordinator. I'll link all related issues that I know about and we can close 
> this issue once we're fully satisfied that the problem is resolved.
> Currently the known issues include:
> * All filters are processed on the coordinator, which quickly becomes a 
> bottleneck
> * We don't track the memory consumed by thrift strings in incoming RPCs



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8544) Expose additional S3A / S3Guard metrics

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8544:

Labels: s3  (was: )

> Expose additional S3A / S3Guard metrics
> ---
>
> Key: IMPALA-8544
> URL: https://issues.apache.org/jira/browse/IMPALA-8544
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
>Priority: Major
>  Labels: s3
>
> S3A / S3Guard internally collects several useful metrics that we should 
> consider exposing to Impala users. The full list of statistics can be found 
> in {{o.a.h.fs.s3a.Statistic}}. The stats include: the number of S3 operations 
> performed (put, get, etc.), invocation counts for various {{FileSystem}} 
> methods, stream statistics (bytes read, written, etc.), etc.
> Some interesting stats that stand out:
>  * "stream_aborted": "Count of times the TCP stream was aborted" - the number 
> of TCP connection aborts, a high value would indicate performance issues
>  * "stream_read_exceptions" : "Number of exceptions invoked on input streams" 
> - incremented whenever an {{IOException}} is caught while reading (these 
> exception don't always get propagated to Impala because they trigger a retry)
>  * "store_io_throttled": "Requests throttled and retried" - looks like it 
> tracks the number of times the fs retries an operation because the original 
> request hit a throttling exception
>  * "s3guard_metadatastore_retry": "S3Guard metadata store retry events" - 
> looks like it tracks the number of times the fs retries S3Guard operations
>  * "s3guard_metadatastore_throttled" : "S3Guard metadata store throttled 
> events" - similar to "store_io_throttled" but looks like it is specific to 
> S3Guard
> We should consider how to expose these metrics via Impala logs / runtime 
> profiles.
> There are a few options:
>  * {{S3AFileSystem}} exposes {{StorageStatistics}} specific to S3A / S3Guard 
> via the {{FileSystem#getStorageStatistics}} method; the 
> {{S3AStorageStatistics}} seems to include all the S3A / S3Guard metrics, 
> however, I think the stats might be aggregated globally, which would make it 
> hard to create per-query specific metrics
>  * {{S3AInstrumentation}} exposes all the metrics as well, and looks like it 
> is per-fs instance, so it is not aggregated globally; {{S3AInstrumentation}} 
> extends {{o.a.h.metrics2.MetricsSource}} so perhaps it is exposed via some 
> API (haven't looked into this yet)
>  * {{S3AInputStream#toString}} dumps the statistics from 
> {{o.a.h.fs.s3a.S3AInstrumentation.InputStreamStatistics}} and 
> {{S3AFileSystem#toString}} dumps them all as well
>  * {{S3AFileSystem}} updates the stats in 
> {{o.a.h.fs.Statistics.StatisticsData}} as well (e.g. bytesRead, bytesWritten, 
> etc.)
> Impala has a {{hdfs-fs-cache}} as well, so {{hdfsFs}} objects get shared 
> across threads.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8561) ScanRanges with mtime=-1 can lead to inconsistent reads when using the file handle cache

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8561?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8561:

Description: 
{color:red}colored text{color}The file handle cache relies on the mtime to 
distinguish between different versions of a file. For example, if file X exists 
with mtime=1, then it is overwritten and the metadata is updated so that now it 
is at mtime=2, the file handle cache treats them as completely different things 
and can never use a single file handle to serve both. However, some codepaths 
generate ScanRanges with an mtime of -1. This removes the ability to 
distinguish these two versions of a file and can read to consistency problems.

A specific example is the code that reads the parquet footer 
[HdfsParquetScanner::ProcessFooter()|https://github.com/apache/impala/blob/832c9de7810b47b5f782bccb761e07264e7548e5/be/src/exec/parquet/hdfs-parquet-scanner.cc#L1354].
 We don't know ahead of time how big the Parquet footer is. So, we read 100KB 
(determined by 
[FOOTER_SIZE|https://github.com/apache/impala/blob/449fe73d2145bd22f0f857623c3652a097f06d73/be/src/exec/hdfs-scanner.h#L331]).
 If the footer size encoded in the last few bytes of the file indicates that 
the footer is larger than that [code 
here|https://github.com/apache/impala/blob/832c9de7810b47b5f782bccb761e07264e7548e5/be/src/exec/parquet/hdfs-parquet-scanner.cc#L1414],
 then we issue a separate read for the actual size of the footer. That separate 
read does not inherit the mtime of the original read and instead uses an mtime 
of -1. I verified this by adding tracing and issuing a select against 
functional_parquet.widetable_1000_cols.

A failure scenario associated with this is that we read the last 100KB using a 
ScanRange with mtime=2, then we find that the footer is larger than 100KB and 
issue a ScanRange with mtime=-1. This uses a file handle that is from a 
previous version of the file equivalent to mtime=1. The data it is reading may 
not come from the end of the file, or it may be at the end of the file but the 
footer has a different length. (There is no validation on the new read to check 
the magic value or metadata size reported by the new buffer.) Either would 
result in a failure to deserialize the thrift for the footer. For example, a 
problem case could produce an error message like:

 
{noformat}
File hdfs://test-warehouse/example_file.parq of length 1048576 bytes has 
invalid file metadata at file offset 462017. Error = couldn't deserialize 
thrift msg:
TProtocolException: Invalid data
.{noformat}
To fix this, we should examine all locations that can result in ScanRanges with 
mtime=-1 and eliminate any that we can. For example, the 
HdfsParquetScanner::ProcessFooter() code should create a ScanRange that 
inherits the mtime from the original footer ScanRange. Also, the file handle 
cache should refuse to cache file handles with mtime=-1.

The code in HdfsParquetScanner::ProcessFooter() should add validation for the 
magic value and metadata size when reading a footer larger than 100KB to verify 
that we are reading something valid. The thrift deserialize failure gives some 
information, but catching this case more specifically would provide a better 
error message.

 

  was:
The file handle cache relies on the mtime to distinguish between different 
versions of a file. For example, if file X exists with mtime=1, then it is 
overwritten and the metadata is updated so that now it is at mtime=2, the file 
handle cache treats them as completely different things and can never use a 
single file handle to serve both. However, some codepaths generate ScanRanges 
with an mtime of -1. This removes the ability to distinguish these two versions 
of a file and can read to consistency problems.

A specific example is the code that reads the parquet footer 
[HdfsParquetScanner::ProcessFooter()|https://github.com/apache/impala/blob/832c9de7810b47b5f782bccb761e07264e7548e5/be/src/exec/parquet/hdfs-parquet-scanner.cc#L1354].
 We don't know ahead of time how big the Parquet footer is. So, we read 100KB 
(determined by 
[FOOTER_SIZE|https://github.com/apache/impala/blob/449fe73d2145bd22f0f857623c3652a097f06d73/be/src/exec/hdfs-scanner.h#L331]).
 If the footer size encoded in the last few bytes of the file indicates that 
the footer is larger than that [code 
here|https://github.com/apache/impala/blob/832c9de7810b47b5f782bccb761e07264e7548e5/be/src/exec/parquet/hdfs-parquet-scanner.cc#L1414],
 then we issue a separate read for the actual size of the footer. That separate 
read does not inherit the mtime of the original read and instead uses an mtime 
of -1. I verified this by adding tracing and issuing a select against 
functional_parquet.widetable_1000_cols.

A failure scenario associated with this is that we read the last 100KB using a 
ScanRange with mtime=2, then we find that the footer is larger than 100KB and 
issue a 

[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Issue Type: New Feature  (was: Bug)

> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are Max 
> Running Queries * * Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Issue Type: Improvement  (was: New Feature)

> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>  Labels: admission-control, resource-management, scalability
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are Max 
> Running Queries * * Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Issue Type: New Feature  (was: Improvement)

> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>  Labels: admission-control, resource-management, scalability
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are Max 
> Running Queries * * Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Labels: admission-control resource-management scalability  (was: )

> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: New Feature
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>  Labels: admission-control, resource-management, scalability
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are Max 
> Running Queries * * Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7486) Admit less memory on dedicated coordinator for admission control purposes

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-7486:

Labels: resource-management scalability  (was: scalability)

> Admit less memory on dedicated coordinator for admission control purposes
> -
>
> Key: IMPALA-7486
> URL: https://issues.apache.org/jira/browse/IMPALA-7486
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Tim Armstrong
>Assignee: Bikramjeet Vig
>Priority: Major
>  Labels: resource-management, scalability
>
> Following on from IMPALA-7349, we should consider handling dedicated 
> coordinators specially rather than admitting a uniform amount of memory on 
> all backends.
> The specific scenario I'm interested in targeting is the case where we a 
> coordinator that is executing many "lightweight" coordinator fragments, e.g. 
> just an ExchangeNode and PlanRootSink, plus maybe other lightweight operators 
> like UnionNode that don't use much memory or CPU. With the current behaviour 
> it's possible for a coordinator to reach capacity from the point-of-view of 
> admission control when at runtime it is actually very lightly loaded.
> This is particularly true if coordinators and executors have different 
> process mem limits. This will be somewhat common since they're often deployed 
> on different hardware or the coordinator will have more memory dedicated to 
> its embedded JVM for the catalog cache.
> More generally we could admit different amounts per backend depending on how 
> many fragments are running, but I think this incremental step would address 
> the most important cases and be a little easier to understand.
> We may want to defer this work until we've implemented distributed runtime 
> filter aggregation, which will significantly reduce coordinator memory 
> pressure, and until we've improved distributed overadmission (since the 
> coordinator behaviour may help throttle overadmission ).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8536) Add Scalable Pool Configuration to Admission Controller.

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8536:

Component/s: Backend

> Add Scalable Pool Configuration to Admission Controller.
> 
>
> Key: IMPALA-8536
> URL: https://issues.apache.org/jira/browse/IMPALA-8536
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.2.0
>Reporter: Andrew Sherman
>Assignee: Andrew Sherman
>Priority: Major
>
> Add configuration parameters to Admission Controller that scale with
> the number of hosts in the resource pool. The planned parameters are Max 
> Running Queries * * Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of concurrently running 
> queries in this pool. 
> * Max Queued Queries Multiple - the Multiple of the current total number of 
> running Impalads which gives the maximum number of queries that can be queued 
> in this pool. 
> * Max Memory Multiple - the Multiple of the current total number of running 
> Impalads which gives the maximum memory available across the cluster for the 
> pool.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-8483) Make coordinator fragment lighter-weight

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-8483:

Labels: resource-management scalability  (was: resource-management)

> Make coordinator fragment lighter-weight
> 
>
> Key: IMPALA-8483
> URL: https://issues.apache.org/jira/browse/IMPALA-8483
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Reporter: Tim Armstrong
>Assignee: Bikramjeet Vig
>Priority: Major
>  Labels: resource-management, scalability
>
> it's possible to write queries with an arbitrarily expensive coordinator 
> fragment. One example is a query with a set of unpartitioned analytic 
> functions each with a different ORDER BY, which results in many SORT nodes in 
> the coordinator fragment.
> This is a problem for dedicated coordinators because it makes the resource 
> consumption unpredictable.
> It would be useful if we could offload that work to executors and guarantee 
> that only "lightweight' operators are part of the coordinator fragment, e.g. 
> operators that take O(#rows returned * log(#executors)) time and 
> O(#executors) memory - regular exchanges, merging exchanges, TOP-N, 
> non-grouping aggregation, etc.
> I think this can be done in general by inserting an additional exchange on 
> top of the "expensive" part of a coordinator fragment, and then ensuring that 
> that fragment gets scheduled on an executor.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-3825) Distribute runtime filter aggregation across cluster

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-3825?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-3825:

Labels: runtime-filters scalability  (was: runtime-filters)

> Distribute runtime filter aggregation across cluster
> 
>
> Key: IMPALA-3825
> URL: https://issues.apache.org/jira/browse/IMPALA-3825
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Distributed Exec
>Affects Versions: Impala 2.6.0
>Reporter: Henry Robinson
>Assignee: Abhishek Rawat
>Priority: Major
>  Labels: runtime-filters, scalability
>
> Runtime filters can be tens of MB or more, and incasting all filters from all 
> shuffle joins to the coordinator can put a lot of memory pressure on that 
> node. To alleviate this we should consider spreading out the aggregation 
> operation across the cluster, so that a different node aggregates each 
> runtime filter.
> This still restricts aggregation to #runtime-filters nodes, which will 
> usually be less than the cluster size. If we want to smooth that out further 
> we could use tree-based aggregation, but let's measure the benefits of simply 
> distributing the aggregation work first.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-7486) Admit less memory on dedicated coordinator for admission control purposes

2019-05-20 Thread Lars Volker (JIRA)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Volker updated IMPALA-7486:

Labels: scalability  (was: )

> Admit less memory on dedicated coordinator for admission control purposes
> -
>
> Key: IMPALA-7486
> URL: https://issues.apache.org/jira/browse/IMPALA-7486
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Reporter: Tim Armstrong
>Assignee: Bikramjeet Vig
>Priority: Major
>  Labels: scalability
>
> Following on from IMPALA-7349, we should consider handling dedicated 
> coordinators specially rather than admitting a uniform amount of memory on 
> all backends.
> The specific scenario I'm interested in targeting is the case where we a 
> coordinator that is executing many "lightweight" coordinator fragments, e.g. 
> just an ExchangeNode and PlanRootSink, plus maybe other lightweight operators 
> like UnionNode that don't use much memory or CPU. With the current behaviour 
> it's possible for a coordinator to reach capacity from the point-of-view of 
> admission control when at runtime it is actually very lightly loaded.
> This is particularly true if coordinators and executors have different 
> process mem limits. This will be somewhat common since they're often deployed 
> on different hardware or the coordinator will have more memory dedicated to 
> its embedded JVM for the catalog cache.
> More generally we could admit different amounts per backend depending on how 
> many fragments are running, but I think this incremental step would address 
> the most important cases and be a little easier to understand.
> We may want to defer this work until we've implemented distributed runtime 
> filter aggregation, which will significantly reduce coordinator memory 
> pressure, and until we've improved distributed overadmission (since the 
> coordinator behaviour may help throttle overadmission ).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



  1   2   3   4   5   >