BTW, the option to use to change slice target is planner.slice_target. For example, alter session set `planner.slice_target` = 60000;
Thanks, Padma On Apr 6, 2017, at 1:38 PM, Padma Penumarthy <[email protected]<mailto:[email protected]>> wrote: Can you send the query profile ? How is your data distributed i.e. how big is the table and how many regions and avg row count per region ? This problem can happen when we don’t have a minor fragment (for scanning) scheduled on a node which is hosting one or more hbase regions. That can happen if we do not have enough work to do (based on total rowCount and slice target) to schedule fragments on all nodes. One thing you can try is lower the slice target so we create fragments on all nodes. Depending upon your configuration, making it close to average rowCount per region might be the ideal thing to do. Thanks, Padma On Apr 6, 2017, at 12:03 PM, John Omernik <[email protected]<mailto:[email protected]>> wrote: By the way, are there any ways to manually prod the data to make it so the queries work again? It seems like an off by one type issue, can I add something to my data make it work? On Thu, Apr 6, 2017 at 2:03 PM, John Omernik <[email protected]<mailto:[email protected]>> wrote: Oh nice, 1.10 from MapR has the fix? Great... Looking forward to that! Thanks! On Thu, Apr 6, 2017 at 1:59 PM, Abhishek Girish <[email protected]<mailto:[email protected]>> wrote: I'm guessing the next release of Apache Drill could be a few months away. MapR Drill 1.10.0 release (which does contain the fix for DRILL-5395) should be out shortly, within the next week or so. On Thu, Apr 6, 2017 at 11:50 AM, John Omernik <[email protected]<mailto:[email protected]>> wrote: Nope no other issues. I was waiting on 1.10 to be available from MapR, do we know the release date for 1.11? On Thu, Apr 6, 2017 at 1:31 PM, Abhishek Girish <[email protected]<mailto:[email protected]>> wrote: Are there any other issues apart from the one being discussed? Builds from Apache should work with MapR-DB tables (when built with mapr profile). Let us know if you are having any trouble. The fix for DRILL-5395 should be available this week, afaik. You could also build off Padma's branch if you need it urgently. On Thu, Apr 6, 2017 at 11:25 AM, John Omernik <[email protected]<mailto:[email protected]>> wrote: Is there any work around except wait? This is unfortunate... I wonder if I could beg a MapR build off the MapR team if I offered them beer/cookies. ( I have been unable to get MapR Tables to work with Builds off the Apache main line) On Thu, Apr 6, 2017 at 1:06 PM, Abhishek Girish <[email protected]<mailto:[email protected]> wrote: Could be related to DRILL-5395 <https://issues.apache.org/jira/browse/DRILL-5395>. Once committed, the fix should be available in Apache master. On Thu, Apr 6, 2017 at 10:56 AM, John Omernik <[email protected]<mailto:[email protected]>> wrote: Hello all, I am using Drill 1.8 and MapR 5.2. I just finished a large load of data into a mapr table. I was able to confirm that the table returns data from the c api for hbase, so no issue there, however, when I select from the table in Drill, either from the table directly, or from a view I created, then I get the NPE as listed below. Any advice on how to troubleshoot further would be appreciated! 0: jdbc:drill:zk:zeta2.brewingintel.com<http://brewingintel.com>:5181,> select * from maprpcaps limit 1; Error: SYSTEM ERROR: NullPointerException [Error Id: 6abaeadb-6e1b-4dce-9d1f-54b99a40becb on zeta4.brewingintel.com<http://zeta4.brewingintel.com>:20005] (org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception during fragment initialization: null org.apache.drill.exec.work.foreman.Foreman.run():281 java.util.concurrent.ThreadPoolExecutor.runWorker():1142 java.util.concurrent.ThreadPoolExecutor$Worker.run():617 java.lang.Thread.run():745 Caused By (java.lang.NullPointerException) null org.apache.drill.exec.store.mapr.db.MapRDBGroupScan. applyAssignments():205 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitGroupScan():116 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitGroupScan():103 org.apache.drill.exec.physical.base. AbstractGroupScan.accept():63 org.apache.drill.exec.physical.base.AbstractPhysicalVisitor. visitChildren():138 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitOp():134 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitOp():103 org.apache.drill.exec.physical.base.AbstractPhysicalVisitor. visitLimit():92 org.apache.drill.exec.physical.config.Limit.accept():57 org.apache.drill.exec.physical.base.AbstractPhysicalVisitor. visitChildren():138 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitOp():134 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitOp():103 org.apache.drill.exec.physical.config. SelectionVectorRemover.accept( ):42 org.apache.drill.exec.physical.base.AbstractPhysicalVisitor. visitChildren():138 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitOp():134 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitExchange():108 org.apache.drill.exec.planner.fragment.Wrapper$ AssignEndpointsToScanAndStore.visitExchange():103 org.apache.drill.exec.physical.base. AbstractExchange.accept():123 org.apache.drill.exec.planner.fragment.Wrapper. assignEndpoints():148 org.apache.drill.exec.planner.fragment. SoftAffinityFragmentParalleliz er.parallelizeFragment():86 org.apache.drill.exec.planner.fragment.SimpleParallelizer. parallelizeFragment():251 org.apache.drill.exec.planner.fragment.SimpleParallelizer. parallelizeFragment():243 org.apache.drill.exec.planner.fragment.SimpleParallelizer. getFragmentsHelper():167 org.apache.drill.exec.planner.fragment.SimpleParallelizer. getFragments():126 org.apache.drill.exec.work.foreman.Foreman. getQueryWorkUnit():596 org.apache.drill.exec.work.foreman.Foreman. runPhysicalPlan():426 org.apache.drill.exec.work.foreman.Foreman.runSQL():1010 org.apache.drill.exec.work.foreman.Foreman.run():264 java.util.concurrent.ThreadPoolExecutor.runWorker():1142 java.util.concurrent.ThreadPoolExecutor$Worker.run():617 java.lang.Thread.run():745 (state=,code=0)
