[ 
https://issues.apache.org/jira/browse/TRAFODION-2017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15304195#comment-15304195
 ] 

ASF GitHub Bot commented on TRAFODION-2017:
-------------------------------------------

Github user DaveBirdsall commented on a diff in the pull request:

    https://github.com/apache/incubator-trafodion/pull/505#discussion_r64923467
  
    --- Diff: core/sql/ustat/hs_globals.cpp ---
    @@ -546,10 +557,13 @@ NABoolean HSGlobalsClass::setHBaseCacheSize(double 
sampleRatio)
       Int64 workableCacheSize = (Int64)(sampleRatio * calibrationFactor);
       if (workableCacheSize < 1)
         workableCacheSize = 1;  // can't go below 1 unfortunately
    +  else if (workableCacheSize > 50)
    +    workableCacheSize = 50; 
     
    -  Int32 max = getDefaultAsLong(HBASE_NUM_CACHE_ROWS_MAX);
    -  if ((workableCacheSize < 10000) && // don't bother if 10000 works
    -      (max == 10000))  // don't do it if user has already set this CQD
    +  // if the user himself set the CQD, don't do anything
    +  NADefaults &defs = ActiveSchemaDB()->getDefaults();
    +  if (defs.getProvenance(HBASE_NUM_CACHE_ROWS_MAX) == 
    +      NADefaults::INIT_DEFAULT_DEFAULTS)
    --- End diff --
    
    Thanks, @selvaganesang. Will change in the next rework.


> Tune HBase row cache sizes so UPDATE STATS completes
> ----------------------------------------------------
>
>                 Key: TRAFODION-2017
>                 URL: https://issues.apache.org/jira/browse/TRAFODION-2017
>             Project: Apache Trafodion
>          Issue Type: Bug
>          Components: sql-cmp
>    Affects Versions: 2.0-incubating
>         Environment: All, though prevalent on clusters with larger tables 
> and/or heavier loads
>            Reporter: David Wayne Birdsall
>            Assignee: David Wayne Birdsall
>             Fix For: 2.1-incubating
>
>
> UPDATE STATISTICS often fails with HBase socket timeout exception and/or 
> scanner timeout exception when run with sampling on larger tables or clusters 
> with heavy concurrent workloads.
> We have experimented in the past with setting various CQDs on large tables to 
> reduce these failures, however we were loathe to set them all the time due to 
> fears that this would lengthen elapsed time in non-failure scenarios. 
> Some recent work by Carol Pearson however shows that the increase in elapsed 
> time is negligible for smaller tables and in failure scenarios the failure 
> does not occur quickly, so paying a small penalty in elapsed time to increase 
> the probability of success seems a better trade-off.
> Carol's work involves tables of less than 1 billion rows. The existing CQD 
> logic is still required for larger tables. But for tables of less than 1 
> billion rows, she recommends setting HBASE_ROWS_CACHED_MIN and 
> HBASE_ROWS_CACHED_MAX to '50'. This JIRA is written to cover this change. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to