[ 
https://issues.apache.org/jira/browse/CASSANDRA-3150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13114220#comment-13114220
 ] 

Mck SembWever edited comment on CASSANDRA-3150 at 9/25/11 11:32 AM:
--------------------------------------------------------------------

{{fullscan-example1.log}} is debug from a "full scan" job. It scans data over a 
full year (and since the cluster's ring range only holds 3 months of data such 
a job guarantees a full scan).

In the debug you see the splits.
{{`nodetool ring`}} gives

{noformat}Address         DC          Rack        Status State   Load           
 Owns    Token                                       
                                                                               
Token(bytes[55555555555555555555555555555554])
152.90.241.22   DC1         RAC1        Up     Normal  16.65 GB        33.33%  
Token(bytes[00])                            
152.90.241.23   DC2         RAC1        Up     Normal  63.22 GB        33.33%  
Token(bytes[2aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa])
152.90.241.24   DC1         RAC1        Up     Normal  72.4 KB         33.33%  
Token(bytes[55555555555555555555555555555554])
{noformat}

The problematic split ends up being 
{noformat}ColumnFamilySplit{startToken='0528cbe0b2b5ff6b816c68b59973bcbc', 
endToken='2aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa', 
dataNodes=[cassandra02.finn.no]}{noformat} This is the split that's receiving 
new data (5-10k rows/second). This new data is being written directly using 
{{StorageProxy.mutate(..)}} with code somewhat similar to the second example in 
[wiki: ScribeToCassandra|http://wiki.apache.org/cassandra/ScribeToCassandra].

      was (Author: michaelsembwever):
    {{fullscan-example1.log}} is debug from a "full scan" job. It scans data 
over a full year (and since the cluster's ring range only holds 3 months of 
data such a job guarantees a full scan).

In the debug you see the splits.
{{`nodetool ring`}} gives

{noformat}Address         DC          Rack        Status State   Load           
 Owns    Token                                       
                                                                               
Token(bytes[55555555555555555555555555555554])
152.90.241.22   DC1         RAC1        Up     Normal  16.65 GB        33.33%  
Token(bytes[00])                            
152.90.241.23   DC2         RAC1        Up     Normal  63.22 GB        33.33%  
Token(bytes[2aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa])
152.90.241.24   DC1         RAC1        Up     Normal  72.4 KB         33.33%  
Token(bytes[55555555555555555555555555555554])
{noformat}

The problematic split ends up being 
{noformat}ColumnFamilySplit{startToken='0528cbe0b2b5ff6b816c68b59973bcbc', 
endToken='2aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa', 
dataNodes=[cassandra02.finn.no]}{noformat} This is the split that's receiving 
new data (5-10k rows/second).
  
> ColumnFormatRecordReader loops forever (StorageService.getSplits(..) out of 
> whack)
> ----------------------------------------------------------------------------------
>
>                 Key: CASSANDRA-3150
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-3150
>             Project: Cassandra
>          Issue Type: Bug
>          Components: Hadoop
>    Affects Versions: 0.8.4, 0.8.5
>            Reporter: Mck SembWever
>            Assignee: Mck SembWever
>            Priority: Critical
>             Fix For: 0.8.6
>
>         Attachments: CASSANDRA-3150.patch, Screenshot-Counters for 
> task_201109212019_1060_m_000029 - Mozilla Firefox.png, Screenshot-Hadoop map 
> task list for job_201109212019_1060 on cassandra01 - Mozilla Firefox.png, 
> attempt_201109071357_0044_m_003040_0.grep-get_range_slices.log, 
> fullscan-example1.log
>
>
> From http://thread.gmane.org/gmane.comp.db.cassandra.user/20039
> {quote}
> bq. Cassandra-0.8.4 w/ ByteOrderedPartitioner
> bq. CFIF's inputSplitSize=196608
> bq. 3 map tasks (from 4013) is still running after read 25 million rows.
> bq. Can this be a bug in StorageService.getSplits(..) ?
> getSplits looks pretty foolproof to me but I guess we'd need to add
> more debug logging to rule out a bug there for sure.
> I guess the main alternative would be a bug in the recordreader paging.
> {quote}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to