Re: NPE during compaction in compare

2013-07-25 Thread Sylvain Lebresne
Are you sure you're using the last version of the 1.2 branch? The
exceptions are in RangeTombstoneList that has been introduced by the patch
of CASSANDRA-5677 and the initial version of the patch were indeed able to
produce that kind of error. Could that be you're using such development
version of the patch?

What I would suggest is to test the current artifacts for 1.2.7 at
https://repository.apache.org/content/repositories/orgapachecassandra-013/org/apache/cassandra/apache-cassandra/1.2.7/.
This is what should be released as soon as the vote period is complete. If
you can reproduce with that version, we'd be very interested to know (you
can even feel free to open a ticket on JIRA).

--
Sylvain


On Wed, Jul 24, 2013 at 10:52 PM, Paul Ingalls paulinga...@gmail.comwrote:

 Hey Chris,

 so I just tried dropping all my data and converting my column families to
 use leveled compaction.  Now I'm getting exceptions like the following once
 I start inserting data.  Have you seen these?



 ERROR 13:13:25,616 Exception in thread Thread[CompactionExecutor:34,1,main]
 java.lang.NullPointerException
 at
 org.apache.cassandra.db.marshal.AbstractCompositeType.compare(AbstractCompositeType.java:69)
 at
 org.apache.cassandra.db.marshal.AbstractCompositeType.compare(AbstractCompositeType.java:31)
 at
 org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:396)
 at
 org.apache.cassandra.db.RangeTombstoneList.addAll(RangeTombstoneList.java:205)
 at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:180)
 at
 org.apache.cassandra.db.AbstractThreadUnsafeSortedColumns.delete(AbstractThreadUnsafeSortedColumns.java:40)
 at
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:51)
 at
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:46)
 at
 org.apache.cassandra.db.compaction.PrecompactedRow.merge(PrecompactedRow.java:115)
 at
 org.apache.cassandra.db.compaction.PrecompactedRow.init(PrecompactedRow.java:98)
 at
 org.apache.cassandra.db.compaction.CompactionController.getCompactedRow(CompactionController.java:160)
 at
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:76)
 at
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:57)
 at
 org.apache.cassandra.utils.MergeIterator$ManyToOne.consume(MergeIterator.java:114)
 at
 org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:97)
 at
 com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
 at
 com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
 at
 org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:145)
 at
 org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48)
 at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
 at
 org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:58)
 at
 org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:60)
 at
 org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:211)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
 at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
 at java.util.concurrent.FutureTask.run(FutureTask.java:138)
 at
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
 at
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
 at java.lang.Thread.run(Thread.java:680)

 and

 ERROR 13:17:11,327 Exception in thread Thread[CompactionExecutor:45,1,main]
 java.lang.ArrayIndexOutOfBoundsException: 2
 at
 org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:396)
 at
 org.apache.cassandra.db.RangeTombstoneList.addAll(RangeTombstoneList.java:205)
 at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:180)
 at
 org.apache.cassandra.db.AbstractThreadUnsafeSortedColumns.delete(AbstractThreadUnsafeSortedColumns.java:40)
 at
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:51)
 at
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:46)
 at
 org.apache.cassandra.db.compaction.PrecompactedRow.merge(PrecompactedRow.java:115)
 at
 org.apache.cassandra.db.compaction.PrecompactedRow.init(PrecompactedRow.java:98)
 at
 org.apache.cassandra.db.compaction.CompactionController.getCompactedRow(CompactionController.java:160)
 at
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:76)
 at
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:57)
 at
 org.apache.cassandra.utils.MergeIterator$ManyToOne.consume(MergeIterator.java:114)
 at
 

Re: unable to compact large rows

2013-07-25 Thread Sylvain Lebresne
Obvisously, that should not happen. That being said, we did fix a bug
yesterday that could produce that kind of trace:
https://issues.apache.org/jira/browse/CASSANDRA-5799. It will be part of
1.2.7 that should be released tomorrow, though the artifacts that are being
voted on are at
https://repository.apache.org/content/repositories/orgapachecassandra-013/org/apache/cassandra/apache-cassandra/1.2.7/if
you want to test sooner than later.

If you still can reproduce with that version, feel free to open a ticket on
JIRA (in which case, the more information you can give us to reproduce, the
better the change we'd be able to fix it quickly).

--
Sylvain


On Wed, Jul 24, 2013 at 10:54 PM, Paul Ingalls paulinga...@gmail.comwrote:

 It is pretty much every row that hits the large threshold.  I don't think
 I can delete every row that hits that…

 you can see the db size in the stack trace, do you want a different type
 of size?

 On Jul 24, 2013, at 11:07 AM, Jason Wee peich...@gmail.com wrote:

 Would it possible to delete this row and reinsert this row? By the way,
 how large is that one row?

 Jason


 On Wed, Jul 24, 2013 at 9:23 AM, Paul Ingalls paulinga...@gmail.comwrote:

 I'm getting constant exceptions during compaction of large rows.  In
 fact, I have not seen one work, even starting from an empty DB.  As soon as
 I start pushing in data, when a row hits the large threshold, it fails
 compaction with this type of stack trace:

  INFO [CompactionExecutor:6] 2013-07-24 01:17:53,592
 CompactionController.java (line 156) Compacting large row
 fanzo/tweets_by_id:352567939972603904 (153360688 bytes) incrementally
 ERROR [CompactionExecutor:6] 2013-07-24 01:18:12,496 CassandraDaemon.java
 (line 192) Exception in thread Thread[CompactionExecutor:6,1,main]
 java.lang.AssertionError: incorrect row data size 5722610 written to
 /mnt/datadrive/lib/cassandra/data/fanzo/tweets_by_id/fanzo-tweets_by_id-tmp-ic-1453-Data.db;
 correct is 5767384
 at
 org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:162)
 at
 org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:162)
 at
 org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48)
 at
 org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
 at
 org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:58)
 at
 org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:60)
 at
 org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:211)
 at
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
 at
 java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
 at java.util.concurrent.FutureTask.run(FutureTask.java:166)
 at
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:724)

 I'm not sure what to do or where to look.  Help…:)

 Thanks,

 Paul







Re: disappointed

2013-07-25 Thread Derek Andree
Yeah, Rob is smart.  don't run crap in production.  Run what others are stable 
at.  If you are running the latest greatest dumbest craziest in prod then you 
ask for fail, and you will get just that.

FAIL

On Jul 24, 2013, at 12:06 PM, Robert Coli rc...@eventbrite.com wrote:

 A better solution would likely involve not running cutting edge code in 
 production.. if you find yourself needing to upgrade production anything on 
 the day of a release, you are probably ahead of the version it is reasonable 
 to run in production.
 
 If you're already comfortable with this high level of risk in production, I 
 don't really see small manual patches as significantly increasing your level 
 of risk...
 
 =Rob 



Re: disappointed

2013-07-25 Thread Colin
Mysql?

--
Colin 
+1 320 221 9531

 

On Jul 25, 2013, at 6:08 AM, Derek Andree dand...@lacunasystems.com wrote:

 Yeah, Rob is smart.  don't run crap in production.  Run what others are 
 stable at.  If you are running the latest greatest dumbest craziest in prod 
 then you ask for fail, and you will get just that.
 
 FAIL
 
 On Jul 24, 2013, at 12:06 PM, Robert Coli rc...@eventbrite.com wrote:
 
 A better solution would likely involve not running cutting edge code in 
 production.. if you find yourself needing to upgrade production anything on 
 the day of a release, you are probably ahead of the version it is reasonable 
 to run in production.
 
 If you're already comfortable with this high level of risk in production, I 
 don't really see small manual patches as significantly increasing your level 
 of risk...
 
 =Rob
 


[BETA RELEASE] Apache Cassandra 2.0.0-beta2 released

2013-07-25 Thread Sylvain Lebresne
The Cassandra team is pleased to announce the release of the second beta for
the future Apache Cassandra 2.0.0.

Let me first stress that this is still beta software and as such is *not*
ready
for production use.

As with the first beta, the goal is to give a preview of what will become
Cassandra 2.0 and to get wider testing before the final release. Please
report
any problem you may encounter[3,4] with this release and have a look at the
change log[1] and release notes[2] to see where Cassandra 2.0 differs from
the
previous series.

Apache Cassandra 2.0.0-beta2[5] is available as usual from the cassandra
website (http://cassandra.apache.org/download/) and a debian package is
available using the 20x branch (see
http://wiki.apache.org/cassandra/DebianPackaging).

Thank you for your help in testing and have fun with it.

[1]: http://goo.gl/jY3FYT (CHANGES.txt)
[2]: http://goo.gl/r1AcVZ (NEWS.txt)
[3]: https://issues.apache.org/jira/browse/CASSANDRA
[4]: user@cassandra.apache.org
[5]:
http://git-wip-us.apache.org/repos/asf?p=cassandra.git;a=shortlog;h=refs/tags/cassandra-2.0.0-beta2


Re: sstable size change

2013-07-25 Thread Keith Wright
Unfortunately the table in question is a CQL3 table so cassandra-cli will not 
output its describe:

WARNING: CQL3 tables are intentionally omitted from 'describe' output.
See https://issues.apache.org/jira/browse/CASSANDRA-4377 for details.

However, I did figure out that apparently I was not setting the change 
properly.  I was using the following alter via CQL3:

alter table shard_user_lookup with compaction_strategy_options = 
{'sstable_size_in_mb':256};

But when I did a describe, I did not see the change.

CREATE TABLE shard_user_lookup (
  shard_user_id int,
  shard text,
  creation_time timestamp,
  status int,
  user_id timeuuid,
  PRIMARY KEY (shard_user_id, shard)
) WITH
  bloom_filter_fp_chance=0.10 AND
  caching='KEYS_ONLY' AND
  comment='' AND
  dclocal_read_repair_chance=0.00 AND
  gc_grace_seconds=86400 AND
  read_repair_chance=0.10 AND
  replicate_on_write='true' AND
  populate_io_cache_on_flush='false' AND
  compaction={'class': 'LeveledCompactionStrategy'} AND
  compression={'chunk_length_kb': '8', 'crc_check_chance': '0.1', 
'sstable_compression': 'LZ4Compressor'};

If I then execute the following via CQL2, I do see the property.  Seems odd to 
me?

alter table shard_user_lookup with 
compaction_strategy_options:sstable_size_in_mb=256;

CREATE TABLE shard_user_lookup (
  shard_user_id int,
  shard text,
  creation_time timestamp,
  status int,
  user_id timeuuid,
  PRIMARY KEY (shard_user_id, shard)
) WITH
  bloom_filter_fp_chance=0.10 AND
  caching='KEYS_ONLY' AND
  comment='' AND
  dclocal_read_repair_chance=0.00 AND
  gc_grace_seconds=86400 AND
  read_repair_chance=0.10 AND
  replicate_on_write='true' AND
  populate_io_cache_on_flush='false' AND
  compaction={'sstable_size_in_mb': '256', 'class': 
'LeveledCompactionStrategy'} AND
  compression={'chunk_length_kb': '8', 'crc_check_chance': '0.1', 
'sstable_compression': 'LZ4Compressor'};


From: Wei Zhu wz1...@yahoo.commailto:wz1...@yahoo.com
Reply-To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
user@cassandra.apache.orgmailto:user@cassandra.apache.org, Wei Zhu 
wz1...@yahoo.commailto:wz1...@yahoo.com
Date: Wednesday, July 24, 2013 8:49 PM
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: Re: sstable size change

what is output of show keyspaces from cassandra-cli, did you see the new value?

  Compaction Strategy: 
org.apache.cassandra.db.compaction.LeveledCompactionStrategy
  Compaction Strategy Options:
sstable_size_in_mb: XXX


From: Keith Wright kwri...@nanigans.commailto:kwri...@nanigans.com
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
user@cassandra.apache.orgmailto:user@cassandra.apache.org
Sent: Wednesday, July 24, 2013 3:44 PM
Subject: Re: sstable size change

Hi all,

   This morning I increased the SSTable size for one of my LCS via an alter 
command and saw at least one compaction run (I did not trigger a compaction via 
nodetool nor upgrades stables nor removing the .json file).  But so far my data 
sizes appear the same at the default 5 MB (see below for output of ls –Sal as 
well as relevant portion of cfstats).   Is this expected?  I was hoping to see 
at least one file at the new 256 MB size I set.

Thanks

SSTable count: 4965
SSTables in each level: [0, 10, 112/100, 1027/1000, 3816, 0, 0, 0]
Space used (live): 29062393142
Space used (total): 29140547702
Number of Keys (estimate): 195103104
Memtable Columns Count: 441483
Memtable Data Size: 205486218
Memtable Switch Count: 243
Read Count: 154226729

-rw-rw-r--  1 cassandra cassandra 5247564 Jul 18 01:33 
users-shard_user_lookup-ib-97153-Data.db
-rw-rw-r--  1 cassandra cassandra 5247454 Jul 23 02:59 
users-shard_user_lookup-ib-109063-Data.db
-rw-rw-r--  1 cassandra cassandra 5247421 Jul 20 14:58 
users-shard_user_lookup-ib-103127-Data.db
-rw-rw-r--  1 cassandra cassandra 5247415 Jul 17 13:56 
users-shard_user_lookup-ib-95761-Data.db
-rw-rw-r--  1 cassandra cassandra 5247379 Jul 21 02:44 
users-shard_user_lookup-ib-104718-Data.db
-rw-rw-r--  1 cassandra cassandra 5247346 Jul 21 21:54 
users-shard_user_lookup-ib-106280-Data.db
-rw-rw-r--  1 cassandra cassandra 5247242 Jul  3 19:41 
users-shard_user_lookup-ib-66049-Data.db
-rw-rw-r--  1 cassandra cassandra 5247235 Jul 21 02:44 
users-shard_user_lookup-ib-104737-Data.db
-rw-rw-r--  1 cassandra cassandra 5247233 Jul 20 14:58 
users-shard_user_lookup-ib-103169-Data.db


From: sankalp kohli kohlisank...@gmail.commailto:kohlisank...@gmail.com
Reply-To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
user@cassandra.apache.orgmailto:user@cassandra.apache.org
Date: Tuesday, July 23, 2013 3:04 PM
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: Re: sstable size change

Will Cassandra force any newly compacted files to my new setting as 
compactions are naturally 

TTL, Tombstones, and gc_grace

2013-07-25 Thread Michael Theroux
Hello,

Quick question on Cassandra, TTLs, tombstones, and GC grace.  If we have a 
column family whose only mechanism of deleting columns is utilizing TTLs, is 
repair really necessary to make tombstones consistent, and therefore would it 
be safe to set the gc grace period of the column family to a very low value?

I ask because of this blog post based on Cassandra .7: 
http://www.datastax.com/dev/blog/whats-new-cassandra-07-expiring-columns.

The first time the expired column is compacted, it is transformed into a 
tombstone. This transformation frees some disk space: the size of the value of 
the expired column. From that moment on, the column is a normal tombstone and 
follows the tombstone rules: it will be totally removed by compaction 
(including minor ones in most cases since Cassandra 0.6.6) after 
GCGraceSeconds.

Since tombstones are not written using a replicated write, but instead written 
during compaction, theoretically, it shouldn't be possible to lose a tombstone? 
 Or is this blog post inaccurate for later versions of cassandra?  We are using 
cassandra 1.1.11.

Thanks,
-Mike




Re: TTL, Tombstones, and gc_grace

2013-07-25 Thread horschi
Hi Michael,

yes, you should never loose a delete, because there are no real deletes. No
matter what version you are using.

btw: There is actually a ticket that builds an optimization on top of that
assumption: CASSANDRA-4917. Basically, if TTLgc_grace then do not create
tombstones for expiring-columns. This works because disappear anyway if TTL
is over.

cheers,
Christian


On Thu, Jul 25, 2013 at 3:24 PM, Michael Theroux mthero...@yahoo.comwrote:

 Hello,

 Quick question on Cassandra, TTLs, tombstones, and GC grace.  If we have a
 column family whose only mechanism of deleting columns is utilizing TTLs,
 is repair really necessary to make tombstones consistent, and therefore
 would it be safe to set the gc grace period of the column family to a very
 low value?

 I ask because of this blog post based on Cassandra .7:
 http://www.datastax.com/dev/blog/whats-new-cassandra-07-expiring-columns.

 The first time the expired column is compacted, it is transformed into a
 tombstone. This transformation frees some disk space: the size of the value
 of the expired column. From that moment on, the column is a normal
 tombstone and follows the tombstone rules: it will be totally removed by
 compaction (including minor ones in most cases since Cassandra 0.6.6) after
 GCGraceSeconds.

 Since tombstones are not written using a replicated write, but instead
 written during compaction, theoretically, it shouldn't be possible to lose
 a tombstone?  Or is this blog post inaccurate for later versions of
 cassandra?  We are using cassandra 1.1.11.

 Thanks,
 -Mike





Re: Strange cassandra-stress results with 2.0.0 beta1

2013-07-25 Thread Sávio Teles
Some bug was fixed in 2.0.0-beta2 by C* developers. Try it!


2013/7/22 Andrew Cobley a.e.cob...@dundee.ac.uk

  I've been noticing some strange casandra-stress results with 2.0.0 beta
 1.  I've set up a single node on a Mac (4 gig ram, 2.8Ghz core 2 duo) and
 installed 2.0.0 beta1.

 When I run ./cassandra-stress  -d 134.36.36.218 I'm seeing the
 interval-op-rate drop from a peek of 11562 at the start to 0 -20 after
 about 358 seconds.  Here's the output:

 ./cassandra-stress  -d 134.36.36.218
 Created keyspaces. Sleeping 1s for propagation.
 total,interval_op_rate,interval_key_rate,latency,95th,99th,elapsed_time
 13629,1362,1362,15.3,117.6,531.1,10
 37545,2391,2391,9.4,74.0,438.4,20
 105410,6786,6786,3.3,38.0,216.4,30
 209014,10360,10360,2.7,26.2,216.5,41
 320024,11101,11101,2.5,14.5,216.5,51
 430216,11019,11019,2.4,10.5,215.4,61
 531905,10168,10168,2.4,8.3,177.1,72
 619672,8776,8776,2.4,6.2,145.8,82
 653977,3430,3430,2.4,5.9,145.4,92
 682648,2867,2867,2.4,6.1,145.4,102
 692771,1012,1012,2.4,6.1,145.4,113
 712163,1939,1939,2.4,6.0,4461.7,123
 723761,1159,1159,2.4,6.0,4461.7,133
 731115,735,735,2.4,6.1,4461.7,143
 737627,651,651,2.4,6.2,4461.7,153
 743853,622,622,2.4,7.0,5056.1,164
 747345,349,349,2.4,7.0,5056.1,174
 755501,815,815,2.5,6.9,6351.7,184
 758692,319,319,2.5,6.9,5056.1,195
 763960,526,526,2.5,7.3,5292.6,205
 767966,400,400,2.5,7.4,5292.6,215
 769514,154,154,2.5,7.4,5292.6,225
 773492,397,397,2.5,7.2,6435.7,236
 775374,188,188,2.5,7.4,5292.6,246
 776035,66,66,2.6,7.5,5347.1,256
 777896,186,186,2.6,7.9,6407.3,266
 778791,89,89,2.6,8.0,6438.9,276
 779646,85,85,2.6,8.0,6523.2,287
 780785,113,113,2.6,11.1,6523.2,297
 781336,55,55,2.6,11.7,6523.2,307
 781684,34,34,2.6,13.3,16734.0,317
 781818,13,13,2.7,13.9,16764.2,328
 781952,13,13,2.7,13.9,16900.1,338
 782195,24,24,2.7,15.3,16900.1,348
 782294,9,9,2.7,15.3,16900.1,358
 782362,6,6,2.7,15.5,21487.3,368
 782589,22,22,2.7,20.1,21487.3,379
 782775,18,18,2.7,28.8,21487.3,389
 783012,23,23,2.8,29.4,21487.3,399
 783248,23,23,2.8,68.2,21487.3,409
 783270,2,2,2.8,68.2,21487.3,419
 783283,1,1,2.8,68.2,21487.3,430
 783283,0,0,2.8,68.2,21487.3,440
 783337,5,5,2.8,198.6,42477.2,450
 783492,15,15,2.9,5057.6,52687.3,460
 783581,8,8,2.9,5349.0,59306.9,471
 783807,22,22,3.0,6429.1,59306.9,481
 783816,0,0,3.0,6438.3,59306.9,491
 783945,12,12,3.2,11566.9,59306.9,501
 783962,1,1,3.2,11612.0,59306.9,512
 783962,0,0,3.2,11612.0,59306.9,522


 I installed 1.2.6 for comparision and that gave  expected results:

 ./cassandra-stress  -d 134.36.36.218
 Unable to create stress keyspace: Keyspace names must be
 case-insensitively unique (Keyspace1 conflicts with Keyspace1)
 total,interval_op_rate,interval_key_rate,latency/95th/99th,elapsed_time
 22158,2215,2215,0.8,8.9,549.0,10
 64192,4203,4203,1.0,7.0,486.2,20
 114236,5004,5004,1.3,6.5,411.9,30
 178407,6417,6417,1.4,5.5,411.9,40
 281714,10330,10330,2.2,5.1,409.6,51
 371684,8997,8997,2.4,5.1,407.5,61
 464401,9271,9271,2.5,5.0,236.7,72
 562797,9839,9839,2.6,5.2,87.5,82
 655755,9295,9295,2.6,5.2,87.5,92
 751560,9580,9580,2.6,4.7,179.9,103
 848022,9646,9646,2.6,4.5,826.0,113
 914539,6651,6651,2.6,4.7,823.2,123
 100,8546,8546,2.6,5.1,1048.5,133

 Any ideas whats causing the slow down ?

 Andy


 The University of Dundee is a registered Scottish Charity, No: SC015096




-- 
Atenciosamente,
Sávio S. Teles de Oliveira
voice: +55 62 9136 6996
http://br.linkedin.com/in/savioteles
Mestrando em Ciências da Computação - UFG
Arquiteto de Software
Laboratory for Ubiquitous and Pervasive Applications (LUPA) - UFG


Re: About column family

2013-07-25 Thread horschi
With 1.2.7 you can use -Dcassandra.unsafesystem. That will speed up cf
creation. So you will get in even more trouble even faster!



On Tue, Jul 23, 2013 at 12:23 PM, bjbylh bjb...@me.com wrote:

 Hi all:
 i have two questions to ask:
 1,how many column families can be created in a cluster?is there a limit to
 the number of it?
 2,it spents 2-5 seconds to create a new cf while the cluster contains
 about 1 cfs(if the cluster is empty,it spents about 0.5s).is it
 normal?how to improve the efficiency of creating cf?
 btw C* is 1.2.4.
 thanks a lot.

 Sent from Samsung Mobile



Re: Strange cassandra-stress results with 2.0.0 beta1

2013-07-25 Thread Andrew Cobley
I'm getting something similar for beta 2, you can see here it's dropping from 
12000 to 400 quite quickly.   The output from cassandra is available n dropbox 
here:

https://dl.dropboxusercontent.com/u/1638201/output.txt (184kBytes)

I'm wondering if it's a GC issue ?

 ./cassandra-stress -d 134.36.36.218 -n 3000
Unable to create stress keyspace: Keyspace names must be case-insensitively 
unique (Keyspace1 conflicts with Keyspace1)
total,interval_op_rate,interval_key_rate,latency/95th/99th,elapsed_time
73761,7376,7376,2.1,7.6,181.0,10
199159,12539,12539,2.4,7.5,102.4,20
316978,11781,11781,2.4,7.8,172.7,30
434060,11708,11708,2.4,8.1,170.6,40
552650,11859,11859,2.5,7.3,170.8,50
602432,4978,4978,2.5,7.3,170.8,60
614364,1193,1193,2.5,7.5,5490.4,71
634199,1983,1983,2.5,8.1,5490.4,81
653293,1909,1909,2.5,8.9,5490.4,91
658398,510,510,2.5,8.7,5490.4,101
671805,1340,1340,2.6,8.7,5490.4,111
678677,687,687,2.6,9.1,5490.4,121
685247,657,657,2.6,10.2,5626.8,131
691519,627,627,2.6,10.7,5626.8,141
693682,216,216,2.6,10.7,5626.8,151
700082,640,640,2.7,11.9,6460.1,161
703952,387,387,2.7,12.5,6460.1,172
706007,205,205,2.7,12.5,6460.1,182
709988,398,398,2.7,12.7,6647.0,192
^Cmacaroon:bin administrator$

Andy C

On 25 Jul 2013, at 15:28, Sávio Teles 
savio.te...@lupa.inf.ufg.brmailto:savio.te...@lupa.inf.ufg.br wrote:

Some bug was fixed in 2.0.0-beta2 by C* developers. Try it!


2013/7/22 Andrew Cobley 
a.e.cob...@dundee.ac.ukmailto:a.e.cob...@dundee.ac.uk
I've been noticing some strange casandra-stress results with 2.0.0 beta 1.  
I've set up a single node on a Mac (4 gig ram, 2.8Ghz core 2 duo) and installed 
2.0.0 beta1.

When I run ./cassandra-stress  -d 134.36.36.218 I'm seeing the interval-op-rate 
drop from a peek of 11562 at the start to 0 -20 after about 358 seconds.  
Here's the output:




The University of Dundee is a registered Scottish Charity, No: SC015096


maximum storage per node

2013-07-25 Thread Pruner, Anne (Anne)
Does anyone have opinions on the maximum amount of data reasonable to store on 
one Cassandra node?  If there are limitations, what are the reasons for it?

Thanks,
Anne


Re: maximum storage per node

2013-07-25 Thread cem
Between 500GB - 1TB is recommended.

But it depends also your hardware, traffic characteristics and
requirements. Can you give some details on that?

Best Regards,
Cem


On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) pru...@avaya.comwrote:

  Does anyone have opinions on the maximum amount of data reasonable to
 store on one Cassandra node?  If there are limitations, what are the
 reasons for it?

 ** **

 Thanks,

 Anne



RE: maximum storage per node

2013-07-25 Thread Pruner, Anne (Anne)
We're storing fairly large files (about 1MB apiece) for a few months and then 
deleting the oldest to get more space to add new ones.  We have large 
requirements (maybe up to 100 TB), so having a 1TB limit would be unworkable.

What is the reason for the limit?  Does something fail after that?

If there are hardware issues, what's recommended?

BTW, we're using Cassandra 1.2

Anne

From: cem [mailto:cayiro...@gmail.com]
Sent: Thursday, July 25, 2013 11:41 AM
To: user@cassandra.apache.org
Subject: Re: maximum storage per node

Between 500GB - 1TB is recommended.

But it depends also your hardware, traffic characteristics and requirements. 
Can you give some details on that?

Best Regards,
Cem

On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) 
pru...@avaya.commailto:pru...@avaya.com wrote:
Does anyone have opinions on the maximum amount of data reasonable to store on 
one Cassandra node?  If there are limitations, what are the reasons for it?

Thanks,
Anne



RE: maximum storage per node

2013-07-25 Thread Kanwar Sangha
Issues with large data nodes would be -


* Nodetool repair will be impossible to run

* Your read i/o will suffer since you will almost always go to disk 
(each read will take 3 IOPS worst case)

* Boot-straping the node in case of failures will take days/weeks



From: Pruner, Anne (Anne) [mailto:pru...@avaya.com]
Sent: 25 July 2013 10:45
To: user@cassandra.apache.org
Subject: RE: maximum storage per node

We're storing fairly large files (about 1MB apiece) for a few months and then 
deleting the oldest to get more space to add new ones.  We have large 
requirements (maybe up to 100 TB), so having a 1TB limit would be unworkable.

What is the reason for the limit?  Does something fail after that?

If there are hardware issues, what's recommended?

BTW, we're using Cassandra 1.2

Anne

From: cem [mailto:cayiro...@gmail.com]
Sent: Thursday, July 25, 2013 11:41 AM
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: Re: maximum storage per node

Between 500GB - 1TB is recommended.

But it depends also your hardware, traffic characteristics and requirements. 
Can you give some details on that?

Best Regards,
Cem

On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) 
pru...@avaya.commailto:pru...@avaya.com wrote:
Does anyone have opinions on the maximum amount of data reasonable to store on 
one Cassandra node?  If there are limitations, what are the reasons for it?

Thanks,
Anne



Re: maximum storage per node

2013-07-25 Thread cem
You will suffer from long compactions if you are planning to get rid of
from old records by TTL.

Best Regards,
Cem.


On Thu, Jul 25, 2013 at 5:51 PM, Kanwar Sangha kan...@mavenir.com wrote:

  Issues with large data nodes would be –

 ** **

 **· **Nodetool repair will be impossible to run

 **· **Your read i/o will suffer since you will almost always go
 to disk (each read will take 3 IOPS worst case)

 **· **Boot-straping the node in case of failures will take
 days/weeks

 ** **

 ** **

 *From:* Pruner, Anne (Anne) [mailto:pru...@avaya.com]
 *Sent:* 25 July 2013 10:45
 *To:* user@cassandra.apache.org
 *Subject:* RE: maximum storage per node

 ** **

 We’re storing fairly large files (about 1MB apiece) for a few months and
 then deleting the oldest to get more space to add new ones.  We have large
 requirements (maybe up to 100 TB), so having a 1TB limit would be
 unworkable.

 ** **

 What is the reason for the limit?  Does something fail after that?

 ** **

 If there are hardware issues, what’s recommended?

 ** **

 BTW, we’re using Cassandra 1.2

 ** **

 Anne

 ** **

 *From:* cem [mailto:cayiro...@gmail.com cayiro...@gmail.com]
 *Sent:* Thursday, July 25, 2013 11:41 AM
 *To:* user@cassandra.apache.org
 *Subject:* Re: maximum storage per node

 ** **

 Between 500GB - 1TB is recommended. 

 ** **

 But it depends also your hardware, traffic characteristics and
 requirements. Can you give some details on that?

 ** **

 Best Regards,

 Cem

 ** **

 On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) pru...@avaya.com
 wrote:

 Does anyone have opinions on the maximum amount of data reasonable to
 store on one Cassandra node?  If there are limitations, what are the
 reasons for it?

  

 Thanks,

 Anne

 ** **



RE: maximum storage per node

2013-07-25 Thread Pruner, Anne (Anne)
I actually wrote my own compactor that deals with this problem.

Anne

From: cem [mailto:cayiro...@gmail.com]
Sent: Thursday, July 25, 2013 11:59 AM
To: user@cassandra.apache.org
Subject: Re: maximum storage per node

You will suffer from long compactions if you are planning to get rid of from 
old records by TTL.

Best Regards,
Cem.

On Thu, Jul 25, 2013 at 5:51 PM, Kanwar Sangha 
kan...@mavenir.commailto:kan...@mavenir.com wrote:
Issues with large data nodes would be -


* Nodetool repair will be impossible to run

* Your read i/o will suffer since you will almost always go to disk 
(each read will take 3 IOPS worst case)

* Boot-straping the node in case of failures will take days/weeks



From: Pruner, Anne (Anne) [mailto:pru...@avaya.commailto:pru...@avaya.com]
Sent: 25 July 2013 10:45
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: RE: maximum storage per node

We're storing fairly large files (about 1MB apiece) for a few months and then 
deleting the oldest to get more space to add new ones.  We have large 
requirements (maybe up to 100 TB), so having a 1TB limit would be unworkable.

What is the reason for the limit?  Does something fail after that?

If there are hardware issues, what's recommended?

BTW, we're using Cassandra 1.2

Anne

From: cem [mailto:cayiro...@gmail.com]
Sent: Thursday, July 25, 2013 11:41 AM
To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
Subject: Re: maximum storage per node

Between 500GB - 1TB is recommended.

But it depends also your hardware, traffic characteristics and requirements. 
Can you give some details on that?

Best Regards,
Cem

On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) 
pru...@avaya.commailto:pru...@avaya.com wrote:
Does anyone have opinions on the maximum amount of data reasonable to store on 
one Cassandra node?  If there are limitations, what are the reasons for it?

Thanks,
Anne




Re: Strange cassandra-stress results with 2.0.0 beta1

2013-07-25 Thread Andrew Cobley
So is that in Cassandra somewhere ? Any idea on how I can go about  pinpointing 
the problem to raise a JIRA issue  ?

Andy

On 25 Jul 2013, at 17:50, Radim Kolar h...@filez.com wrote:


 I'm wondering if it's a GC issue ?
 yes it is:

 1039280992 used; max is 1052770304

 most likely memory leak.




The University of Dundee is a registered Scottish Charity, No: SC015096



Re: Strange cassandra-stress results with 2.0.0 beta1

2013-07-25 Thread Radim Kolar

Dne 25.7.2013 20:03, Andrew Cobley napsal(a):

Any idea on how I can go about  pinpointing the problem to raise a JIRA issue  ?

http://www.ehow.com/how_8705297_create-java-heap-dump.html


deleting columns with CAS (2.0.0-beta2)

2013-07-25 Thread Kalpana Suraesh
Hi,

I've been playing around with CAS (compare-and-swap) in the 2.0 beta, using
the Thrift API. I can't at all figure out how to delete columns while using
CAS, however. I'm able to update/insert columns by adding them to the list
of updates that is passed as a parameter to
org.apache.cassandra.thrift.Cassandra.Client#cas(). But there seems to be
no indication as to how to delete.

I presume it can be done because conditional deletes work through CQL3.
However, I'm interested in using Hector/Java (particularly as CQL3 does not
yet allow batched conditional statements).

Any help would be great!

Thanks,
Kalpana


cassandra 1.2.5- virtual nodes (num_token) pros/cons?

2013-07-25 Thread rash aroskar
Hi,
I am upgrading my cassandra cluster from 0.8 to 1.2.5.
In cassandra 1.2.5 the 'num_token' attribute confuses me.
I understand that it distributes multiple tokens per node but I am not
clear how that is helpful for performance or load balancing. Can anyone
elaborate? has anyone used this feature  and knows its
advantages/disadvantages?

Thanks,
Rashmi


Re: NPE during compaction in compare

2013-07-25 Thread Paul Ingalls
Hey Sylvain,

I pulled the latest from the 1.2 branch and these exceptions were fixed.  The 
only issue that remains is compacting large rows which fails with an assertion 
error.

Thanks for the tip!

Paul

On Jul 25, 2013, at 12:26 AM, Sylvain Lebresne sylv...@datastax.com wrote:

 Are you sure you're using the last version of the 1.2 branch? The exceptions 
 are in RangeTombstoneList that has been introduced by the patch of 
 CASSANDRA-5677 and the initial version of the patch were indeed able to 
 produce that kind of error. Could that be you're using such development 
 version of the patch?
 
 What I would suggest is to test the current artifacts for 1.2.7 at 
 https://repository.apache.org/content/repositories/orgapachecassandra-013/org/apache/cassandra/apache-cassandra/1.2.7/.
  This is what should be released as soon as the vote period is complete. If 
 you can reproduce with that version, we'd be very interested to know (you can 
 even feel free to open a ticket on JIRA).
 
 --
 Sylvain
 
 
 On Wed, Jul 24, 2013 at 10:52 PM, Paul Ingalls paulinga...@gmail.com wrote:
 Hey Chris,
 
 so I just tried dropping all my data and converting my column families to use 
 leveled compaction.  Now I'm getting exceptions like the following once I 
 start inserting data.  Have you seen these?
 
 
 
 ERROR 13:13:25,616 Exception in thread Thread[CompactionExecutor:34,1,main]
 java.lang.NullPointerException
   at 
 org.apache.cassandra.db.marshal.AbstractCompositeType.compare(AbstractCompositeType.java:69)
   at 
 org.apache.cassandra.db.marshal.AbstractCompositeType.compare(AbstractCompositeType.java:31)
   at 
 org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:396)
   at 
 org.apache.cassandra.db.RangeTombstoneList.addAll(RangeTombstoneList.java:205)
   at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:180)
   at 
 org.apache.cassandra.db.AbstractThreadUnsafeSortedColumns.delete(AbstractThreadUnsafeSortedColumns.java:40)
   at 
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:51)
   at 
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:46)
   at 
 org.apache.cassandra.db.compaction.PrecompactedRow.merge(PrecompactedRow.java:115)
   at 
 org.apache.cassandra.db.compaction.PrecompactedRow.init(PrecompactedRow.java:98)
   at 
 org.apache.cassandra.db.compaction.CompactionController.getCompactedRow(CompactionController.java:160)
   at 
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:76)
   at 
 org.apache.cassandra.db.compaction.CompactionIterable$Reducer.getReduced(CompactionIterable.java:57)
   at 
 org.apache.cassandra.utils.MergeIterator$ManyToOne.consume(MergeIterator.java:114)
   at 
 org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:97)
   at 
 com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
   at 
 com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
   at 
 org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:145)
   at 
 org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48)
   at 
 org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
   at 
 org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:58)
   at 
 org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:60)
   at 
 org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:211)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439)
   at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
   at java.util.concurrent.FutureTask.run(FutureTask.java:138)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
   at java.lang.Thread.run(Thread.java:680)
  
 and
 
 ERROR 13:17:11,327 Exception in thread Thread[CompactionExecutor:45,1,main]
 java.lang.ArrayIndexOutOfBoundsException: 2
   at 
 org.apache.cassandra.db.RangeTombstoneList.insertFrom(RangeTombstoneList.java:396)
   at 
 org.apache.cassandra.db.RangeTombstoneList.addAll(RangeTombstoneList.java:205)
   at org.apache.cassandra.db.DeletionInfo.add(DeletionInfo.java:180)
   at 
 org.apache.cassandra.db.AbstractThreadUnsafeSortedColumns.delete(AbstractThreadUnsafeSortedColumns.java:40)
   at 
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:51)
   at 
 org.apache.cassandra.db.AbstractColumnContainer.delete(AbstractColumnContainer.java:46)
   at 
 

Re: cassandra 1.2.5- virtual nodes (num_token) pros/cons?

2013-07-25 Thread Sávio Teles
It is very useful to upgrade the apps perfomance.
For example, if you have a machine with X capacity, you can put the
num_token=256. If you add a machine in your cluster with (X*2) capacity you
can put the num_token=512.
So, this new machine will receive twice the load automatically.
Moreover, you can minimize how long a machine replacement is going to take
and will minimize the time to repair a disk failure.

Read the docs to more informations:
http://www.datastax.com/dev/blog/virtual-nodes-in-cassandra-1-2

Cheers!



2013/7/25 rash aroskar rashmi.aros...@gmail.com

 Hi,
 I am upgrading my cassandra cluster from 0.8 to 1.2.5.
 In cassandra 1.2.5 the 'num_token' attribute confuses me.
 I understand that it distributes multiple tokens per node but I am not
 clear how that is helpful for performance or load balancing. Can anyone
 elaborate? has anyone used this feature  and knows its
 advantages/disadvantages?

 Thanks,
 Rashmi




-- 
Atenciosamente,
Sávio S. Teles de Oliveira
voice: +55 62 9136 6996
http://br.linkedin.com/in/savioteles
Mestrando em Ciências da Computação - UFG
Arquiteto de Software
Laboratory for Ubiquitous and Pervasive Applications (LUPA) - UFG


Re: unable to compact large rows

2013-07-25 Thread Paul Ingalls
I ran with the latest from the 1.2 branch, and although the NPE and index out 
of bounds exception went away, these still stuck around.

I'll see if I can figure out something to put into JIRA...

On Jul 25, 2013, at 12:32 AM, Sylvain Lebresne sylv...@datastax.com wrote:

 Obvisously, that should not happen. That being said, we did fix a bug 
 yesterday that could produce that kind of trace: 
 https://issues.apache.org/jira/browse/CASSANDRA-5799. It will be part of 
 1.2.7 that should be released tomorrow, though the artifacts that are being 
 voted on are at 
 https://repository.apache.org/content/repositories/orgapachecassandra-013/org/apache/cassandra/apache-cassandra/1.2.7/
  if you want to test sooner than later.
 
 If you still can reproduce with that version, feel free to open a ticket on 
 JIRA (in which case, the more information you can give us to reproduce, the 
 better the change we'd be able to fix it quickly).
 
 --
 Sylvain
 
 
 On Wed, Jul 24, 2013 at 10:54 PM, Paul Ingalls paulinga...@gmail.com wrote:
 It is pretty much every row that hits the large threshold.  I don't think I 
 can delete every row that hits that…
 
 you can see the db size in the stack trace, do you want a different type of 
 size?
 
 On Jul 24, 2013, at 11:07 AM, Jason Wee peich...@gmail.com wrote:
 
 Would it possible to delete this row and reinsert this row? By the way, how 
 large is that one row?
 
 Jason
 
 
 On Wed, Jul 24, 2013 at 9:23 AM, Paul Ingalls paulinga...@gmail.com wrote:
 I'm getting constant exceptions during compaction of large rows.  In fact, I 
 have not seen one work, even starting from an empty DB.  As soon as I start 
 pushing in data, when a row hits the large threshold, it fails compaction 
 with this type of stack trace:
 
  INFO [CompactionExecutor:6] 2013-07-24 01:17:53,592 
 CompactionController.java (line 156) Compacting large row 
 fanzo/tweets_by_id:352567939972603904 (153360688 bytes) incrementally
 ERROR [CompactionExecutor:6] 2013-07-24 01:18:12,496 CassandraDaemon.java 
 (line 192) Exception in thread Thread[CompactionExecutor:6,1,main]
 java.lang.AssertionError: incorrect row data size 5722610 written to 
 /mnt/datadrive/lib/cassandra/data/fanzo/tweets_by_id/fanzo-tweets_by_id-tmp-ic-1453-Data.db;
  correct is 5767384
 at 
 org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:162)
 at 
 org.apache.cassandra.db.compaction.CompactionTask.runWith(CompactionTask.java:162)
 at 
 org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48)
 at 
 org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
 at 
 org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:58)
 at 
 org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:60)
 at 
 org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionTask.run(CompactionManager.java:211)
 at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
 at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
 at java.util.concurrent.FutureTask.run(FutureTask.java:166)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:724)
 
 I'm not sure what to do or where to look.  Help…:)
 
 Thanks,
 
 Paul
 
 
 
 
 



Cassandra 2.0: Paxos Prepare response always false

2013-07-25 Thread Soumava Ghosh
Hi,

I have test setup where clients randomly make a controlled number of cas()
requests (among other requests) at a cluster of cassandra 2.0 servers.
After one point, I'm seeing that all requests are pending and my client's
throughput has reduced to 0.0 for all kinds of requests. For this specific
case I had 10 clients each making around 30 cas() requests per second at a
cluster of 72 instances of cassandra.

Clients are set up to register a request as a success after the cas() call
returns with CASResult.success = true, else an exception is thrown. Since I
see that no client requests were actually registered and no exceptions were
thrown, which indicates that the cas() call itself is hung.

On the server side, I see Paxos logs as follows - they go on for 50 log
files for each of the servers involved, and they span at least an hour. I
have marked a particular instance where the prepare response is true but
the propose response is false from all the involved servers:

*At the Paxos Initiator: * None of the files among the 50 system logs have
the phrase 'Propose response true', these logs just go on and on.
*
*
DEBUG [RequestResponseStage:110] 2013-07-25 15:09:05,332
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d2093120-f576-11e2-a57e-a154d605509d, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.195
*
*
DEBUG [RequestResponseStage:92] 2013-07-25 15:09:05,346
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d2093120-f576-11e2-a57e-a154d605509d, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.184

DEBUG [RequestResponseStage:98] 2013-07-25 15:09:05,347
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d2093120-f576-11e2-a57e-a154d605509d, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.20

DEBUG [RequestResponseStage:93] 2013-07-25 15:09:05,350
ProposeCallback.java (line 44) Propose response false from /17.163.7.20
DEBUG [RequestResponseStage:100] 2013-07-25 15:09:05,350
ProposeCallback.java (line 44) Propose response false from /17.163.7.184
DEBUG [RequestResponseStage:111] 2013-07-25 15:09:05,350
ProposeCallback.java (line 44) Propose response false from /17.163.7.195

DEBUG [RequestResponseStage:102] 2013-07-25 15:09:05,351
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d20c3e60-f576-11e2-9bbe-bf2ad4fe6707, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.195

DEBUG [RequestResponseStage:107] 2013-07-25 15:09:05,352
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d20c3e60-f576-11e2-9bbe-bf2ad4fe6707, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.20

DEBUG [RequestResponseStage:108] 2013-07-25 15:09:05,352
PrepareCallback.java (line 58) Prepare response PrepareResponse(true,
Commit(d145fe46f5d02a54b5ea95852f94c402,
1a0c4220-f561-11e2-a409-019f62d610d7, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])),
Commit(d145fe46f5d02a54b5ea95852f94c402,
d20c3e60-f576-11e2-9bbe-bf2ad4fe6707, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,]))) from /
17.163.7.184

DEBUG [RequestResponseStage:104] 2013-07-25 15:09:05,352
ProposeCallback.java (line 44) Propose response false from /17.163.7.20
DEBUG [RequestResponseStage:99] 2013-07-25 15:09:05,353
ProposeCallback.java (line 44) Propose response false from /17.163.7.195
DEBUG [RequestResponseStage:105] 2013-07-25 15:09:05,353
ProposeCallback.java (line 44) Propose response false from /17.163.7.184

*At 17.163.7.20:*
*
*
DEBUG [MutationStage:58] 2013-07-25 15:09:05,347 PaxosState.java (line 100)
accept requested for Commit(d145fe46f5d02a54b5ea95852f94c402,
d20b05e0-f576-11e2-9bbe-bf2ad4fe6707, ColumnFamily(P
[81d271b2125c59cb0003:false:27@1374780817218000,])) but
inProgress is 

Re: cassandra 1.2.6 - Start key's token sorts after end token

2013-07-25 Thread Marcelo Elias Del Valle
We managed to make it work with RandomPartitioner... Guess we will rely on
it for now...


2013/7/23 Hiller, Dean dean.hil...@nrel.gov

 Oh, and in the past 0.20.x has been pretty stable by the wayŠ..they
 finally switched their numbering scheme thank god.

 Dean

 On 7/23/13 2:13 PM, Hiller, Dean dean.hil...@nrel.gov wrote:

 Perhaps try 0.20.2 as
 
  1.  The maven pom files have cassandra depending on 0.20.2
  2.  The 0.20.2 default was murmur and we had to change it to random
 partitioner or it wouldn't work for us
 
 Ie. I suspect they will change the pom file to a more recent version of
 hadoop at some point but I wonder if test suites suck in 0.20.2 because
 the pom file points to that versionŠ.depends on if they actually have
 tests for map/reduce which is probably a bit hard.
 
 Dean
 
 From: Marcelo Elias Del Valle
 mvall...@gmail.commailto:mvall...@gmail.com
 Reply-To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
 user@cassandra.apache.orgmailto:user@cassandra.apache.org
 Date: Tuesday, July 23, 2013 1:54 PM
 To: user@cassandra.apache.orgmailto:user@cassandra.apache.org
 user@cassandra.apache.orgmailto:user@cassandra.apache.org
 Subject: Re: cassandra 1.2.6 - Start key's token sorts after end token
 
 Dean,
 
 I am using hadoop 1.0.3.
 Indeed, using Cassandra 1.2.3 with Random partitioner, it worked.
 However, it's the only reason for me to use randompartitioner, I really
 would like to move forward. Besides, I tried to use Cassandra 1.2.6 with
 RandomPartitioner and I got problems when inserting data, even stopping
 Cassandra, cleaning my entire data folder and then starting it again.
 I am also really curious to know if there is anyone else having these
 problems or if it is just me...
 
 Best regards,
 Marcelo.
 
 
 2013/7/23 Hiller, Dean dean.hil...@nrel.govmailto:dean.hil...@nrel.gov
 
 Out of curiosity, what version of hadoop are you using with cassandra?  I
 think we are trying 0.20.2 if I remember(I have to ask my guy working on
 it to be sure).  I do remember him saying the cassandra maven dependency
 was odd in that it is in the older version and not a newer hadoop version.
 
 We are using RandomPartitioner though right now which I have personally
 used in the past with success.  We are in the process of map/reducing to
 a cassandra with MurmurPartitioner  (our real reason to map/reduce is
 some refactorings in our model though and we just thought we would switch
 to murmur).
 
 Has anyone else used map/reduce with murmur partitioner?
 
 Dean
 
 From: Marcelo Elias Del Valle
 mvall...@gmail.commailto:mvall...@gmail.commailto:mvall...@gmail.com
 m
 ailto:mvall...@gmail.com
 Reply-To:
 user@cassandra.apache.orgmailto:user@cassandra.apache.orgmailto:
 user@c
 assandra.apache.orgmailto:user@cassandra.apache.org
 user@cassandra.apache.orgmailto:user@cassandra.apache.orgmailto:
 user@c
 assandra.apache.orgmailto:user@cassandra.apache.org
 Date: Monday, July 22, 2013 4:04 PM
 To:
 user@cassandra.apache.orgmailto:user@cassandra.apache.orgmailto:
 user@c
 assandra.apache.orgmailto:user@cassandra.apache.org
 user@cassandra.apache.orgmailto:user@cassandra.apache.orgmailto:
 user@c
 assandra.apache.orgmailto:user@cassandra.apache.org
 Subject: cassandra 1.2.6 - Start key's token sorts after end token
 
 Hello,
 
 I am trying to figure what might be cause this error. I am using
 Cassandra 1.2.6 (tried with 1.2.3 as well) and I am trying to read data
 from cassandra on hadoop using column family input format. I also got the
 same error using pure astyanax on a test.
 I am using Murmur3Partitioner and I created the keyspace using
 Cassandra 1.2.6, there is nothing from prior versions. I created the
 keyspace with SimpleStrategy and replication factor 1.
 Here is the exception I am getting:
 2013-07-22 21:53:05,824 WARN org.apache.hadoop.mapred.Child (main): Error
 running child
 java.lang.RuntimeException: InvalidRequestException(why:Start key's token
 sorts after end token)
 at
 org.apache.cassandra.hadoop.ColumnFamilyRecordReader$WideRowIterator.maybe
 Init(ColumnFamilyRecordReader.java:453)
 at
 org.apache.cassandra.hadoop.ColumnFamilyRecordReader$WideRowIterator.compu
 teNext(ColumnFamilyRecordReader.java:459)
 at
 org.apache.cassandra.hadoop.ColumnFamilyRecordReader$WideRowIterator.compu
 teNext(ColumnFamilyRecordReader.java:406)
 at
 com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterat
 or.java:143)
 at
 com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:1
 38)
 at
 org.apache.cassandra.hadoop.ColumnFamilyRecordReader.getProgress(ColumnFam
 ilyRecordReader.java:103)
 at
 org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.getProgress(MapTa
 sk.java:522)
 at
 org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapT
 ask.java:547)
 at org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67)
 at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
 at 

Re: maximum storage per node

2013-07-25 Thread sankalp kohli
Try putting multiple instances per machine with each instance mapped to its
own disk. This might not work with v-nodes


On Thu, Jul 25, 2013 at 9:04 AM, Pruner, Anne (Anne) pru...@avaya.comwrote:

  I actually wrote my own compactor that deals with this problem.

 ** **

 Anne

 ** **

 *From:* cem [mailto:cayiro...@gmail.com]
 *Sent:* Thursday, July 25, 2013 11:59 AM

 *To:* user@cassandra.apache.org
 *Subject:* Re: maximum storage per node

 ** **

 You will suffer from long compactions if you are planning to get rid of
 from old records by TTL.

 ** **

 Best Regards,

 Cem.

 ** **

 On Thu, Jul 25, 2013 at 5:51 PM, Kanwar Sangha kan...@mavenir.com wrote:
 

 Issues with large data nodes would be –

  

 · Nodetool repair will be impossible to run

 · Your read i/o will suffer since you will almost always go to
 disk (each read will take 3 IOPS worst case)

 · Boot-straping the node in case of failures will take days/weeks*
 ***

  

  

 *From:* Pruner, Anne (Anne) [mailto:pru...@avaya.com]
 *Sent:* 25 July 2013 10:45
 *To:* user@cassandra.apache.org
 *Subject:* RE: maximum storage per node

  

 We’re storing fairly large files (about 1MB apiece) for a few months and
 then deleting the oldest to get more space to add new ones.  We have large
 requirements (maybe up to 100 TB), so having a 1TB limit would be
 unworkable.

  

 What is the reason for the limit?  Does something fail after that?

  

 If there are hardware issues, what’s recommended?

  

 BTW, we’re using Cassandra 1.2

  

 Anne

  

 *From:* cem [mailto:cayiro...@gmail.com cayiro...@gmail.com]
 *Sent:* Thursday, July 25, 2013 11:41 AM
 *To:* user@cassandra.apache.org
 *Subject:* Re: maximum storage per node

  

 Between 500GB - 1TB is recommended. 

  

 But it depends also your hardware, traffic characteristics and
 requirements. Can you give some details on that?

  

 Best Regards,

 Cem

  

 On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) pru...@avaya.com
 wrote:

 Does anyone have opinions on the maximum amount of data reasonable to
 store on one Cassandra node?  If there are limitations, what are the
 reasons for it?

  

 Thanks,

 Anne

  

 ** **