Re: vnodes on aws

2013-12-05 Thread Jayadev Jayaraman
Availability zones are analogous to racks not data centres . EC2 regions are equivalent to data centres. You can use vnodes if you want to deploy a cluster across multiple regions (data centres) with one availability zone per region. Each region maintains a separate ring. I don't know if you can

Re: Cassandra remove column using thrift

2013-11-01 Thread Jayadev Jayaraman
Hey guys, False alarm, sorry about that. Our column-names are byte-concatenations of short integers and we had been constructing the column names wrongly before attempting a delete. We fixed the problem and we've been able to delete the columns without issue. On Fri, Nov 1, 2013 at 4:19 PM, Robe

Re: what does nodetool compact command do for leveled compactions?

2013-10-24 Thread Jayadev Jayaraman
Thanks for the reply. I have one more question. If multiple columns with identical names but with different timestamps are bulk loaded (with sstableloader) into a CF, and we had LCS running in the background, would a slice predicate query retrieve multiple columns with the same name assuming compac

Re: Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter

2013-09-26 Thread Jayadev Jayaraman
es not have a TTL. > > Cheers > > > - > Aaron Morton > New Zealand > @aaronmorton > > Co-Founder & Principal Consultant > Apache Cassandra Consulting > http://www.thelastpickle.com > > On 26/09/2013, at 12:44 AM, Jayadev Jayaraman wrote: >

Re: Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter

2013-09-25 Thread Jayadev Jayaraman
the column will expire as required) ? If not , what is the TTL attribute used for in the Column object ? """ Thanks, Jayadev On Tue, Sep 24, 2013 at 2:48 PM, Jayadev Jayaraman wrote: > Let's say I've initialized a *SSTableSimpleWriter* instance and a new > column w

Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter

2013-09-24 Thread Jayadev Jayaraman
Let's say I've initialized a *SSTableSimpleWriter* instance and a new column with TTL set : *SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here */);* *Column column;* What is the difference between calling *writer.addColumn()* on the column's name and value, and *writer.addE

Re: Cassandra 1.2.9 cluster with vnodes is heavily unbalanced.

2013-09-20 Thread Jayadev Jayaraman
curs. ( it shouldn't be the fault of the Murmur3 partitioner which guarantees a uniform distribution of keys across token-ranges according to the doc. ) On Thu, Sep 19, 2013 at 8:56 PM, Jayadev Jayaraman wrote: > We're using NetworkTopologyStrategy for our placement_strategy and our > prim

Re: Cassandra 1.2.9 cluster with vnodes is heavily unbalanced.

2013-09-19 Thread Jayadev Jayaraman
We ran nodetool repair on all nodes for all Keyspaces / CFs, restarted cassandra and this is what we get for nodetool status : bin/nodetool -h localhost status Datacenter: us-east === Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns

Re: What is the ideal value for sstable_size_in_mb when using LeveledCompactionStrategy ?

2013-09-18 Thread Jayadev Jayaraman
es to stay under 1T per node or you run into big > troubles and most people stay under 500G per node. > > Later, > Dean > > From: Jayadev Jayaraman mailto:jdisal...@gmail.com>> > Reply-To: "user@cassandra.apache.org<mailto:user@cassandra.apache.org>"

Re: What is the ideal value for sstable_size_in_mb when using LeveledCompactionStrategy ?

2013-09-18 Thread Jayadev Jayaraman
th our > 300T per node. > > Dean > > From: Jayadev Jayaraman mailto:jdisal...@gmail.com>> > Reply-To: "user@cassandra.apache.org<mailto:user@cassandra.apache.org>" < > user@cassandra.apache.org<mailto:user@cassandra.apache.org>> > Date: Wednesd

What is the ideal value for sstable_size_in_mb when using LeveledCompactionStrategy ?

2013-09-18 Thread Jayadev Jayaraman
We have set up a 24 node (m1.xlarge nodes, 1.7 TB per node) cassandra cluster on Amazon EC2 : version=1.2.9 replication factor = 2 snitch=EC2Snitch placement_strategy=NetworkTopologyStrategy (with 12 nodes each in 2 availability zones) Background on our use-case : We plan on using hadoop with ss