Re: memtable mem usage off by 10?

2014-06-11 Thread Idrén , Johan
Sorry for the slow reply, here’s the output: java version "1.7.0_55" Java(TM) SE Runtime Environment (build 1.7.0_55-b13) Java HotSpot(TM) 64-Bit Server VM (build 24.55-b03, mixed mode) What behaviour should I expect to see if I started using 2.1.x instead? Should I see a closer correlation of r

Re: Migration 1.2.14 to 2.0.8 causes "Tried to create duplicate hard link" at startup

2014-06-11 Thread Tom van den Berge
No, unfortunately I haven't. On Tue, Jun 10, 2014 at 5:35 PM, Chris Burroughs wrote: > Were you able to solve or work around this problem? > > > On 06/05/2014 11:47 AM, Tom van den Berge wrote: > >> Hi, >> >> I'm trying to migrate a development cluster from 1.2.14 to 2.0.8. When >> starting u

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread DuyHai Doan
Hello Or Sher, The behavior is quite normal: 1) insert into test_table (p1,p2,c1,d1,d2) values ('a','b','c','d','e'); --> Insert 5 columns without any TTL 2) update test_table using ttl 10 set d1='---', d2='---' where p1='a' and p2='b' and c1='c'; --> Re-insert columns d1 and d2 with new value

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread Or Sher
Hi, Does that mean internally there is a TTL for an entire CQL row?? I thought ttl are only attached to CQL row values (Columns which are not in the PK). I thought when all values of a row are deleted, it should mean that that row does not exists. Please correct me where I'm wrong. On Wed, Jun

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread DuyHai Doan
Yes, the TTL is also set on an internal row marker. More details on this here: https://issues.apache.org/jira/browse/CASSANDRA-6668 On Wed, Jun 11, 2014 at 10:38 AM, Or Sher wrote: > Hi, > > Does that mean internally there is a TTL for an entire CQL row?? > I thought ttl are only attached to CQ

RPC timeout paging secondary index query results

2014-06-11 Thread Phil Luckhurst
Is paging through the results of a secondary index query broken in Cassandra 2.0.7 or are we doing something wrong? We have table with a few hundred thousand records and an indexed low-cardinality column. The relevant bits of the table definition are shown below CREATE TABLE measurement ( measure

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread Or Sher
Thanks DuyHai, Didn't had an idea it works like that is CQL. So per my understanding, the only way to set a ttl on an entire existing CQL row is to Insert the row again with the same values. Is that correct? Wouldn't it be simpler if Cassandra just let us change the ttl on the row marker? On Wed

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread DuyHai Doan
The only way to set a ttl on an entire existing CQL row is to Insert the row again with the same values. Is that correct? -> What do you mean by "entire existing CQL3 row" ? Do you mean setting TTL on every column of this row ? If so, the answer is yes, you need either to use INSERT or UPDATE by sp

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread Or Sher
That's what I mean, but you stated "you need either to use INSERT or UPDATE by specifying all the columns in your statement". How can I use all columns in an update statement? Primary key parts can't be set.. "This is internal impl details, not supposed to be exposed as public API" Of course it's

Re: Setting TTL to entire row: UPDATE vs INSERT

2014-06-11 Thread DuyHai Doan
Yes, you're right for update, only non primary key columns. For the idea of exposing row marker to public API, you can create a JIRA to make suggestion. On Wed, Jun 11, 2014 at 3:36 PM, Or Sher wrote: > That's what I mean, but you stated "you need either to use INSERT or > UPDATE by specifying

Re: Large number of row keys in query kills cluster

2014-06-11 Thread Jeremy Jongsma
I'm using Astyanax with a query like this: clusterContext .getClient() .getKeyspace("instruments") .prepareQuery(INSTRUMENTS_CF) .setConsistencyLevel(ConsistencyLevel.CL_LOCAL_QUORUM) .getKeySlice(new String[] { "ROW1", "ROW2", // 20,000 keys here... "ROW2" }) .ex

Re: VPC AWS

2014-06-11 Thread Peter Sanford
Tinc's developers acknowledge that there are some fairly serious unfixed security issues in their protocol: http://www.tinc-vpn.org/security/. As such, I do not consider tinc to be a good choice for production systems. Either IPSec or OpenVPN are reasonable for connecting VPCs in different regions

ANN: Cassandra monitoring & anomaly detection

2014-06-11 Thread Otis Gospodnetic
Hi, We've added Cassandra performance monitoring and anomaly detection to SPM recently. Here's the announcement with several screenshots: http://blog.sematext.com/2014/06/02/announcement-cassandra-performance-monitoring-in-spm/ Storm, Kafka, Elasticsearch, Hadoop... m

Re: Large number of row keys in query kills cluster

2014-06-11 Thread Jeremy Jongsma
The big problem seems to have been requesting a large number of row keys combined with a large number of named columns in a query. 20K rows with 20K columns destroyed my cluster. Splitting it into slices of 100 sequential queries fixed the performance issue. When updating 20K rows at a time, I saw

Re: Large number of row keys in query kills cluster

2014-06-11 Thread Robert Coli
On Wed, Jun 11, 2014 at 10:12 AM, Jeremy Jongsma wrote: > Is there any documentation on this? Obviously these limits will vary by > cluster capacity, but for new users it would be great to know that you can > run into problems with large queries, and how they present themselves when > you hit the

Re: RPC timeout paging secondary index query results

2014-06-11 Thread Robert Coli
On Wed, Jun 11, 2014 at 2:24 AM, Phil Luckhurst < phil.luckhu...@powerassure.com> wrote: > Is paging through the results of a secondary index query broken in > Cassandra > 2.0.7 or are we doing something wrong? > General feedback on questions of this type : http://mail-archives.apache.org/mod_mb

Re: RPC timeout paging secondary index query results

2014-06-11 Thread DuyHai Doan
I like the "- Provides the illusion that you are using a RDBMS." part ;-) On Wed, Jun 11, 2014 at 8:52 PM, Robert Coli wrote: > On Wed, Jun 11, 2014 at 2:24 AM, Phil Luckhurst < > phil.luckhu...@powerassure.com> wrote: > >> Is paging through the results of a secondary index query broken in >> C

Hadoop MultipleOutputs with CqlOutputFormat

2014-06-11 Thread Deepak Shetty
Hi In Hadoop the MultipleOutputs class can be used to write to more than one destination - However it appears you cannot use it to write to two Cassandra tables using CQLOutputFormat This seems to have been working previously with ColumnFamilyOutputFormat (https://gist.github.com/rstrickland/37637

Re: RPC timeout paging secondary index query results

2014-06-11 Thread Phil Luckhurst
Thanks Rob. I understand that we will probably end up either creating our own index or duplicating the data and we have done that to remove a reliance on secondary indexes in other places. It just seems that what we are trying to do here is such basic functionality of an index that I thought we mu

Re: RPC timeout paging secondary index query results

2014-06-11 Thread Robert Coli
On Wed, Jun 11, 2014 at 12:43 PM, Phil Luckhurst < phil.luckhu...@powerassure.com> wrote: > It just seems that what we are trying to do here is > such basic functionality of an index that I thought we must be doing > something wrong for it to appear to be this broken. > To be clear, I did not rea

Re: Large number of row keys in query kills cluster

2014-06-11 Thread Peter Sanford
On Wed, Jun 11, 2014 at 10:12 AM, Jeremy Jongsma wrote: > The big problem seems to have been requesting a large number of row keys > combined with a large number of named columns in a query. 20K rows with 20K > columns destroyed my cluster. Splitting it into slices of 100 sequential > queries fix

embedding cassandra in junit ..

2014-06-11 Thread Kevin Burton
Maybe this is an anti-pattern.. please feel free to flame me if that's the case ;) I'd like to embed cassandra in junit so that I can test my database code from my laptop and also get it working in our integration. The general idea is to start with a clean slate, run cassandra, create tables, sim

Re: embedding cassandra in junit ..

2014-06-11 Thread James Carman
Take a look at cassandra-unit. https://github.com/savoirtech/hecate/blob/master/hecate-cql3/src/test/java/com/savoirtech/hecate/cql3/test/CassandraTestCase.java On Wed, Jun 11, 2014 at 11:02 PM, Kevin Burton wrote: > Maybe this is an anti-pattern.. please feel free to flame me if that's the >

Re: embedding cassandra in junit ..

2014-06-11 Thread Johan Edstrom
Cassandra-unit 2.0X works awesomely, if you are willing to spend the slightly few more cycles, - Look at farsandra. :) I copied Farsandra (since there was no release) into hecate. On Jun 11, 2014, at 11:02 PM, Kevin Burton wrote: > Maybe this is an anti-pattern.. please feel free to flame me

How do I listen for new queues created on apollo…

2014-06-11 Thread Kevin Burton
I am trying to listen for advisory messages on apollo and for the life of me I can't figure it out… I want to find out when new queues are being created. I need to get all queues and then new queues as they are created. Is there an example or piece of code I could start with? I was banging my h

Re: Large number of row keys in query kills cluster

2014-06-11 Thread Jack Krupansky
Hmmm... that multipl-gets section is not present in the 2.0 doc: http://www.datastax.com/documentation/cassandra/2.0/cassandra/architecture/architecturePlanningAntiPatterns_c.html Was that intentional – is that anti-pattern no longer relevant to C* 2.0? Matt’s slideshare refers to “unbounded batc