as I understand, the read repair is a background task triggered by the read
request, but once the consistency requirement has been met you will be given
a response.
the coordinator at CL.ONE is allowed to return your responce once it has one
response (empty or not) from any replica. if the first
also there is a configuration parameter that controls the probability of any
read request triggering a read repair
- Stephen
---
Sent from my Android phone, so random spelling mistakes, random nonsense
words and other nonsense are a direct result of using swype to type on the
screen
On 7 Apr
Hi,
I am using the thrift client batch_mutate method with Cassandra 0.7.0 on
Ubuntu 10.10.
When the size of the mutations gets too large, the client fails with the
following exception:
Caused by: org.apache.thrift.transport.TTransportException:
java.net.SocketException: Connection reset
at
that makes sense. thanks!
On Apr 7, 2011, at 8:36 AM, Stephen Connolly wrote:
also there is a configuration parameter that controls the probability of any
read request triggering a read repair
- Stephen
---
Sent from my Android phone, so random spelling mistakes, random nonsense
El mié, 06-04-2011 a las 21:04 -0500, Jonathan Ellis escribió:
out of sequence response is thrift's way of saying I got a response
for request Y when I expected request X.
my money is on using a single connection from multiple threads. don't do
that.
I'm not using thrift directly, and
El mié, 06-04-2011 a las 21:04 -0500, Jonathan Ellis escribió:
out of sequence response is thrift's way of saying I got a response
for request Y when I expected request X.
my money is on using a single connection from multiple threads. don't do
that.
I'm not using thrift directly, and
Thank both of you. I'll look it up from another view including
non-cassandra processes on the servers. Just guessing Nagios or
iptables or others causes it.
Kazuo
(11/04/05 22:55), Sasha Dolgy wrote:
I've been seeing this EOF in my system.log file occasionally as well.
Doesn't seem to be
I would look at your client and see if it can handle multiple pools. So it will
try to connect in the first pool, if that fails retry the other nodes in that
pool and then move to the next pool. I think hector has grabbed some of the
performance detection features from the DynamicSnitch so may
2011/4/7 Jonathan Ellis jbel...@gmail.com
Hypothesis: it's probably the flush causing the CMS, not the snapshot
linking.
Confirmation possibility #1: Add a logger.warn to
CLibrary.createHardLinkWithExec -- with JNA enabled it shouldn't be
called, but let's rule it out.
Confirmation
Hi all,
dead nodes which I removed via nodetool's removetoken are reappearing in the
ring after few days.
Is this normal? Is there a way how to remove them for good?
Thanks,
Viliam
Nice explanation.
Wanted to add the importance of been the first node in the ordered node list,
even for CL ONE, is that this is the node sent the data request and it has to
return before the CL is considered satisfied. e.g. CL One with RR running, read
sent to all 5 replicas, if 3 digest
Peter, thank you for the extremely detailed reply.
To now answer my own question, the critical points that are different
from what I said earlier are: that CL.ONE does prefer *one* node (which
one depending on snitch) and that RR uses digests (which are not
mentioned on the wiki page [1]) instead
Can you turn the logging up to DEBUG level and look for a message from
CassandraServer that says ... timed out ?
Also check the thread pool stats nodetool tpstats to see if the node is
keeping up.
Aaron
On 7 Apr 2011, at 13:43, Sheng Chen wrote:
Thank you Aaron.
It does not seem to be
Sounds like this
http://comments.gmane.org/gmane.comp.db.cassandra.user/14498
https://issues.apache.org/jira/browse/CASSANDRA-2371
Aaron
On 7 Apr 2011, at 22:36, Viliam Holub wrote:
Hi all,
dead nodes which I removed via nodetool's removetoken are reappearing in the
ring after few days.
These types of exceptions is seen sporadically in our cassandra logs. They
occur especially after running a repair with the nodetool.
I assume there are a few corrupt rows. Is this cause for panic?
Will a repair fix this, or is it best to do a decomission + bootstrap via a
move for
No, 2252 is not suitable for backporting to 0.7.
On Thu, Apr 7, 2011 at 7:33 AM, ruslan usifov ruslan.usi...@gmail.com wrote:
2011/4/7 Jonathan Ellis jbel...@gmail.com
Hypothesis: it's probably the flush causing the CMS, not the snapshot
linking.
Confirmation possibility #1: Add a
For anyone using Codeigniter and interested. I've written a little library
to integrate
Codeigniter with PHPcassa and consequently Cassandra.
It provides you with access to code igniter's $this-db instance that only
has
the library's methods and phpcassa's.
Follow up tutorial
Creating an index, validator, and default validator then renaming/dropping the
index later results in read errors.
Is there an easy way around this problem without having to keep an invalid
definition for a column that will get deleted or expired?
1) create a secondary index on a column with
On Wed, Apr 6, 2011 at 11:49 PM, Ross Black ross.w.bl...@gmail.com wrote:
Hi,
I am using the thrift client batch_mutate method with Cassandra 0.7.0 on
Ubuntu 10.10.
When the size of the mutations gets too large, the client fails with the
following exception:
Caused by:
You could simulate it thoug. Just Add some Meta Column with a boolean Value
indicating if the referred Column is in the Row or Not. Then Add an Index in
that Meta Column and query for it.
I. E. Row a: (c=1234),(has_c=Yes)
Quert : List cf where has_c=Yes
Am 06.04.2011 um 18:52 schrieb
I'll capture what I we're seeing here for anyone else who may look
into this in more detail later.
Our standard heap growth is ~300K in between collections with regular
ParNew collections happening on average about every 4 seconds. All
very healthy.
The memtable flush (where we see almost all
Hi,
I am working on a Question/Answers web app using Cassandra(consider very
similar to StackOverflow sites). I need to built the reputation system for
users on the application. This way the user's reputation increases when s/he
answers correctly somebody's question. Thus if I keep the reputation
On Thu, Apr 7, 2011 at 2:27 PM, Erik Onnen eon...@gmail.com wrote:
1) Does this seem like a sane amount of garbage (512MB) to generate
when flushing a 64MB table to disk?
Sort of -- that's just about exactly the amount of space you'd expect
64MB of serialized data to take, in memory. (Not very
Pelops uses a single connection per operation from a pool that is backed by
Apache Commons Pool (assuming you're using Cassandra 0.7). I'm not saying it's
perfect but it's NOT sharing a connection over multiple threads.
Dan Hendry mentioned that he sees these errors. Is he also using Pelops?
Connect with jconsole and watch the memory consumption graph. Click the
force GC button watch what the low point is, that is how much memory is
being used for persistent stuff, the rest is garbage generated while
satisfying queries. Run a query, watch how the graph spikes up when you
run your
Hi All,
I have written a code for connecting to mbean server runnning on cassandra
node.
I get the following error:
Exception in thread main java.lang.reflect.UndeclaredThrowableException
at $Proxy1.getReadOperations(Unknown Source)
at
The correct object name is org.apache.cassandra.db:type=StorageProxy
-Naren
On Thu, Apr 7, 2011 at 4:36 PM, Anurag Gujral anurag.guj...@gmail.comwrote:
Hi All,
I have written a code for connecting to mbean server runnning on cassandra
node.
I get the following error:
Exception in thread
Addressed on the issue you created,
https://issues.apache.org/jira/browse/CASSANDRA-2436.
On Thu, Apr 7, 2011 at 12:19 PM, Fryar, Dexter dexter.fr...@hp.com wrote:
I have also attached the debug log with each step attached. I've even tried
going back and updating the CF with the old index to
Hello
I just upgraded a 1-node setup from rc2 to 0.7.4 and ran scrub without any
error. Now 'list CF' in CLI does not return any data as followings:
list User;
Using default limit of 100
Input length = 1
I don't see any errors or exceptions in the log.
If I run CLi from 0.7.0 against 0.7.4
29 matches
Mail list logo