I finally solved this by using nodetool move and specifying token
number explicitly. Thanks for all your help !!
Cheers,
Aishwarya
On Tue, Aug 2, 2011 at 5:31 PM, aaron morton wrote:
> initial_token is read from the yaml file once only, during bootstrap. It is
> then stored in the LocationInf
Thank you!
Will this situation work only for UTF8Type comparator?
On Wed, Aug 3, 2011 at 4:50 AM, Tyler Hobbs wrote:
> A minor correction:
>
> To get all columns starting with "ABC_", you would set column_start="ABC_"
> and column_finish="ABC`" (the '`' character comes after '_'), and ignore th
Hi,
Can the schema disagreement problem be avoided?
Thanks.
--
Dikang Gu
0086 - 18611140205
On Tue, Aug 2, 2011 at 23:10, Jeremiah Jordan
wrote:
> If you have RF=1, taking one node down is going to cause 25% of your
> data to be unavailable. If you want to tolerate a machines going down
> you need to have at least RF=2, if you want to use quorum and have a
> machine go down, you need at
Kill -9 (SIGKILL) is the worst signal to use. It has the advantage to
kill quickly the process, but no shutdown hook are called. You should
better kill -15 (SIGTERM, which is the default).
2011/7/26 mcasandra :
> I need to write cassandra start/stop script. Currently I run "cassandra" to
> start
Thanks Aaron.
I'll take a look at that. I'll refer this to the hector list as well, as I'm
interested in an OOTB solution from the Hector API itself.
On Wed, Aug 3, 2011 at 4:05 AM, Aaron Morton wrote:
> You want to run the thrift function describe_schema_versions it will tell
> you which nodes
As per subject.
Thanks.
--
Dikang Gu
0086 - 18611140205
I got this error when processing a lot operations….
2011-08-03 11:26:35,786 ERROR [com.iw.nebula.dao.simpledb.SimpleDBAdapter] -
2011-08-03 11:48:21,998 ERROR [com.iw.nebula.dao.simpledb.SimpleDBAdapter] -
I did not see anything wrong in the cassandra/system.log
What's your suggestions?
--
I followed the instructions in the FAQ, but got the following when "describe
cluster;"
Snitch: org.apache.cassandra.locator.SimpleSnitch
Partitioner: org.apache.cassandra.dht.RandomPartitioner
Schema versions:
dd73c740-bd84-11e0--98dab94442fb: [192.168.1.28, 192.168.1.9, 192.168.1.27]
UNREAC
That class got committed prematurely from an old version of 2252.
If you're interested you should look at the latest patch posted with
https://issues.apache.org/jira/browse/CASSANDRA-2252?focusedCommentId=13065060&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-130650
On Tue, 2011-08-02 at 22:46 -0500, Brandon Williams wrote:
> On Tue, Aug 2, 2011 at 10:22 PM, Bill Hastings wrote:
> > Oops. Sorry. Any information would be great.
>
> The class does not exist in trunk and appears unused in the 0.8 branch.
>
> --
> Eric Evans
Doh!
--
Eric Evans
eev...@rackspa
On Tue, Aug 2, 2011 at 10:22 PM, Bill Hastings wrote:
> Oops. Sorry. Any information would be great.
The class does not exist in trunk and appears unused in the 0.8 branch.
--
Eric Evans
Have you seen http://wiki.apache.org/cassandra/FAQ#schema_disagreement ?
On Tue, Aug 2, 2011 at 10:25 PM, Dikang Gu wrote:
> I also encounter the schema disagreement in my 0.8.1 cluster today…
>
> The disagreement occurs when I create a column family using the hector api,
> and I found the follow
Hi guys,
Now that dynamic composite have been introduced as part of cassandra
core, has any thought been given to implementing "order by" semantics to
secondary indexing? The dynamic composite allows asc/desc order sorting
of each component in the composite, allowing for mixed asc/desc order in
I also encounter the schema disagreement in my 0.8.1 cluster today…
The disagreement occurs when I create a column family using the hector api, and
I found the following errors in my cassandra/system.log
ERROR [pool-2-thread-99] 2011-08-03 11:21:18,051 Cassandra.java (line 3378)
Internal error
Oops. Sorry. Any information would be great.
Regards
Bill.
On Tue, Aug 2, 2011 at 8:20 PM, Avinash Lakshman wrote:
> Bill
>
> You have done it again :). Jonathan is the man you need to contact for this
> or as I advised the mailing list is the best. But don't sign off as me. This
> is the Nth t
Bill
You have done it again :). Jonathan is the man you need to contact for this
or as I advised the mailing list is the best. But don't sign off as me. This
is the Nth time :).
Avinash
On Tue, Aug 2, 2011 at 8:17 PM, Bill Hastings wrote:
> Is this part of trunk and used? I see the code for Me
Yes, that should work. For better/worse, CF metadata is stored in the
system schema, not individual sstables.
On Tue, Aug 2, 2011 at 10:00 PM, Todd Nine wrote:
> Hi Johnathan. Are these the correct steps for the upgrade?
>
> 1. Copy SSTable from 0.7.4 snapshot
> 2. Drop CFS in all nodes in 0.7.
Is this part of trunk and used? I see the code for MemtableAllocator but I
don't see it being used.
Avinash
Hi Johnathan. Are these the correct steps for the upgrade?
1. Copy SSTable from 0.7.4 snapshot
2. Drop CFS in all nodes in 0.7.4 cluster
3. Upgrade to 0.8.2 and start cluster (with no clients connected)
4. Create CFs using cli in cluster
5. Shutdown all nodes
6. Move the SSTables into place on al
On Wed, 2011-08-03 at 03:58 +0200, Jens Hartung wrote:
> Is there a way, to insert Strings with single quotes to Cassandra via
> CQL?
Yes, escape them with another single quote ('aa''bb').
--
Eric Evans
eev...@rackspace.com
If the bytes are compatible why not just use the existing sstable
files after redefining the schema?
On Tue, Aug 2, 2011 at 8:06 PM, Todd Nine wrote:
> Hi all,
> I seem to have backed myself into a corner and I cannot easily upgrade
> from a custom 0.7.4 installation to the default 0.8.2. Any
Hi all,
first, i'm new to Cassandra. I'm use Cassandra 0.8.2 with CQL JDBC Driver
Version 1.0.3.
My problem is, that I have to insert a large amount of weather data to
Cassandra via JAVA. Some of the weather station have single quotes in their
name.
When I insert them directly via a Statement
A minor correction:
To get all columns starting with "ABC_", you would set column_start="ABC_"
and column_finish="ABC`" (the '`' character comes after '_'), and ignore the
last column in your results if it happened to be "ABC`".
column_finish, or the "slice end" in other clients, is inclusive. Y
Hi all,
I seem to have backed myself into a corner and I cannot easily upgrade
from a custom 0.7.4 installation to the default 0.8.2. Any help I could
get would be greatly appreciated. Below is an outline of the problem.
Current installation:
0.7.4 with Ed Anuff's custom composite comparators.
Columns are reconciled using the following logic:
1) DeletedColumns always win. If there are two deletes the highest timestamp
wins
2) For non deleted columns the one with the highest timestamp wins.
3) In the case of timestamp collision the column with the highest value (by
comparing bytes) w
initial_token is read from the yaml file once only, during bootstrap. It is
then stored in the LocationInfo system CF and used from there.
It sounds like when you did the move you deleted these files, but then started
the nodes each with their own seed. So you created 3 separate clusters, when
Not sure I understand your question exactly, but will take a shot…
Writes are sent to every UP node, the consistency level is how many nodes we
require to complete before we say the request completed successfully. So we
also make sure that CL nodes are UP before we start the request. If you run
Yup, thats a pretty common pattern. How exactly depends on the client you are
using.
Say you were using pycassam, you would do a get()
http://pycassa.github.com/pycassa/api/pycassa/columnfamily.html#pycassa.columnfamily.ColumnFamily.get
with column_start="ABC_" , count to whatever, and column_
You want to run the thrift function describe_schema_versions it will tell you
which nodes are on which schemas, and wait until there is a single version.
From memory this is what the Cassandra Cli does, have a look in the code base
in the o.a.c.cli package.
Cheers
-
Aaron Mor
Try dropping the min compaction threshold via node tool to 2, that should force
every compaction bucket to get processed.
Use nodetool flush to trigger minor compaction for all cf's.
Clean will only remove load if you have moved tokens.
Cheers
-
Aaron Morton
Freelance Cassandra
I believe that what would happen is that whichever data center has the later
clock will win. Every modification you make gets a time stamp (generally set by
your client to the current time, if you are using one). I believe that whatever
modification happened with the last time stamp is canonical
Have not used Cassandra much yet, but it seems pretty clear to me that the
entry with the latest timestamp (sent in by the client) will be the one that
will be the "winner" eventually.
Mark here that this is the timestamp the client send in which is not
necessarily the actual time.
So eg if the D
I stand corrected. There are several dozen reasons to upgrade, AND that one. :)
On Tue, Aug 2, 2011 at 4:42 PM, Yiming Sun wrote:
> Hi Jonathan,
>
> Good to know. We will certainly upgrade to 0.7.8.
>
> Also, here is the link to that post I came across earlier:
>
> http://cassandra-user-incubat
I corrected the seed list and checked the cluster name. They are all
good now. But still nodetool -ring shows only one node.
INFO 21:36:59,735 Starting Messaging Service on port 7000
INFO 21:36:59,748 Using saved token 113427455640312814857969558651062452224
Nodes a_ipadrr and b_ipaddr have th
Hi Jonathan,
Good to know. We will certainly upgrade to 0.7.8.
Also, here is the link to that post I came across earlier:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Files-not-deleted-after-compaction-and-GCed-td5960453.html
best,
-- Y.
On Tue, Aug 2, 2011 at 5:36 PM, Jo
Hello,
I'm trying to save geo-data in Cassandra,
according to SimpleGeo they did that using nested tree:
http://www.readwriteweb.com/cloud/2011/02/video-simplegeo-cassandra.php
I wonder if someone already implement something like that and how he
accomplished that without transaction supports (whi
I don't remember a removing-compacted-files bug in 0.7.0, but you
should absolutely upgrade to 0.7.8 for several dozen other fixes,
including some severe ones -- see NEWS.txt.
On Tue, Aug 2, 2011 at 4:29 PM, Yiming Sun wrote:
> Hi Jeremiah,
>
> Thank you for the information - it certainly is a re
Yes.
Different cluster names could also cause this.
On Tue, Aug 2, 2011 at 4:21 PM, Jeremiah Jordan
wrote:
> All of the nodes should have the same seedlist. Don't use localhost as
> one of the items in it if you have multiple nodes.
>
> On Tue, 2011-08-02 at 10:10 -0700, Aishwarya Venkataraman
Is consistency level "All" for write actually grenty that my data is updated
in all of my node?
is it apply to read actions as-well?
I've read it on the wiki, I just want to make sure.
Thanks!
Hi Jeremiah,
Thank you for the information - it certainly is a relief. Two questions
though:
1. I came across an old thread which seemed to be saying 0.7.0 cassandra has
a bug and doesn't remove these compact files properly. Should we upgrade to
a newer version that has this bug fixed?
2. Do w
Hello,
I wonder if I can select a column or all columns that start with X.
E.g I have columns ABC_1, ABC_2, ZZZ_1 and I want to select all columns that
start with ABC_ - is that possible?
Thanks!
All of the nodes should have the same seedlist. Don't use localhost as
one of the items in it if you have multiple nodes.
On Tue, 2011-08-02 at 10:10 -0700, Aishwarya Venkataraman wrote:
> Nodetool does not show me all the nodes. Assuming I have three nodes
> A, B and C. The seedlist of A is loca
0.6.13 flushes first, and I don't remember changing that earlier in
0.6, but you'd have to check ColumnFamilyStore.snapshot in 0.6.8 to be
100% sure
On Tue, Aug 2, 2011 at 4:08 PM, Jeremiah Jordan
wrote:
> Does snapshot in 0.6 cause a flush to happen first? If not there could
> be data in the dat
Connect with jconsole and run garbage collection.
All of the files that have a -Compacted with the same name will get
deleted the next time a full garbage collection runs, or when the node
is restarted. They have already been combined into new files, the old
ones just haven't been deleted yet.
On
If you have RF=1, taking one node down is going to cause 25% of your
data to be unavailable. If you want to tolerate a machines going down
you need to have at least RF=2, if you want to use quorum and have a
machine go down, you need at least RF=3.
On Tue, 2011-08-02 at 16:22 +0200, Patrik Modest
Does snapshot in 0.6 cause a flush to happen first? If not there could
be data in the database that won't be in the snapshot. Though that
seems like a long time for data to be sitting in the commit log and not
make it to the sstables.
On Thu, 2011-07-28 at 17:30 -0500, Jonathan Ellis wrote:
> Doe
Hi,
Let’s say that I have 2 datacenters, a key is changed on both of my
datacenters in the exact same time (even in 1-2 seconds diff).
Datacenter #1 add column "abc" with value X Datacenter #2 add column "abc"
with value Y.
What is the result of that situation?
Is there any different if the change
> Very interesting. After the second host goes down do you see
> "me.prettyprint.hector.api.exceptions.HectorException: All host pools
> marked down. Retry burden pushed out to client"?
No, the last message is:
2011-08-02 08:43:06,561 INFO
[me.prettyprint.cassandra.connection.HConnectionManager]
On Tue, Aug 2, 2011 at 6:13 PM, Anthony Ikeda
wrote:
> The link (which I may be misreading)
> is http://groups.google.com/group/hector-users/browse_thread/thread/8d7004b6f85a0f2e
I hadn't found that one, but I doubt that our issue is related to that.
> It's only started happening today and happe
Hi,
I am new to Cassandra, and am hoping someone could help me understand the
(large amount of small) data files on disk that Cassandra generates.
The reason we are using Cassandra is because we are dealing with thousands
to millions of small text files on disk, so we are experimenting with
Cassa
As mentioned, there is an init.d script in the RPM package to start and stop
Cassandra (it is what we use). If you do not use the RPM and don't want to or
cannot install the full package, you can get just the script at:
https://svn.apache.org/repos/asf/cassandra/trunk/redhat/cassandra
- Ori
On Tue, Aug 2, 2011 at 10:30 AM, Adi wrote:
>> The seedlist of A is localhost.
>>
>> Seedlist of B is localhost, A_ipaddr and
>>
>> seedlist of C is localhost,B_ipaddr,A_ipaddr.
>
>
> Using localhost(or own IP address for non-seed nodes) is not a good
> practice.
> Try
>>
>> The seedlist of A :
I was thinking, it's only the one function that it has happened to so far
with the other functions unaffected. My understanding is that the hector
client pool should be shared across functions meaning that other functions
should also be affected but they are not.
Here is the connection configurati
The link (which I may be misreading) is
http://groups.google.com/group/hector-users/browse_thread/thread/8d7004b6f85a0f2e
It's only started happening today and happened on 2 occassions (8:43 and
10:21) performing the same function (querying a column family).
It seems to be trying to access a conn
I ran across a problem, when trying to execute the following code through
Hector.
private boolean createCF(String CFName) {
BasicColumnFamilyDefinition columnFamilyDefinition = new
BasicColumnFamilyDefinition();
columnFamilyDefinition.setColumnType(ColumnType.STANDARD);
col
On Tue, Aug 2, 2011 at 4:36 PM, Anthony Ikeda
wrote:
> I'm not sure if this is a problem with Hector or with Cassandra.
> We seem to be seeing broken pipe issues with our connections on the client
> side (Exception below). A bit of googling finds possibly a problem with the
> amount of data we are
>
> The seedlist of A is localhost.
Seedlist of B is localhost, A_ipaddr and
seedlist of C is localhost,B_ipaddr,A_ipaddr.
>
Using localhost(or own IP address for non-seed nodes) is not a good
practice.
Try
> The seedlist of A : A_ipaddr.
Seedlist of B : A_ipaddr
seedlist of C : A_ipaddr
Nodetool does not show me all the nodes. Assuming I have three nodes
A, B and C. The seedlist of A is localhost. Seedlist of B is
localhost, A_ipaddr and seedlist of C is localhost,B_ipaddr,A_ipaddr.
I have autobootstrap set to false for all 3 nodes since they all have
the correct data and do not
"ERROR 08:53:47,678 Internal error processing batch_mutate
>> java.lang.IllegalStateException: replication factor (3) exceeds number
>> of endpoints (1)"
>>
> You already answered
> "It always keeps showing only one node and mentions that it is handling
> 100% of the load."
>
Cluster think only on
I'm not sure if this is a problem with Hector or with Cassandra.
We seem to be seeing broken pipe issues with our connections on the client
side (Exception below). A bit of googling finds possibly a problem with the
amount of data we are trying to store, although I'm certain our datasets are
not a
> "ERROR 08:53:47,678 Internal error processing batch_mutate
> java.lang.IllegalStateException: replication factor (3) exceeds number
> of endpoints (1)"
>
You already answered
"It always keeps showing only one node and mentions that it is handling 100%
of the load."
> On Tue, Aug 2, 2011 at 7:2
On Mon, Aug 1, 2011 at 6:12 PM, Ryan King wrote:
> On Fri, Jul 29, 2011 at 12:02 PM, Chris Burroughs
> wrote:
>> On 07/25/2011 01:53 PM, Ryan King wrote:
>>> Actually I was wrong– our patch will disable gosisp and thrift but
>>> leave the process running:
>>>
>>> https://issues.apache.org/jira/br
Sorry to follow-up to my own post but I just saw this issue:
https://issues.apache.org/jira/browse/CASSANDRA-2118 linked in a
neighboring thread (cassandra server disk full). It certainly implies
that a disk IO failure resulting in a "zombie" node is a possibility.
Jim
On Tue, Aug 2, 2011 at 4:19
Ideally, I would hope that a bad disk wouldn't hang a node but would
instead just cause writes to fail, but if that is not the case,
perhaps the bad disk somehow wedged that server node completely so
that requests were not being processed at all (maybe not even being
timed out). At that point you'd
this is expected behaviour. Either increase RF or do a nodetool decommission
on a node to remove it from the ring.
On Tue, Aug 2, 2011 at 3:22 PM, Patrik Modesto wrote:
> Hi all!
>
> I've a test cluster of 4 nodes running cassandra 0.7.8, with one
> keyspace with RF=1, each node owns 25% of the d
Also I see these in the logs
"ERROR 08:53:47,678 Internal error processing batch_mutate
java.lang.IllegalStateException: replication factor (3) exceeds number
of endpoints (1)"
On Tue, Aug 2, 2011 at 7:21 AM, Aishwarya Venkataraman
wrote:
> Replies inline.
>
> Thanks,
> Aishwarya
>
> On Tue, Aug
Hi all!
I've a test cluster of 4 nodes running cassandra 0.7.8, with one
keyspace with RF=1, each node owns 25% of the data. As long as all
nodes are alive, there is no problem, but when I shut down just one
node I get UnavailableException in my application. cassandra-cli
returns "null" and hadoop
Replies inline.
Thanks,
Aishwarya
On Tue, Aug 2, 2011 at 7:12 AM, Sorin Julean wrote:
> Hi,
>
> Until someone answers with more details, few questions:
> 1. did you moved the system keyspace as well ?
Yes. But I deleted the LocationInfo* files under the system folder.
Shall I go ahead and del
Cassandra 0.7.6 , 4 node cluster
I ran a nodetool repair for a column family a week ago which caused the load
on that node to double. The diagnosis was that I had hit
https://issues.apache.org/jira/browse/CASSANDRA-2280
and that minor compactions will get rid of the unnecessary data and reduce
the
Hi,
Until someone answers with more details, few questions:
1. did you moved the system keyspace as well ?
2. the gossip IP of the new nodes are the same as the old ones ?
3. which cassandra version are you running ?
If 1. is yes and 2. is no, for a quick fix: take down the cluster, remove
s
We simply copy the snapshot data to another cluster to replace its data
directory. This cluster does not have any incoming traffic.
I haven't compared the md5 sum yet. Will try that.
Thanks,
John
On Thu, Jul 28, 2011 at 10:27 PM, Zhu Han wrote:
>
>
> On Thu, Jul 28, 2011 at 10:47 PM, Jian Fan
Hello,
I recently migrated 400 GB of data that was on a different cassandra
cluster (3 node with RF= 3) to a new cluster. I have a 3 node cluster
with replication factor set to three. When I run nodetool ring, it
does not show me all the nodes in the cluster. It always keeps showing
only one node
Rafael Almeida wrote:
> I find it hard to explain that without an example. So, let's say the token
> space
> is actually from 0 to 100 and we have 4 nodes (let's do this in order to make
> things more manageble). In our example, we have the following initial_tokens:
>
> node A = 0
> node B = 20
>
what is in the server log /var/log/cassandra/system.log ?
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 2 Aug 2011, at 20:40, Dikang Gu wrote:
> I'm using the hector 0.8.0-2 against cassandra 0.8.1.
>
> When I executed the code:
>
>
it's probably a proxy thing, check ant and or ivy to see how to set them.
Also, please stop using 0.6.3 there is little / no support available.
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 2 Aug 2011, at 18:51, Donna Li wrote:
> All:
Hang on, using brain now.
That is triggering a small bug in the code see
https://issues.apache.org/jira/browse/CASSANDRA-2984
For not just remove the column meta data.
Cheers
-
Aaron Morton
Freelance Cassandra Developer
@aaronmorton
http://www.thelastpickle.com
On 2 Aug 2011
What do you see when you run describe cluster; in the cassandra-cli ? Whats the
exact error you get and is there anything in the server side logs ?
Have you added other CF's before adding this one ? Did the schema agree before
starting this statement?
I ran the statement below on the current tr
I'm using the hector 0.8.0-2 against cassandra 0.8.1.
When I executed the code:
rangeSlicesQuery.setColumnFamily(columnFamily).setKeys(key, key)
.setRange(startColumn, null, reversed, count);
I got the following errors:
Caused by: me.prettyprint.hector.api.exceptions.HCassandraInternalExceptio
79 matches
Mail list logo