Re: How many Column Families can Cassandra handle?
At 600 CF's I would expect to see very frequent flushing to disk, as the algorithm that drives this from a memory standpoint is sensitive to the number of CF's. Additionally, and from experience on earlier versions, you can expect it to take over half an hour to make schema changes to over 500 CFs. Finally, if you ever have performance problems it's a lot harder to diagnose on a system with 600 CF's that it is one with 60. Hope that helps. - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 7:05 AM, Krishna Pisupat krishna.pisu...@gmail.com wrote: I don't know the full use case. However, for a generic time series scenario, we can make the timestamp (may be unto second part) part of the key and write all the data into the same CF(one CF for all data). Again, it may not make sense in your case, given the full use case. Just my 2 cents. Thanks and Regards, Krishna Pisupat krishna.pisu...@gmail.com On Sep 26, 2013, at 11:18 AM, Hiller, Dean dean.hil...@nrel.gov wrote: 600 is probably doable but each CF takes up memory……PlayOrm goes with a strategy that can virtualize CF's into one CF allowing less memory usage….we have 80,000 virtual CF's in cassandra through playorm….you can copy playorm's pattern if desired. But 600 is probably doable but high. 10,000 is not very doable. But you would have to try out 600 to see if it works for you….it may not work…try and find out in your load and context. NOTE: We have changed the 80,000 virtual CF's such that are in 10 real CF's these days so we get more parallel compaction going on. Dean From: Raihan Jamal jamalrai...@gmail.commailto:jamalrai...@gmail.com Reply-To: user@cassandra.apache.orgmailto:user@cassandra.apache.org user@cassandra.apache.orgmailto:user@cassandra.apache.org Date: Thursday, September 26, 2013 11:39 AM To: user@cassandra.apache.orgmailto:user@cassandra.apache.org user@cassandra.apache.orgmailto:user@cassandra.apache.org Subject: How many Column Families can Cassandra handle? I am working on a use case for Timeline series data. I have been told to create 600 column families in Cassandra. Meaning for 10 minutes, I will be having column families in Cassandra. Each second will have its own column family, so till 10 minutes which is 600 second, I will be having 600 column families... In each second, we will write into that particular second column family.. so at 10 minutes (which is 600 second), we will write into 600 second column family.. I am wondering whether Cassandra will be able to handle 600 column families or not.. Right now, I am not sure how much data each column family will have... What I know so far is write will be coming at a rate of 20,000 writes per second... Can anyone shed some light into this?
Re: Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter
Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer private though? Yes but I thought you had it in your examples, was included for completeness. use the official overloads. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 4:12 PM, Jayadev Jayaraman jdisal...@gmail.com wrote: Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer private though? I know what to do now in order to construct a column with a TTL now. Thanks. On Sep 26, 2013 9:00 PM, Aaron Morton aa...@thelastpickle.com wrote: org.apache.cassandra.thrift.Column column; // initialize this with name, value, timestamp, TTL This is the wrong object to use. one overload of addColumn() accepts IColumn which is from org.apache.cassanda.db . The thrift classes are only use for the thrift API. What is the difference between calling writer.addColumn() on the column's name, value and timestamp, and writer.addExpiringColumn() on the column's name, value, TTL, timestamp and expiration timestamp ? They both add an column to the row. addExpiringColumn() adds an expiring column, and addColumn adds a normal one. only addExpiringColumn accepts a TTL (in seconds) for the column. Does the former result in the column expiring still , in cassandra 1.2.x (i.e. does setting the TTL on a Column object change the name or value in a way so as to ensure the column will expire as required) ? No. An expiring column must be an ExpiringColumn column instance. The base IColumn interface does not have a TTL, only expiring columns do. If not , what is the TTL attribute used for in the Column object ? The org.apache.cassandra.db.Column class does not have a TTL. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 26/09/2013, at 12:44 AM, Jayadev Jayaraman jdisal...@gmail.com wrote: Can someone answer this doubt reg. SSTableSimpleWriter ? I'd asked about this earlier but it probably missed. Apologies for repeating the question (with minor additions) : Let's say I've initialized a SSTableSimpleWriter instance and a new column with TTL set : org.apache.cassandra.io.sstable.SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here */); org.apache.cassandra.thrift.Column column; // initialize this with name, value, timestamp, TTL What is the difference between calling writer.addColumn() on the column's name, value and timestamp, and writer.addExpiringColumn() on the column's name, value, TTL, timestamp and expiration timestamp ? Does the former result in the column expiring still , in cassandra 1.2.x (i.e. does setting the TTL on a Column object change the name or value in a way so as to ensure the column will expire as required) ? If not , what is the TTL attribute used for in the Column object ? Thanks, Jayadev On Tue, Sep 24, 2013 at 2:48 PM, Jayadev Jayaraman jdisal...@gmail.com wrote: Let's say I've initialized a SSTableSimpleWriter instance and a new column with TTL set : SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here */); Column column; What is the difference between calling writer.addColumn() on the column's name and value, and writer.addExpiringColumn() on the column and its TTL ? Does the former result in the column expiring still , in cassandra 1.2.x ? Or does it not ?
Re: Connecting to a remote cassandra node..
See the comments for rpc_address and listen_address in the yaml file. These control which interfaces the server binds to. If you leave them as blank and DNS is correctly setup they will bind to the correct address. If you set them to 0.0.0.0 they will bind to all interfaces, you should not do this in production. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 8:27 PM, Krishna Chaitanya bnsk1990r...@gmail.com wrote: Thank you...but cant we dynamically connect to any system instead of changing the .yaml file each time?? I want to take the input from the user regarding the host and port that he wants to use...Thanks. On Sep 27, 2013 1:38 PM, Kais Ahmed k...@neteck-fr.com wrote: hello, you have to check listen_address in cassandra.yaml, change the localhost value by the ip of the machine and restart cassandra 2013/9/27 Krishna Chaitanya bnsk1990r...@gmail.com Hello, I am relatively new to cassandra. I am using a library called libQtCassandra for accesing the cassandra database from my c++ programs. When I try to connect to the localhost cassandra , everything seems fine but when I try to connect to a remote node on which cassandra is up and running , it says connection refused. Any help would be of a great value. Thank You... -- Regards, BNSK.
Re: 2.0.1 counter replicate on write error
ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main] java.lang.AssertionError: DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) in /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db at org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114) at org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNamesIterator.java:62) When reading from an SSTable the position returned from the -Index.db / KEYS cache pointed to a row in the -Data.db component that was for a different row. DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) Was what we were searching for DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) Is what was found in the data component. The first part is the Token (M3 hash) the second is the key. It looks like a collision, but it could also be a bug somewhere else. Code in SSTableReader.getPosition() points to https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion that is only trigger if TRACE logging is running. Can you add to the 4687 ticket and update the thread ? Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 10:50 PM, Christopher Wirt chris.w...@struq.com wrote: Hello, I’ve started to see a slightly worrying error appear in our logs occasionally. We’re writing at 400qps per machine and I only see this appear every 5-10minutes. Seems to have started when I switched us to using the hsha thrift server this morning. We’ve been running 2.0.1 ran off the sync thrift server since yesterday without seeing this error. But might not be related. There are some machines in another DC still running 1.2.10. Anyone seen this before? Have any insight? ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main] java.lang.AssertionError: DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) in /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db at org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114) at org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNamesIterator.java:62) at org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(NamesQueryFilter.java:87) at org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFilter.java:62) at org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:249) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:53) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:1468) at org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1294) at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332) at org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadCommand.java:55) at org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutation.java:100) at org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java:1107) at org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:1897) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724)
Re: What is the best way to install upgrade Cassandra on Ubuntu ?
I am not sure if I should use datastax's DSC or official Debian packages from Cassandra. How do I choose between them for a production server ? They are technically the same. The DSC update will come out a little after the Apache release, and I _think_ they release for every Apache release. 1. when I upgrade to a newer version, would that retain my previous configurations so that I don't need to configure everything again ? Yes if you select that when doing the package install. 2. would that smoothly replace the previous installation by itself ? Yes 3. what's the way (kindly, if you can tell the command) to upgrade ? http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck 4. when should I prefer datastax's dsc to that ? (I need to install for production env.) Above Hope that helps. - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 11:01 PM, Ertio Lew ertio...@gmail.com wrote: I am not sure if I should use datastax's DSC or official Debian packages from Cassandra. How do I choose between them for a production server ? On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew ertio...@gmail.com wrote: Could you please clarify that: 1. when I upgrade to a newer version, would that retain my previous configurations so that I don't need to configure everything again ? 2. would that smoothly replace the previous installation by itself ? 3. what's the way (kindly, if you can tell the command) to upgrade ? 4. when should I prefer datastax's dsc to that ? (I need to install for production env.) On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli rc...@eventbrite.com wrote: On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew ertio...@gmail.com wrote: How do you install Cassandra on Ubuntu later how do you upgrade the installation on the node when an update has arrived ? Do you simply download replace the latest tar.gz, untar it to replace the older cassandra files? How do you do it ? How does this upgrade process differ for a major version upgrade, like say switching from 1.2 series to 2.0 series ? Use the deb packages. To upgrade, install the new package. Only upgrade a single major version. and be sure to consult NEWS.txt for any upgrade caveats. Also be aware of this sub-optimal behavior of the debian packages : https://issues.apache.org/jira/browse/CASSANDRA-2356 =Rob
Re: HintedHandoff process does not finish
What can be the reason for the handoff process not to finish? Check for other errors about timing out during hint reply. What would be the best way to recover from this situation? If they are really causing trouble drop the hints via HintedHandoffManager JMX MBean or stopping the node and deleting the files on disk. Then use repair later. What can be done to prevent this from happening again? Hints are stored when either the node is down before the request starts or when the coordinator times out waiting for the remote node. Check the logs for nodes going down, and check the MessagingService MBean for TimedOuts from other nodes. This may indicate issues with a cross DC connection. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 11:18 PM, Tom van den Berge t...@drillster.com wrote: Hi, One one of my nodes, the (storage) load increased dramatically (doubled), within one or two hours. The hints column family was causing the growth. I noticed one HintedHandoff process that was started some two hours ago, but hadn't finished. Normally, these processes take only a few seconds, 15 seconds max, in my cluster. The not-finishing process was handing the hints over to a host in another data center. There were no warning or error messages in the logs, other than the repeated flushing high-traffic column family hints. I'm using Cassandra 1.2.3. What can be the reason for the handoff process not to finish? What would be the best way to recover from this situation? What can be done to prevent this from happening again? Thanks in advance, Tom
Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target
It's an error in the antlr compilation, check the antlr versions. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera mianmarjun.mailingl...@gmail.com wrote: hi all: Environment • apache-cassandra-2.0.1-src • EC2 • Linux version 3.2.30-49.59.amzn1.x86_64 (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3) When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I have this error at the target, gen-cql2-grammar: gen-cql2-grammar: [echo] Building Grammar /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g …. [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1: Multiple token rules can match input such as '0'..'9': INTEGER, FLOAT, UUID [java] [java] As a result, token(s) FLOAT,UUID were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as 'I': K_INSERT, K_IN, K_INDEX, K_INTO, IDENT, COMPIDENT [java] [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as {'R', 'r'}: K_REVERSED, IDENT, COMPIDENT [java] [java] As a result, token(s) IDENT,COMPIDENT were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as 'T': K_LEVEL, K_TRUNCATE, K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT [java] [java] As a result, token(s) K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were disabled for that input [java] error(208): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1: The following token definitions can never be matched because prior tokens match the same input: T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT BUILD FAILED /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1 If I execute these targets in the next order first: • ant generate-cql-html • ant maven-ant-tasks-init and then execute ant default target • ant the project build succesfully. Regards. note: I do not have this issue in my mac. Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com
Re: Connecting to a remote cassandra node..
Thanks... On Sep 30, 2013 12:35 PM, Aaron Morton aa...@thelastpickle.com wrote: See the comments for rpc_address and listen_address in the yaml file. These control which interfaces the server binds to. If you leave them as blank and DNS is correctly setup they will bind to the correct address. If you set them to 0.0.0.0 they will bind to all interfaces, you should not do this in production. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 8:27 PM, Krishna Chaitanya bnsk1990r...@gmail.com wrote: Thank you...but cant we dynamically connect to any system instead of changing the .yaml file each time?? I want to take the input from the user regarding the host and port that he wants to use...Thanks. On Sep 27, 2013 1:38 PM, Kais Ahmed k...@neteck-fr.com wrote: hello, you have to check listen_address in cassandra.yaml, change the localhost value by the ip of the machine and restart cassandra 2013/9/27 Krishna Chaitanya bnsk1990r...@gmail.com Hello, I am relatively new to cassandra. I am using a library called libQtCassandra for accesing the cassandra database from my c++ programs. When I try to connect to the localhost cassandra , everything seems fine but when I try to connect to a remote node on which cassandra is up and running , it says connection refused. Any help would be of a great value. Thank You... -- Regards, BNSK.
Re: What is the best way to install upgrade Cassandra on Ubuntu ?
Thanks Aaron! Does DSC include other things like Opscenter by default ? I installed DSC on linux, but Opscenter wasn't installed there but when tried on Windows it was installed along with JRE python, using the windows installer. Would it be possible to remove any of these installations but keeping the data intact easily switch to the another, I mean switching from DSC package to apache one or vice versa ? On Mon, Sep 30, 2013 at 1:10 PM, Aaron Morton aa...@thelastpickle.comwrote: I am not sure if I should use datastax's DSC or official Debian packages from Cassandra. How do I choose between them for a production server ? They are technically the same. The DSC update will come out a little after the Apache release, and I _think_ they release for every Apache release. 1. when I upgrade to a newer version, would that retain my previous configurations so that I don't need to configure everything again ? Yes if you select that when doing the package install. 2. would that smoothly replace the previous installation by itself ? Yes 3. what's the way (kindly, if you can tell the command) to upgrade ? http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck 4. when should I prefer datastax's dsc to that ? (I need to install for production env.) Above Hope that helps. - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 11:01 PM, Ertio Lew ertio...@gmail.com wrote: I am not sure if I should use datastax's DSC or official Debian packages from Cassandra. How do I choose between them for a production server ? On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew ertio...@gmail.com wrote: Could you please clarify that: 1. when I upgrade to a newer version, would that retain my previous configurations so that I don't need to configure everything again ? 2. would that smoothly replace the previous installation by itself ? 3. what's the way (kindly, if you can tell the command) to upgrade ? 4. when should I prefer datastax's dsc to that ? (I need to install for production env.) On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli rc...@eventbrite.comwrote: On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew ertio...@gmail.com wrote: How do you install Cassandra on Ubuntu later how do you upgrade the installation on the node when an update has arrived ? Do you simply download replace the latest tar.gz, untar it to replace the older cassandra files? How do you do it ? How does this upgrade process differ for a major version upgrade, like say switching from 1.2 series to 2.0 series ? Use the deb packages. To upgrade, install the new package. Only upgrade a single major version. and be sure to consult NEWS.txt for any upgrade caveats. Also be aware of this sub-optimal behavior of the debian packages : https://issues.apache.org/jira/browse/CASSANDRA-2356 =Rob
RE: 2.0.1 counter replicate on write error
Thanks Aaron, I've added to the ticket. We were not running on TRACE logging. From: Aaron Morton [mailto:aa...@thelastpickle.com] Sent: 30 September 2013 08:37 To: user@cassandra.apache.org Subject: Re: 2.0.1 counter replicate on write error ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main] java.lang.AssertionError: DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) in /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195 3-Data.db at org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName sIterator.java:114) at org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNa mesIterator.java:62) When reading from an SSTable the position returned from the -Index.db / KEYS cache pointed to a row in the -Data.db component that was for a different row. DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) Was what we were searching for DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) Is what was found in the data component. The first part is the Token (M3 hash) the second is the key. It looks like a collision, but it could also be a bug somewhere else. Code in SSTableReader.getPosition() points to https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion that is only trigger if TRACE logging is running. Can you add to the 4687 ticket and update the thread ? Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 10:50 PM, Christopher Wirt chris.w...@struq.com wrote: Hello, I've started to see a slightly worrying error appear in our logs occasionally. We're writing at 400qps per machine and I only see this appear every 5-10minutes. Seems to have started when I switched us to using the hsha thrift server this morning. We've been running 2.0.1 ran off the sync thrift server since yesterday without seeing this error. But might not be related. There are some machines in another DC still running 1.2.10. Anyone seen this before? Have any insight? ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main] java.lang.AssertionError: DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0) in /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195 3-Data.db at org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName sIterator.java:114) at org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNa mesIterator.java:62) at org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(Nam esQueryFilter.java:87) at org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFil ter.java:62) at org.apache.cassandra.db.CollationController.collectAllData(CollationControll er.java:249) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationCont roller.java:53) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilySto re.java:1468) at org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore. java:1294) at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332) at org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadComma nd.java:55) at org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutat ion.java:100) at org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java: 1107) at org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy .java:1897) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:11 45) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:6 15) at java.lang.Thread.run(Thread.java:724)
Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target
hi: is that mean that antlr-3.2.jar is not the correct version? what is the correct version? In the build file, I see that cassandra uses the jar lib at the * ${build.lib}* folder, in this case antlr-3.2.jar ... target name=gen-cql2-grammar depends=check-gen-cql2-grammar unless=cql2current echoBuilding Grammar ${build.src.java}/org/apache/cassandra/cql/Cql.g .../echo java classname=org.antlr.Tool * classpath=${build.lib}/antlr-3.2.jar* fork=true failonerror=true arg value=${build.src.java}/org/apache/cassandra/cql/Cql.g / arg value=-fo / arg value=${build.src.gen-java}/org/apache/cassandra/cql/ / /java /target ... thanks in advance Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com 2013/9/30 Aaron Morton aa...@thelastpickle.com It's an error in the antlr compilation, check the antlr versions. Cheers - Aaron Morton New Zealand @aaronmorton Co-Founder Principal Consultant Apache Cassandra Consulting http://www.thelastpickle.com On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera mianmarjun.mailingl...@gmail.com wrote: hi all: Environment • apache-cassandra-2.0.1-src • EC2 • Linux version 3.2.30-49.59.amzn1.x86_64 (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3) When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I have this error at the target, gen-cql2-grammar: gen-cql2-grammar: [echo] Building Grammar /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g …. [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1: Multiple token rules can match input such as '0'..'9': INTEGER, FLOAT, UUID [java] [java] As a result, token(s) FLOAT,UUID were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as 'I': K_INSERT, K_IN, K_INDEX, K_INTO, IDENT, COMPIDENT [java] [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as {'R', 'r'}: K_REVERSED, IDENT, COMPIDENT [java] [java] As a result, token(s) IDENT,COMPIDENT were disabled for that input [java] warning(209): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1: Multiple token rules can match input such as 'T': K_LEVEL, K_TRUNCATE, K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT [java] [java] As a result, token(s) K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were disabled for that input [java] error(208): /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1: The following token definitions can never be matched because prior tokens match the same input: T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT BUILD FAILED /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1 If I execute these targets in the next order first: • ant generate-cql-html • ant maven-ant-tasks-init and then execute ant default target • ant the project build succesfully. Regards. note: I do not have this issue in my mac. Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com
Re: Among Datastax community Cassandra debian package, which to choose for production install ?
OpsCenter should be a separate package as you would only install it on a single node, not necessarily even one that is running Cassandra. On Sat, Sep 28, 2013 at 2:12 PM, Ertio Lew ertio...@gmail.com wrote: I think both provide the same thing except Datastax Community also provides some extras like Opscenter, etc. But I cannot find opscenter installed when I installled DSC on ubuntu. Although on windows installation, I saw opscenter JRE as well , so I think for DSC, there is no such prerequisite for Oracle JRE as required for Cassandra debain package, is it so ? Btw which is usually preferred for production installs ? I may need to use Opscenter but just *occasionally*. -- *Ken Hancock *| System Architect, Advanced Advertising SeaChange International 50 Nagog Park Acton, Massachusetts 01720 ken.hanc...@schange.com | www.schange.com | NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image: LinkedIn]http://www.linkedin.com/in/kenhancock [image: SeaChange International] http://www.schange.com/This e-mail and any attachments may contain information which is SeaChange International confidential. The information enclosed is intended only for the addressees herein and may not be copied or forwarded without permission from SeaChange International.
Cassandra Summit EU 2013
For those in the Europe area, there will be a Cassandra Summit EU 2013 in London in the month of October. On 17 October, there will be the main conference sessions and the 16th and 18th there will be Cassandra workshops. http://www.datastax.com/cassandraeurope2013 The speakers have been announced and the presentation abstracts are all on there. Like always, the presentations will be recorded and posted on Planet Cassandra, but it's great to meet and interact with people in the community - in my opinion that's the best part of any conference. Anyway, just wanted to make sure people knew. Cheers, Jeremy
Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1
hi: I try to test cassandra 2.0.1 and pig 0.11.1 but always when i try to load any keyspacei have this error: Example creating a test keyspace and data: I have this error: *grunt define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();* *grunt rows = LOAD 'cql://keyspace0/test' USING org.apache.cassandra.hadoop.pig.CqlStorage();* *2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt - ERROR 2999: Unexpected internal error. InvalidRequestException(why:Undefined name key_alias in selection clause)* *Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log* and in the log file there are not more info Pig Stack Trace --- ERROR 2999: Unexpected internal error. InvalidRequestException(why:Undefined name key_alias in selection clause) java.lang.RuntimeException: InvalidRequestException(why:Undefined name key_alias in selection clause) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511) at org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246) at org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280) at org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151) at org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110) at org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100) at org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219) at org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75) at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50) at org.apache.pig.newplan.logical.visitor.CastLineageSetter.init(CastLineageSetter.java:57) at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635) at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566) at org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538) at org.apache.pig.PigServer.registerQuery(PigServer.java:540) at org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970) at org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165) at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69) at org.apache.pig.Main.run(Main.java:490) at org.apache.pig.Main.main(Main.java:111) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.util.RunJar.main(RunJar.java:156) Caused by: InvalidRequestException(why:Undefined name key_alias in selection clause) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898) at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78) at org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658) at org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500) ... 25 more any idea? thanks Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com
Re: Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1
I have the same issue using cassandraStorage() and other dataset for this *note*: the example cql3 datatest are CREATE TABLE test ( id text PRIMARY KEY, title text, age int ); and insert some dummy data insert into test (id, title, age) values('1', 'child', 21); insert into test (id, title, age) values('2', 'support', 21); insert into test (id, title, age) values('3', 'manager', 31); insert into test (id, title, age) values('4', 'QA', 41); insert into test (id, title, age) values('5', 'QA', 30); insert into test (id, title, age) values('6', 'QA', 30); Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com 2013/9/30 Miguel Angel Martin junquera mianmarjun.mailingl...@gmail.com hi: I try to test cassandra 2.0.1 and pig 0.11.1 but always when i try to load any keyspacei have this error: Example creating a test keyspace and data: I have this error: *grunt define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();* *grunt rows = LOAD 'cql://keyspace0/test' USING org.apache.cassandra.hadoop.pig.CqlStorage();* *2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt - ERROR 2999: Unexpected internal error. InvalidRequestException(why:Undefined name key_alias in selection clause)* *Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log* and in the log file there are not more info Pig Stack Trace --- ERROR 2999: Unexpected internal error. InvalidRequestException(why:Undefined name key_alias in selection clause) java.lang.RuntimeException: InvalidRequestException(why:Undefined name key_alias in selection clause) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511) at org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246) at org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280) at org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151) at org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110) at org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100) at org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219) at org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75) at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50) at org.apache.pig.newplan.logical.visitor.CastLineageSetter.init(CastLineageSetter.java:57) at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635) at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566) at org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538) at org.apache.pig.PigServer.registerQuery(PigServer.java:540) at org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970) at org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189) at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165) at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69) at org.apache.pig.Main.run(Main.java:490) at org.apache.pig.Main.main(Main.java:111) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.util.RunJar.main(RunJar.java:156) Caused by: InvalidRequestException(why:Undefined name key_alias in selection clause) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983) at org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898) at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78) at org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658) at org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573) at org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500) ... 25 more any idea? thanks Miguel Angel Martín Junquera Analyst Engineer. miguelangel.mar...@brainsins.com
Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
Solr's data is stored on the file system as a set of index files[ http://stackoverflow.com/a/7685579/530153]. Then why do we need anything like Solandra or DataStax Enterprise Search? Isn't Solr complete solution in itself ? What do we need to integrate with Cassandra ?
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
the main reason is scalability and performance. If your Solr indexes fit fine on a single system and doesn't need to scale out, Cassandra/HDFS isn't necessary. On Mon, Sep 30, 2013 at 11:50 AM, Ertio Lew ertio...@gmail.com wrote: Solr's data is stored on the file system as a set of index files[ http://stackoverflow.com/a/7685579/530153]. Then why do we need anything like Solandra or DataStax Enterprise Search? Isn't Solr complete solution in itself ? What do we need to integrate with Cassandra ?
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew ertio...@gmail.com wrote: Solr's data is stored on the file system as a set of index files[ http://stackoverflow.com/a/7685579/530153]. Then why do we need anything like Solandra or DataStax Enterprise Search? Isn't Solr complete solution in itself ? What do we need to integrate with Cassandra ? Solr's index sitting on a single machine, even if that single machine can vertically scale, is a single point of failure. The value add of DES is that the index has the same availability characteristics as the underlying data, because it is stored in the same cluster. =Rob
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
To clarify, solr indexes are not distributed in the same way that Cassandra data is stored. With Cassandra, each node receives a fraction of the keyspace (based on your replication factor and token assignment). With DSE Search, writes to Cassandra are hooked and each node independently indexes its data and keeps this index on the local file system. If you have keyspace with RF=3 then three nodes will index each document. Indexes, unlike from Solr, only store the docids and the actual field values are stored in Cassandra. When it comes to search, DSE splits up the search so in the example above only one of those RF=3 nodes will be queried for a particular token range so that data can be unioned across all the nodes with different token ranges. Not sure about Solandra, but you do need to be aware that there's a number of Solr search options that are not supported on distribute searches/DSE Search. http://wiki.apache.org/solr/DistributedSearch http://wiki.apache.org/solr/FieldCollapsing Also, be aware that while Cassandra has knobs to allow you to get consistent read results (CL=QUORUM), DSE Search does not. If a node drops messages for whatever reason, outtage, mutation, etc. its solr indexes will be inconsistent with other nodes in its replication group. On Mon, Sep 30, 2013 at 1:06 PM, Robert Coli rc...@eventbrite.com wrote: On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew ertio...@gmail.com wrote: Solr's data is stored on the file system as a set of index files[ http://stackoverflow.com/a/7685579/530153]. Then why do we need anything like Solandra or DataStax Enterprise Search? Isn't Solr complete solution in itself ? What do we need to integrate with Cassandra ? Solr's index sitting on a single machine, even if that single machine can vertically scale, is a single point of failure. The value add of DES is that the index has the same availability characteristics as the underlying data, because it is stored in the same cluster. =Rob -- *Ken Hancock *| System Architect, Advanced Advertising SeaChange International 50 Nagog Park Acton, Massachusetts 01720 ken.hanc...@schange.com | www.schange.com | NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image: LinkedIn]http://www.linkedin.com/in/kenhancock [image: SeaChange International] http://www.schange.com/This e-mail and any attachments may contain information which is SeaChange International confidential. The information enclosed is intended only for the addressees herein and may not be copied or forwarded without permission from SeaChange International.
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
Also, be aware that while Cassandra has knobs to allow you to get consistent read results (CL=QUORUM), DSE Search does not. If a node drops messages for whatever reason, outtage, mutation, etc. its solr indexes will be inconsistent with other nodes in its replication group. Will repair fix it?
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
Yes. On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh ailin...@gmail.com wrote: Also, be aware that while Cassandra has knobs to allow you to get consistent read results (CL=QUORUM), DSE Search does not. If a node drops messages for whatever reason, outtage, mutation, etc. its solr indexes will be inconsistent with other nodes in its replication group. Will repair fix it? -- *Ken Hancock *| System Architect, Advanced Advertising SeaChange International 50 Nagog Park Acton, Massachusetts 01720 ken.hanc...@schange.com | www.schange.com | NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image: LinkedIn]http://www.linkedin.com/in/kenhancock [image: SeaChange International] http://www.schange.com/This e-mail and any attachments may contain information which is SeaChange International confidential. The information enclosed is intended only for the addressees herein and may not be copied or forwarded without permission from SeaChange International.
How to log the details of the updated data locally
Hi all, In cassandra cluster, once a write/update is successful to the particular instance locally, i want to log the data that is updated and its timestamps separately in a file.. In which class it will be more appropriate to do this.. To the extent i explored the codebase, it's possible to do that in Keyspace class (apply method) in db package.. But i don't know how to retrieve the timestamp details from the mutation object.. Is Keyspace class is appropriate for my purpose.. If yes, please kindly provide me some ideas to retrieve timestamp details from mutation object(RowMutation).. Any help is appreciable.. Looking forward for your kind replies.. Thanks in advance.. Thank you.
Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?
Solr's index sitting on a single machine, even if that single machine can vertically scale, is a single point of failure. And about Cloud Solr? 2013/9/30 Ken Hancock ken.hanc...@schange.com Yes. On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh ailin...@gmail.comwrote: Also, be aware that while Cassandra has knobs to allow you to get consistent read results (CL=QUORUM), DSE Search does not. If a node drops messages for whatever reason, outtage, mutation, etc. its solr indexes will be inconsistent with other nodes in its replication group. Will repair fix it? -- *Ken Hancock *| System Architect, Advanced Advertising SeaChange International 50 Nagog Park Acton, Massachusetts 01720 ken.hanc...@schange.com | www.schange.com | NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image: LinkedIn] http://www.linkedin.com/in/kenhancock [image: SeaChange International] http://www.schange.com/This e-mail and any attachments may contain information which is SeaChange International confidential. The information enclosed is intended only for the addressees herein and may not be copied or forwarded without permission from SeaChange International. -- Atenciosamente, Sávio S. Teles de Oliveira voice: +55 62 9136 6996 http://br.linkedin.com/in/savioteles Mestrando em Ciências da Computação - UFG Arquiteto de Software Laboratory for Ubiquitous and Pervasive Applications (LUPA) - UFG
Segmentation fault when trying to store into cassandra...
Hello, I modified a network probe which collects network packets to store them into cassandra. So there are many packets that are coming in, I capture the packets in the program and store them into cassandra. I am using libQtCassandra library. The program is crashing with segmentation fault as soon as I run it. Can someone help as to what all can go wrong here?? Could there be a problem with row/col keys or is it some configuration parameter or the speed at which the packets or coming? I am not able to figure it out. Thank you. -- Regards, BNSK*. *
Re: Segmentation fault when trying to store into cassandra...
Java version issue? Using sun jdk or open jdk? -Vivek On Tue, Oct 1, 2013 at 6:16 AM, Krishna Chaitanya bnsk1990r...@gmail.comwrote: Hello, I modified a network probe which collects network packets to store them into cassandra. So there are many packets that are coming in, I capture the packets in the program and store them into cassandra. I am using libQtCassandra library. The program is crashing with segmentation fault as soon as I run it. Can someone help as to what all can go wrong here?? Could there be a problem with row/col keys or is it some configuration parameter or the speed at which the packets or coming? I am not able to figure it out. Thank you. -- Regards, BNSK*. *