Re: How many Column Families can Cassandra handle?

2013-09-30 Thread Aaron Morton
At 600 CF's I would expect to see very frequent flushing to disk, as the 
algorithm that drives this from a memory standpoint is sensitive to the number 
of CF's. 

Additionally, and from experience on earlier versions, you can expect it to 
take over half an hour to make schema changes to over 500 CFs. 

Finally, if you ever have performance problems it's a lot harder to diagnose on 
a system with 600 CF's that it is one with 60. 

Hope that helps. 

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 7:05 AM, Krishna Pisupat krishna.pisu...@gmail.com wrote:

 I don't know the full use case. However, for a generic time series scenario, 
 we can make the timestamp (may be unto second part) part of the key and write 
 all the data into the same CF(one CF for all data). Again, it may not make 
 sense in your case, given the full use case. Just my 2 cents. 
 
 
 Thanks and Regards,
 Krishna Pisupat
 krishna.pisu...@gmail.com
 
 
 
 On Sep 26, 2013, at 11:18 AM, Hiller, Dean dean.hil...@nrel.gov wrote:
 
 600 is probably doable but each CF takes up memory……PlayOrm goes with a 
 strategy that can virtualize CF's into one CF allowing less memory usage….we 
 have 80,000 virtual CF's in cassandra through playorm….you can copy 
 playorm's pattern if desired.  But 600 is probably doable but high.  10,000 
 is not very doable.
 
 But you would have to try out 600 to see if it works for you….it may not 
 work…try and find out in your load and context.
 
 NOTE: We have changed the 80,000 virtual CF's such that are in 10 real CF's 
 these days so we get more parallel compaction going on.
 
 Dean
 
 From: Raihan Jamal jamalrai...@gmail.commailto:jamalrai...@gmail.com
 Reply-To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
 user@cassandra.apache.orgmailto:user@cassandra.apache.org
 Date: Thursday, September 26, 2013 11:39 AM
 To: user@cassandra.apache.orgmailto:user@cassandra.apache.org 
 user@cassandra.apache.orgmailto:user@cassandra.apache.org
 Subject: How many Column Families can Cassandra handle?
 
 I am working on a use case for Timeline series data. I have been told to 
 create 600 column families in Cassandra. Meaning for 10 minutes, I will be 
 having column families in Cassandra. Each second will have its own column 
 family, so till 10 minutes which is 600 second, I will be having 600 column 
 families...
 
 In each second, we will write into that particular second column family.. so 
 at 10 minutes (which is 600 second), we will write into 600 second column 
 family..
 
 I am wondering whether Cassandra will be able to handle 600 column families 
 or not.. Right now, I am not sure how much data each column family will 
 have... What I know so far is write will be coming at a rate of 20,000 
 writes per second...
 
 Can anyone shed some light into this?
 



Re: Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter

2013-09-30 Thread Aaron Morton
 Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer 
 private though?
 
 

Yes but I thought you had it in your examples, was included for completeness. 
use the official overloads. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 4:12 PM, Jayadev Jayaraman jdisal...@gmail.com wrote:

 Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer 
 private though? I know what to do now in order to construct a column with a 
 TTL now. Thanks.
 
 On Sep 26, 2013 9:00 PM, Aaron Morton aa...@thelastpickle.com wrote:
  org.apache.cassandra.thrift.Column column; // initialize this with name, 
  value, timestamp, TTL
 This is the wrong object to use.
 
 one overload of addColumn() accepts IColumn which is from 
 org.apache.cassanda.db . The thrift classes are only use for the thrift API.
 
  What is the difference between calling writer.addColumn() on the column's 
  name, value and timestamp, and writer.addExpiringColumn() on the column's 
  name, value, TTL, timestamp and expiration timestamp ?
 They both add an column to the row. addExpiringColumn() adds an expiring 
 column, and addColumn adds a normal one.
 
 only addExpiringColumn accepts a TTL (in seconds) for the column.
 
 
  Does the former result in the column expiring still , in cassandra 1.2.x 
  (i.e. does setting the TTL on a Column object change the name or value in a 
  way so as to ensure the column will expire as required) ?
 No.
 An expiring column must be an ExpiringColumn column instance.
 The base IColumn interface does not have a TTL, only expiring columns do.
 
   If not , what is the TTL attribute used for in the Column object ?
 The org.apache.cassandra.db.Column class does not have a TTL.
 
 Cheers
 
 
 -
 Aaron Morton
 New Zealand
 @aaronmorton
 
 Co-Founder  Principal Consultant
 Apache Cassandra Consulting
 http://www.thelastpickle.com
 
 On 26/09/2013, at 12:44 AM, Jayadev Jayaraman jdisal...@gmail.com wrote:
 
  Can someone answer this doubt reg. SSTableSimpleWriter ? I'd asked about 
  this earlier but it probably missed. Apologies for repeating the question 
  (with minor additions)  :
 
  
  Let's say I've initialized a SSTableSimpleWriter instance and a new column 
  with TTL set :
 
  org.apache.cassandra.io.sstable.SSTableSimpleWriter writer = new 
  SSTableSimpleWriter( ... /* params here */);
  org.apache.cassandra.thrift.Column column; // initialize this with name, 
  value, timestamp, TTL
 
  What is the difference between calling writer.addColumn() on the column's 
  name, value and timestamp, and writer.addExpiringColumn() on the column's 
  name, value, TTL, timestamp and expiration timestamp ? Does the former 
  result in the column expiring still , in cassandra 1.2.x (i.e. does setting 
  the TTL on a Column object change the name or value in a way so as to 
  ensure the column will expire as required) ? If not , what is the TTL 
  attribute used for in the Column object ?
  
 
  Thanks,
  Jayadev
 
 
  On Tue, Sep 24, 2013 at 2:48 PM, Jayadev Jayaraman jdisal...@gmail.com 
  wrote:
  Let's say I've initialized a SSTableSimpleWriter instance and a new column 
  with TTL set :
 
  SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here 
  */);
  Column column;
 
  What is the difference between calling writer.addColumn() on the column's 
  name and value, and writer.addExpiringColumn() on the column and its TTL ? 
  Does the former result in the column expiring still , in cassandra 1.2.x ? 
  Or does it not ?
 
 
 
 



Re: Connecting to a remote cassandra node..

2013-09-30 Thread Aaron Morton
See the comments for rpc_address and listen_address in the yaml file. These 
control which interfaces the server binds to. 

If you leave them as blank and DNS is correctly setup they will bind to the 
correct address. 

If you set them to 0.0.0.0 they will bind to all interfaces, you should not do 
this in production. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 8:27 PM, Krishna Chaitanya bnsk1990r...@gmail.com wrote:

 Thank you...but cant we dynamically connect to any system instead of changing 
 the .yaml file each time?? I want to take the input from the user regarding 
 the host and port that he wants to use...Thanks.
 
 On Sep 27, 2013 1:38 PM, Kais Ahmed k...@neteck-fr.com wrote:
 hello,
 
 you have to check listen_address in cassandra.yaml, change the localhost 
 value by the ip of the machine and restart cassandra  
 
 
 2013/9/27 Krishna Chaitanya bnsk1990r...@gmail.com
 Hello,
   I am relatively new to cassandra. I am using a library called 
 libQtCassandra for accesing the cassandra database from my c++ programs. When 
 I try to connect to the localhost cassandra , everything seems fine but when 
 I try to connect to a remote node on which cassandra is up and running , it 
 says connection refused. Any help would be of a great value. Thank You...
 
 -- 
 Regards,
 BNSK.
 



Re: 2.0.1 counter replicate on write error

2013-09-30 Thread Aaron Morton
 ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java 
 (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main]
 java.lang.AssertionError: DecoratedKey(-1754949563326053382, 
 a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, 
 aeadcec8184445d4ab631ef4250927d0) in 
 /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db
 at 
 org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114)
 at 
 org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNamesIterator.java:62)

When reading from an SSTable the position returned from the -Index.db / KEYS 
cache pointed to a row in the -Data.db component that was for a different row. 

 DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0)
Was what we were searching for

 DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6)
Is what was found in the data component. 

The first part is the Token (M3 hash) the second is the key. It looks like a 
collision, but it could also be a bug somewhere else. 

Code in SSTableReader.getPosition() points to 
https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion that 
is only trigger if TRACE logging is running. Can you add to the 4687 ticket and 
update the thread ? 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 10:50 PM, Christopher Wirt chris.w...@struq.com wrote:

 Hello,
  
 I’ve started to see a slightly worrying error appear in our logs 
 occasionally. We’re writing at 400qps per machine and I only see this appear 
 every 5-10minutes.
  
 Seems to have started when I switched us to using the hsha thrift server this 
 morning. We’ve been running 2.0.1 ran off the sync thrift server since 
 yesterday without seeing this error.  But might not be related.
  
 There are some machines in another DC still running 1.2.10.
  
 Anyone seen this before? Have any insight?
  
 ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java 
 (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main]
 java.lang.AssertionError: DecoratedKey(-1754949563326053382, 
 a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, 
 aeadcec8184445d4ab631ef4250927d0) in 
 /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db
 at 
 org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114)
 at 
 org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNamesIterator.java:62)
 at 
 org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(NamesQueryFilter.java:87)
 at 
 org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFilter.java:62)
 at 
 org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:249)
 at 
 org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:53)
 at 
 org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:1468)
 at 
 org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1294)
 at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332)
 at 
 org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadCommand.java:55)
 at 
 org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutation.java:100)
 at 
 org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java:1107)
 at 
 org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:1897)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:724)



Re: What is the best way to install upgrade Cassandra on Ubuntu ?

2013-09-30 Thread Aaron Morton
 I am not sure if I should use datastax's DSC or official Debian packages from 
 Cassandra. How do I choose between them for a production server ?
They are technically the same. 
The DSC update will come out a little after the Apache release, and I _think_ 
they release for every Apache release.

  1.  when I upgrade to a newer version, would that retain my previous 
 configurations so that I don't need to configure everything again ? 

Yes if you select that when doing the package install. 

 2.  would that smoothly replace the previous installation by itself ?

Yes

 3.  what's the way (kindly, if you can tell the command) to upgrade ?


http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck

 4. when should I prefer datastax's dsc to that ? (I need to install for 
 production env.)

Above

Hope that helps. 


-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:01 PM, Ertio Lew ertio...@gmail.com wrote:

 I am not sure if I should use datastax's DSC or official Debian packages from 
 Cassandra. How do I choose between them for a production server ?
 
 
 
 On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew ertio...@gmail.com wrote:
 
  Could you please clarify that:
 1.  when I upgrade to a newer version, would that retain my previous 
 configurations so that I don't need to configure everything again ? 
 2.  would that smoothly replace the previous installation by itself ?
 3.  what's the way (kindly, if you can tell the command) to upgrade ?
 4. when should I prefer datastax's dsc to that ? (I need to install for 
 production env.)
 
 
 On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli rc...@eventbrite.com wrote:
 On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew ertio...@gmail.com wrote:
 How do you install Cassandra on Ubuntu  later how do you upgrade the 
 installation on the node when an update has arrived ? Do you simply download 
  replace the latest tar.gz, untar it to replace the older cassandra files? 
 How do you do it ? How does this upgrade process differ for a major version 
 upgrade, like say switching from 1.2 series to 2.0 series ?
 
 Use the deb packages. To upgrade, install the new package. Only upgrade a 
 single major version. and be sure to consult NEWS.txt for any upgrade caveats.
 
 Also be aware of this sub-optimal behavior of the debian packages :
 
 https://issues.apache.org/jira/browse/CASSANDRA-2356
 
 =Rob
 
 
 



Re: HintedHandoff process does not finish

2013-09-30 Thread Aaron Morton
 What can be the reason for the handoff process not to finish?
Check for other errors about timing out during hint reply. 

 What would be the best way to recover from this situation?
If they are really causing trouble drop the hints via HintedHandoffManager JMX 
MBean or stopping the node and deleting the files on disk. Then use repair 
later. 

 What can be done to prevent this from happening again?
Hints are stored when either the node is down before the request starts or when 
the coordinator times out waiting for the remote node. Check the logs for nodes 
going down, and check the MessagingService MBean for TimedOuts from other 
nodes. This may indicate issues with a cross DC connection. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:18 PM, Tom van den Berge t...@drillster.com wrote:

 Hi,
 
 One one of my nodes, the (storage) load increased dramatically (doubled), 
 within one or two hours. The hints column family was causing the growth. I 
 noticed one HintedHandoff process that was started some two hours ago, but 
 hadn't finished. Normally, these processes take only a few seconds, 15 
 seconds max, in my cluster.
 
 The not-finishing process was handing the hints over to a host in another 
 data center. There were no warning or error messages in the logs, other than 
 the repeated flushing high-traffic column family hints.
 I'm using Cassandra 1.2.3.
 What can be the reason for the handoff process not to finish?
 What would be the best way to recover from this situation?
 What can be done to prevent this from happening again?
 
 Thanks in advance,
 Tom
 
 
 
 
 



Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target

2013-09-30 Thread Aaron Morton
It's an error in the antlr compilation, check the antlr versions. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder  Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera 
mianmarjun.mailingl...@gmail.com wrote:

 
 
 
 
 hi all:
 
 Environment
 
   • apache-cassandra-2.0.1-src
   • EC2
   • Linux version 3.2.30-49.59.amzn1.x86_64 
 (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3) 
 
 When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I  have 
 this error at  the target, gen-cql2-grammar:
 
 
 gen-cql2-grammar:
  [echo] Building Grammar 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g
 
 ….

   [java] warning(209): 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1:
  Multiple token rules can match input such as '0'..'9': INTEGER, FLOAT, UUID
  [java] 
  [java] As a result, token(s) FLOAT,UUID were disabled for that input
  [java] warning(209): 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
  Multiple token rules can match input such as 'I': K_INSERT, K_IN, K_INDEX, 
 K_INTO, IDENT, COMPIDENT
  [java] 
  [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT were 
 disabled for that input
  [java] warning(209): 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
  Multiple token rules can match input such as {'R', 'r'}: K_REVERSED, 
 IDENT, COMPIDENT
  [java] 
  [java] As a result, token(s) IDENT,COMPIDENT were disabled for that input
  [java] warning(209): 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
  Multiple token rules can match input such as 'T': K_LEVEL, K_TRUNCATE, 
 K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT
  [java] 
  [java] As a result, token(s) 
 K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were 
 disabled for that input
  [java] error(208): 
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1:
  The following token definitions can never be matched because prior tokens 
 match the same input: 
 T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT
 
 BUILD FAILED
 /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1
 
 
 
 
 
 
 If  I  execute these targets in the next order first:
 
 
 
   • ant generate-cql-html
   • ant maven-ant-tasks-init
 and then  execute ant default target
 
 
   • ant
 
 
 the project build succesfully.
 
 Regards.
 
 
  
 
 
 note:
 
 I do not have this issue in my mac.
 
 
 
 
 
 
 Miguel Angel Martín Junquera
 Analyst Engineer.
 miguelangel.mar...@brainsins.com
 



Re: Connecting to a remote cassandra node..

2013-09-30 Thread Krishna Chaitanya
Thanks...
On Sep 30, 2013 12:35 PM, Aaron Morton aa...@thelastpickle.com wrote:

 See the comments for rpc_address and listen_address in the yaml file.
 These control which interfaces the server binds to.

 If you leave them as blank and DNS is correctly setup they will bind to
 the correct address.

 If you set them to 0.0.0.0 they will bind to all interfaces, you should
 not do this in production.

 Cheers

 -
 Aaron Morton
 New Zealand
 @aaronmorton

 Co-Founder  Principal Consultant
 Apache Cassandra Consulting
 http://www.thelastpickle.com

 On 27/09/2013, at 8:27 PM, Krishna Chaitanya bnsk1990r...@gmail.com
 wrote:

  Thank you...but cant we dynamically connect to any system instead of
 changing the .yaml file each time?? I want to take the input from the user
 regarding the host and port that he wants to use...Thanks.
 
  On Sep 27, 2013 1:38 PM, Kais Ahmed k...@neteck-fr.com wrote:
  hello,
 
  you have to check listen_address in cassandra.yaml, change the localhost
 value by the ip of the machine and restart cassandra
 
 
  2013/9/27 Krishna Chaitanya bnsk1990r...@gmail.com
  Hello,
I am relatively new to cassandra. I am using a library called
 libQtCassandra for accesing the cassandra database from my c++ programs.
 When I try to connect to the localhost cassandra , everything seems fine
 but when I try to connect to a remote node on which cassandra is up and
 running , it says connection refused. Any help would be of a great value.
 Thank You...
 
  --
  Regards,
  BNSK.
 




Re: What is the best way to install upgrade Cassandra on Ubuntu ?

2013-09-30 Thread Ertio Lew
Thanks Aaron!

Does DSC include other things like Opscenter by default ? I installed DSC
on linux, but Opscenter wasn't installed there but when tried on Windows it
was installed along with JRE  python, using the windows installer.

Would it be possible to remove any of these installations but keeping the
data intact  easily switch to the another, I mean switching from DSC
package to apache one or vice versa ?


On Mon, Sep 30, 2013 at 1:10 PM, Aaron Morton aa...@thelastpickle.comwrote:

 I am not sure if I should use datastax's DSC or official Debian packages
 from Cassandra. How do I choose between them for a production server ?

 They are technically the same.
 The DSC update will come out a little after the Apache release, and I
 _think_ they release for every Apache release.

  1.  when I upgrade to a newer version, would that retain my previous
 configurations so that I don't need to configure everything again ?

 Yes if you select that when doing the package install.

 2.  would that smoothly replace the previous installation by itself ?


 Yes


 3.  what's the way (kindly, if you can tell the command) to upgrade ?



 http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck

 4. when should I prefer datastax's dsc to that ? (I need to install for
 production env.)

 Above

 Hope that helps.


 -
 Aaron Morton
 New Zealand
 @aaronmorton

 Co-Founder  Principal Consultant
 Apache Cassandra Consulting
 http://www.thelastpickle.com

 On 27/09/2013, at 11:01 PM, Ertio Lew ertio...@gmail.com wrote:

 I am not sure if I should use datastax's DSC or official Debian packages
 from Cassandra. How do I choose between them for a production server ?



 On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew ertio...@gmail.com wrote:


  Could you please clarify that:
 1.  when I upgrade to a newer version, would that retain my previous
 configurations so that I don't need to configure everything again ?
 2.  would that smoothly replace the previous installation by itself ?
 3.  what's the way (kindly, if you can tell the command) to upgrade ?
 4. when should I prefer datastax's dsc to that ? (I need to install for
 production env.)


 On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli rc...@eventbrite.comwrote:

 On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew ertio...@gmail.com wrote:

 How do you install Cassandra on Ubuntu  later how do you upgrade the
 installation on the node when an update has arrived ? Do you simply
 download  replace the latest tar.gz, untar it to replace the older
 cassandra files? How do you do it ? How does this upgrade process differ
 for a major version upgrade, like say switching from 1.2 series to 2.0
 series ?


 Use the deb packages. To upgrade, install the new package. Only upgrade
 a single major version. and be sure to consult NEWS.txt for any upgrade
 caveats.

 Also be aware of this sub-optimal behavior of the debian packages :

 https://issues.apache.org/jira/browse/CASSANDRA-2356

 =Rob







RE: 2.0.1 counter replicate on write error

2013-09-30 Thread Christopher Wirt
Thanks Aaron, I've added to the ticket. We were not running on TRACE
logging. 

 

From: Aaron Morton [mailto:aa...@thelastpickle.com] 
Sent: 30 September 2013 08:37
To: user@cassandra.apache.org
Subject: Re: 2.0.1 counter replicate on write error

 

ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778
CassandraDaemon.java (line 185) Exception in thread
Thread[ReplicateOnWriteStage:19,5,main]

java.lang.AssertionError: DecoratedKey(-1754949563326053382,
a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382,
aeadcec8184445d4ab631ef4250927d0) in
/disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195
3-Data.db

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName
sIterator.java:114)

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNa
mesIterator.java:62)

 

When reading from an SSTable the position returned from the -Index.db / KEYS
cache pointed to a row in the -Data.db component that was for a different
row. 

 

DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0)

Was what we were searching for

 

DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6)

Is what was found in the data component. 

 

The first part is the Token (M3 hash) the second is the key. It looks like a
collision, but it could also be a bug somewhere else. 

 

Code in SSTableReader.getPosition() points to
https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion
that is only trigger if TRACE logging is running. Can you add to the 4687
ticket and update the thread ? 

 

Cheers

 

-

Aaron Morton

New Zealand

@aaronmorton

 

Co-Founder  Principal Consultant

Apache Cassandra Consulting

http://www.thelastpickle.com

 

On 27/09/2013, at 10:50 PM, Christopher Wirt chris.w...@struq.com wrote:





Hello,

 

I've started to see a slightly worrying error appear in our logs
occasionally. We're writing at 400qps per machine and I only see this appear
every 5-10minutes.

 

Seems to have started when I switched us to using the hsha thrift server
this morning. We've been running 2.0.1 ran off the sync thrift server since
yesterday without seeing this error.  But might not be related.

 

There are some machines in another DC still running 1.2.10.

 

Anyone seen this before? Have any insight?

 

ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778
CassandraDaemon.java (line 185) Exception in thread
Thread[ReplicateOnWriteStage:19,5,main]

java.lang.AssertionError: DecoratedKey(-1754949563326053382,
a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382,
aeadcec8184445d4ab631ef4250927d0) in
/disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195
3-Data.db

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName
sIterator.java:114)

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.init(SSTableNa
mesIterator.java:62)

at
org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(Nam
esQueryFilter.java:87)

at
org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFil
ter.java:62)

at
org.apache.cassandra.db.CollationController.collectAllData(CollationControll
er.java:249)

at
org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationCont
roller.java:53)

at
org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilySto
re.java:1468)

at
org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.
java:1294)

at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332)

at
org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadComma
nd.java:55)

at
org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutat
ion.java:100)

at
org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java:
1107)

at
org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy
.java:1897)

at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:11
45)

at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:6
15)

at java.lang.Thread.run(Thread.java:724)

 



Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target

2013-09-30 Thread Miguel Angel Martin junquera
hi:



is that mean that antlr-3.2.jar is not the correct version?

what is the correct version?

In the build file,  I see that cassandra uses the jar lib at the *
${build.lib}*  folder, in this case antlr-3.2.jar


...

 target name=gen-cql2-grammar depends=check-gen-cql2-grammar
unless=cql2current
  echoBuilding Grammar
${build.src.java}/org/apache/cassandra/cql/Cql.g  .../echo
  java classname=org.antlr.Tool
   * classpath=${build.lib}/antlr-3.2.jar*
fork=true
failonerror=true
 arg value=${build.src.java}/org/apache/cassandra/cql/Cql.g /
 arg value=-fo /
 arg value=${build.src.gen-java}/org/apache/cassandra/cql/ /
  /java
/target

...



thanks in advance


Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com



2013/9/30 Aaron Morton aa...@thelastpickle.com

 It's an error in the antlr compilation, check the antlr versions.

 Cheers

 -
 Aaron Morton
 New Zealand
 @aaronmorton

 Co-Founder  Principal Consultant
 Apache Cassandra Consulting
 http://www.thelastpickle.com

 On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera 
 mianmarjun.mailingl...@gmail.com wrote:

 
 
 
 
  hi all:
 
  Environment
 
• apache-cassandra-2.0.1-src
• EC2
• Linux version 3.2.30-49.59.amzn1.x86_64
 (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3)
 
  When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I
  have this error at  the target, gen-cql2-grammar:
 
 
  gen-cql2-grammar:
   [echo] Building Grammar
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g
 
  ….
 
[java] warning(209):
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1:
 Multiple token rules can match input such as '0'..'9': INTEGER, FLOAT,
 UUID
   [java]
   [java] As a result, token(s) FLOAT,UUID were disabled for that input
   [java] warning(209):
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
 Multiple token rules can match input such as 'I': K_INSERT, K_IN,
 K_INDEX, K_INTO, IDENT, COMPIDENT
   [java]
   [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT
 were disabled for that input
   [java] warning(209):
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
 Multiple token rules can match input such as {'R', 'r'}: K_REVERSED,
 IDENT, COMPIDENT
   [java]
   [java] As a result, token(s) IDENT,COMPIDENT were disabled for that
 input
   [java] warning(209):
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
 Multiple token rules can match input such as 'T': K_LEVEL, K_TRUNCATE,
 K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT
   [java]
   [java] As a result, token(s)
 K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were
 disabled for that input
   [java] error(208):
 /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1:
 The following token definitions can never be matched because prior tokens
 match the same input:
 T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT
 
  BUILD FAILED
  /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1
 
 
 
 
 
 
  If  I  execute these targets in the next order first:
 
 
 
• ant generate-cql-html
• ant maven-ant-tasks-init
  and then  execute ant default target
 
 
• ant
 
 
  the project build succesfully.
 
  Regards.
 
 
 
 
 
  note:
 
  I do not have this issue in my mac.
 
 
 
 
 
 
  Miguel Angel Martín Junquera
  Analyst Engineer.
  miguelangel.mar...@brainsins.com
 




Re: Among Datastax community Cassandra debian package, which to choose for production install ?

2013-09-30 Thread Ken Hancock
OpsCenter should be a separate package as you would only install it on a
single node, not necessarily even one that is running Cassandra.




On Sat, Sep 28, 2013 at 2:12 PM, Ertio Lew ertio...@gmail.com wrote:

 I think both provide the same thing except Datastax Community also
 provides some extras like Opscenter, etc. But I cannot find opscenter
 installed when I installled DSC on ubuntu. Although on windows
 installation, I saw opscenter  JRE as well , so I think for DSC, there is
 no such prerequisite for Oracle JRE as required for Cassandra debain
 package, is it so ?

 Btw which is usually preferred for production installs ?

 I may need to use Opscenter but just *occasionally*.




-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]http://www.linkedin.com/in/kenhancock

[image: SeaChange International]
 http://www.schange.com/This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


Cassandra Summit EU 2013

2013-09-30 Thread Jeremy Hanna
For those in the Europe area, there will be a Cassandra Summit EU 2013 in 
London in the month of October.  On 17 October, there will be the main 
conference sessions and the 16th and 18th there will be Cassandra workshops.

http://www.datastax.com/cassandraeurope2013

The speakers have been announced and the presentation abstracts are all on 
there.  Like always, the presentations will be recorded and posted on Planet 
Cassandra, but it's great to meet and interact with people in the community - 
in my opinion that's the best part of any conference.

Anyway, just wanted to make sure people knew.

Cheers,

Jeremy

Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1

2013-09-30 Thread Miguel Angel Martin junquera
hi:


I try to test cassandra 2.0.1 and pig 0.11.1
but always when i try to load  any keyspacei have this error:

Example creating a test keyspace and data:



I have this error:




*grunt define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();*
*grunt rows = LOAD 'cql://keyspace0/test' USING
org.apache.cassandra.hadoop.pig.CqlStorage();*
*2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt -
ERROR 2999: Unexpected internal error.
InvalidRequestException(why:Undefined name key_alias in selection clause)*
*Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log*




and in the log file there are not  more info



Pig Stack Trace
---
ERROR 2999: Unexpected internal error.
InvalidRequestException(why:Undefined name key_alias in selection clause)

java.lang.RuntimeException: InvalidRequestException(why:Undefined name
key_alias in selection clause)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511)
at
org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246)
at
org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280)
at
org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151)
at
org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110)
at
org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100)
at
org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219)
at
org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75)
at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50)
at
org.apache.pig.newplan.logical.visitor.CastLineageSetter.init(CastLineageSetter.java:57)
at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635)
at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566)
at org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538)
at org.apache.pig.PigServer.registerQuery(PigServer.java:540)
at
org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970)
at
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165)
at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
at org.apache.pig.Main.run(Main.java:490)
at org.apache.pig.Main.main(Main.java:111)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
Caused by: InvalidRequestException(why:Undefined name key_alias in
selection clause)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898)
at
org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
at
org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658)
at
org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500)
... 25 more




any idea?


thanks




Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com


Re: Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1

2013-09-30 Thread Miguel Angel Martin junquera
I have the same issue using cassandraStorage()  and other dataset for this


*note*:

the example cql3 datatest are



CREATE TABLE test (
  id text PRIMARY KEY,
  title text,
  age int
);
and insert some dummy data

insert into test (id, title, age) values('1', 'child', 21);
insert into test (id, title, age) values('2', 'support', 21);
insert into test (id, title, age) values('3', 'manager', 31);
insert into test (id, title, age) values('4', 'QA', 41);
insert into test (id, title, age) values('5', 'QA', 30);
insert into test (id, title, age) values('6', 'QA', 30);




Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com



2013/9/30 Miguel Angel Martin junquera mianmarjun.mailingl...@gmail.com




 hi:


 I try to test cassandra 2.0.1 and pig 0.11.1
 but always when i try to load  any keyspacei have this error:

 Example creating a test keyspace and data:



 I have this error:




 *grunt define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();*
 *grunt rows = LOAD 'cql://keyspace0/test' USING
 org.apache.cassandra.hadoop.pig.CqlStorage();*
 *2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt -
 ERROR 2999: Unexpected internal error.
 InvalidRequestException(why:Undefined name key_alias in selection clause)*
 *Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log*




 and in the log file there are not  more info



 Pig Stack Trace
 ---
 ERROR 2999: Unexpected internal error.
 InvalidRequestException(why:Undefined name key_alias in selection clause)

 java.lang.RuntimeException: InvalidRequestException(why:Undefined name
 key_alias in selection clause)
 at
 org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511)
 at
 org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246)
 at
 org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280)
 at
 org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151)
 at
 org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110)
 at
 org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100)
 at
 org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219)
 at
 org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75)
 at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50)
 at
 org.apache.pig.newplan.logical.visitor.CastLineageSetter.init(CastLineageSetter.java:57)
 at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635)
 at
 org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566)
 at
 org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538)
 at org.apache.pig.PigServer.registerQuery(PigServer.java:540)
 at
 org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970)
 at
 org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386)
 at
 org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189)
 at
 org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165)
 at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
 at org.apache.pig.Main.run(Main.java:490)
 at org.apache.pig.Main.main(Main.java:111)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
 Caused by: InvalidRequestException(why:Undefined name key_alias in
 selection clause)
 at
 org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006)
 at
 org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983)
 at
 org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898)
 at
 org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
 at
 org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658)
 at
 org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643)
 at
 org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573)
 at
 org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500)
 ... 25 more




 any idea?


 thanks




 Miguel Angel Martín Junquera
 Analyst Engineer.
 miguelangel.mar...@brainsins.com




Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ertio Lew
Solr's data is stored on the file system as a set of index files[
http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
in itself ?  What do we need to integrate with Cassandra ?


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Peter Lin
the main reason is scalability and performance.

If your Solr indexes fit fine on a single system and doesn't need to scale
out, Cassandra/HDFS isn't necessary.


On Mon, Sep 30, 2013 at 11:50 AM, Ertio Lew ertio...@gmail.com wrote:

 Solr's data is stored on the file system as a set of index files[
 http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
 like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
 in itself ?  What do we need to integrate with Cassandra ?






Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Robert Coli
On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew ertio...@gmail.com wrote:

 Solr's data is stored on the file system as a set of index files[
 http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
 like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
 in itself ?  What do we need to integrate with Cassandra ?


Solr's index sitting on a single machine, even if that single machine can
vertically scale, is a single point of failure.

The value add of DES is that the index has the same availability
characteristics as the underlying data, because it is stored in the same
cluster.

=Rob


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ken Hancock
To clarify, solr indexes are not distributed in the same way that Cassandra
data is stored.

With Cassandra, each node receives a fraction of the keyspace (based on
your replication factor and token assignment).  With DSE Search, writes to
Cassandra are hooked and each node independently indexes its data and keeps
this index on the local file system.  If you have keyspace with RF=3 then
three nodes will index each document. Indexes, unlike from Solr, only store
the docids and the actual field values are stored in Cassandra.

When it comes to search, DSE splits up the search so in the example above
only one of those RF=3 nodes will be queried for a particular token range
so that data can be unioned across all the nodes with different token
ranges.

Not sure about Solandra, but you do need to be aware that there's a number
of Solr search options that are not supported on distribute searches/DSE
Search.

http://wiki.apache.org/solr/DistributedSearch
http://wiki.apache.org/solr/FieldCollapsing

Also, be aware that while Cassandra has knobs to allow you to get
consistent read results (CL=QUORUM), DSE Search does not. If a node drops
messages for whatever reason, outtage, mutation, etc. its solr indexes will
be inconsistent with other nodes in its replication group.



On Mon, Sep 30, 2013 at 1:06 PM, Robert Coli rc...@eventbrite.com wrote:

 On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew ertio...@gmail.com wrote:

 Solr's data is stored on the file system as a set of index files[
 http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
 like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
 in itself ?  What do we need to integrate with Cassandra ?


 Solr's index sitting on a single machine, even if that single machine can
 vertically scale, is a single point of failure.

 The value add of DES is that the index has the same availability
 characteristics as the underlying data, because it is stored in the same
 cluster.

 =Rob





-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]http://www.linkedin.com/in/kenhancock

[image: SeaChange International]
 http://www.schange.com/This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Andrey Ilinykh
 Also, be aware that while Cassandra has knobs to allow you to get
 consistent read results (CL=QUORUM), DSE Search does not. If a node drops
 messages for whatever reason, outtage, mutation, etc. its solr indexes will
 be inconsistent with other nodes in its replication group.

 Will repair fix it?


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ken Hancock
Yes.


On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh ailin...@gmail.com wrote:


 Also, be aware that while Cassandra has knobs to allow you to get
 consistent read results (CL=QUORUM), DSE Search does not. If a node drops
 messages for whatever reason, outtage, mutation, etc. its solr indexes will
 be inconsistent with other nodes in its replication group.

 Will repair fix it?




-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]http://www.linkedin.com/in/kenhancock

[image: SeaChange International]
 http://www.schange.com/This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


How to log the details of the updated data locally

2013-09-30 Thread sathiya prabhu
Hi all,

In cassandra cluster, once a write/update is successful to the particular 
instance locally, i want to log the data that is updated and its timestamps 
separately in a file.. In which class it will be more appropriate to do this.. 

To the extent i explored the codebase, it's possible to do that in Keyspace 
class (apply method) in db package.. But i don't know how to retrieve the 
timestamp details from the mutation object.. Is Keyspace class is appropriate 
for my purpose.. If yes, please kindly provide me some ideas to retrieve 
timestamp details from mutation object(RowMutation)..

Any help is appreciable.. Looking forward for your kind replies.. Thanks in 
advance..

Thank you.

Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Sávio Teles
 Solr's index sitting on a single machine, even if that single machine can
 vertically scale, is a single point of failure.


And about Cloud Solr?


2013/9/30 Ken Hancock ken.hanc...@schange.com

 Yes.


 On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh ailin...@gmail.comwrote:


 Also, be aware that while Cassandra has knobs to allow you to get
 consistent read results (CL=QUORUM), DSE Search does not. If a node drops
 messages for whatever reason, outtage, mutation, etc. its solr indexes will
 be inconsistent with other nodes in its replication group.

 Will repair fix it?




 --
 *Ken Hancock *| System Architect, Advanced Advertising
 SeaChange International
 50 Nagog Park
 Acton, Massachusetts 01720
 ken.hanc...@schange.com | www.schange.com | 
 NASDAQ:SEAChttp://www.schange.com/en-US/Company/InvestorRelations.aspx

 Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com
  | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
 LinkedIn] http://www.linkedin.com/in/kenhancock

 [image: SeaChange International]
  http://www.schange.com/This e-mail and any attachments may contain
 information which is SeaChange International confidential. The information
 enclosed is intended only for the addressees herein and may not be copied
 or forwarded without permission from SeaChange International.




-- 
Atenciosamente,
Sávio S. Teles de Oliveira
voice: +55 62 9136 6996
http://br.linkedin.com/in/savioteles
Mestrando em Ciências da Computação - UFG
Arquiteto de Software
Laboratory for Ubiquitous and Pervasive Applications (LUPA) - UFG


Segmentation fault when trying to store into cassandra...

2013-09-30 Thread Krishna Chaitanya
Hello,
   I modified a network probe which collects network packets to
store them into cassandra. So there are many packets that are coming in, I
capture the packets in the program and store them into cassandra. I am
using libQtCassandra library. The program is crashing with segmentation
fault as soon as I run it. Can someone help as to what all can go wrong
here?? Could there be a problem with row/col keys or is it some
configuration parameter or the speed at which the packets or coming? I am
not able to figure it out. Thank you.

-- 
Regards,
BNSK*.
*


Re: Segmentation fault when trying to store into cassandra...

2013-09-30 Thread Vivek Mishra
Java version issue?
Using sun jdk or open jdk?

-Vivek


On Tue, Oct 1, 2013 at 6:16 AM, Krishna Chaitanya bnsk1990r...@gmail.comwrote:

 Hello,
I modified a network probe which collects network packets to
 store them into cassandra. So there are many packets that are coming in, I
 capture the packets in the program and store them into cassandra. I am
 using libQtCassandra library. The program is crashing with segmentation
 fault as soon as I run it. Can someone help as to what all can go wrong
 here?? Could there be a problem with row/col keys or is it some
 configuration parameter or the speed at which the packets or coming? I am
 not able to figure it out. Thank you.

 --
 Regards,
 BNSK*.
 *