Re:Hadoop+Cassandra

2013-03-11 Thread Шамим
http://frommyworkshop.blogspot.ru/2012/07/single-node-hadoop-cassandra-pig-setup.html > I use Cassandra 1.2.2 and Hadoop 1.0.4 > > 2013/3/11 Renato Marroquín Mogrovejo > >> Hi there, >> >> Check this out [1]. It´s kinda old but I think it will help you get started. >> >> Renato M. >> >> [1] htt

Pig_cassandra : Map task only running on one node

2013-02-25 Thread Шамим
Dear users, We have got very strange beheviour of hadoop cluster after upgrading Cassandra from 1.1.5 to Cassandra 1.2.1. We have 5 nodes cluster of Cassandra, where three of them are hodoop slaves. Now when we are submitting job through Pig script, only one map task runs on one of the hadoop

Re: ETL Tools to transfer data from Cassandra into other relational databases

2012-12-14 Thread Шамим
Hello Chin, you can extract delta using pig script and save it in another CF in Cassandra. By using Pentaho kettle you can then load the data from the CF to RDBMS. Pentaho Kettle is open source project. All of the process you can automate through Azkaban or Ozzie. Kafka is also an alternatives

Re: Filter data on row key in Cassandra Hadoop's Random Partitioner

2012-12-12 Thread Шамим
You can use Apache PIG to load data and filter it by row key, filter in pig is very fast. Regards Shamim 11.12.2012, 20:46, "Ayush V." : > I'm working on Cassandra Hadoop intergration (MapReduce). We have used Random > Partioner to insert data to gain faster write. Now we have to read that data

Re:Data not replicating to all datacenters

2012-12-03 Thread Шамим
x27;t understand if we have > specified Options: [dc1:3, dc2:3], surely after a while all the data > will be on every server? > > Thanks, > > Owen > > On 3 December 2012 14:06, Шамим wrote: > >> Hello Owen, >> Seems you did not configure token for al

Data not replicating to all datacenters

2012-12-03 Thread Шамим
Hello Owen, Seems you did not configure token for all nodes correctly. See the section Calculating Tokens for multiple data centers here http://www.datastax.com/docs/0.8/install/cluster_init Best regards Shamim --- On Mon, Dec 3, 2012 at 4:42 PM, Owen Davies wrote: We have a 2 data cent

Re: Changing compression_parameters of exsisting CF

2012-11-27 Thread Шамим
te a ticket on https://issues.apache.org/jira/browse/CASSANDRA ? > > Thanks > > - > Aaron Morton > Freelance Cassandra Developer > New Zealand > > @aaronmorton > http://www.thelastpickle.com > > On 27/11/2012, at 2:40 AM, Шамим wrote: > >>

Changing compression_parameters of exsisting CF

2012-11-26 Thread Шамим
Hello users,   faced very strange behaviour when chnaging compression_parameters of exisiting CF. After changing the compaction strategy, compression_strategy returning back to the "SnappyCompressor". Using version 1.1.5. [cqlsh 2.2.0 | Cassandra 1.1.5 | CQL spec 2.0.0 | Thrift protocol 19.32.0] I 

java.io.IOException: InvalidRequestException(why:Expected 8 or 0 byte long for date (4)) when inserting data to CF with compound key from pig

2012-11-11 Thread Шамим
Hello All, we are using pig (pig-0.10.0) to store some data in CF with compound key. Cassandra version is 1.1.15. Here is the script for creating CF CREATE TABLE clicks_c ( user_id varchar, time timestamp, url varchar, PRIMARY KEY (user_id, time) ) WITH COMPACT STORAGE; Here is descript