Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Giampaolo Trapasso
"If not exists" was an oversight of a previous test. Removing it solved the problem. Thanks a lot, Jim! giampaolo 2016-02-09 1:21 GMT+01:00 Jim Ancona : > The "if not exists" in your INSERT means that you are incurring a > performance hit by using Paxos. Do you need that? Have you tried your tes

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Jack Krupansky
Bucket size is not disclosed. My recommendation is that partitions not be more than about 10 MB (some people say 100MB or 50MB.) I think I'd recommend a smaller chunk size, like 128K or 256K. I would note that Mongo's GridFS uses 256K chunks. I don't know enough about the finer nuances of Cassand

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Giampaolo Trapasso
Sorry Jack for my poor description, I write 600 times the same array of 1M of bytes to make my life easier. This allows me to simulate a 600Mb file. It's just a simplification. Instead of generating 600Mb random array (or reading a real 600Mb file), and dividing it into 600 chunks, I write the same

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Jack Krupansky
I'm a little lost now. Where are you specifying chunk size, which is what should be varying, as opposed to blob size? And what exactly is the number of records? Seems like you should be computing number of chunks from blob size divided by chunk size. And it still seems like you are writing the same

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Jim Ancona
The "if not exists" in your INSERT means that you are incurring a performance hit by using Paxos. Do you need that? Have you tried your test without it? Jim

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Giampaolo Trapasso
I write at every step MyConfig.blobsize number of bytes, that I configured to be from 10 to 100. This allows me to "simulate" the writing of a 600Mb file, as configuration on github ( https://github.com/giampaolotrapasso/cassandratest/blob/master/src/main/resources/application.conf *)* G

Re: Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Jack Krupansky
You appear to be writing the entire bob on each chunk rather than the slice of the blob. -- Jack Krupansky On Mon, Feb 8, 2016 at 1:45 PM, Giampaolo Trapasso < giampaolo.trapa...@radicalbit.io> wrote: > Hi to all, > > I'm trying to put a large binary file (> 500MB) on a C* cluster as fast as > I

Writing a large blob returns WriteTimeoutException

2016-02-08 Thread Giampaolo Trapasso
Hi to all, I'm trying to put a large binary file (> 500MB) on a C* cluster as fast as I can but I get some (many) WriteTimeoutExceptions. I created a small POC that isolates the problem I'm facing. Here you will find the code: https://github.com/giampaolotrapasso/cassandratest, *Main details abo