atically tune your CMS
initiating occupancy, or you'd probably see horrible, horrible pauses.
On Tue, Feb 13, 2018 at 8:44 AM, James Rothering
mailto:jrother...@codojo.me>> wrote:
Wow, an 84GB heap! Would you mind disclosing the kind of data requirements
behind this choice? And what
In most cases, Cassandra is pretty efficient about memory usage.
However, if your use case does require/need/demand more memory for your
workload, I would not hesitate to use heap > 32 GB.
FYI, we have configured our heap for 84 GB.
However there's more tuning that we have done beyond just the hea
#x27;t happen. Seems like it's possibly not looking in the
correct location for data directories. Try setting CASSANDRA_INCLUDE=http://cassandra.in.sh>> prior to running the script?
e.g:
CASSANDRA_INCLUDE=/cassandra.in.sh<http://cassandra.in.sh>
sstableexpiredblockers ae raw_log
Thanks Kurt and Kenneth.
Now only if they would work as expected.
node111.ord.ae.tsg.cnvr.net:/ae/disk1/data/ae/raw_logs_by_user-f58b9960980311e79ac26928246f09c1>ls
-lt | tail
-rw-r--r--. 1 vchadoop vchadoop286889260 Sep 18 14:14 mc-1070-big-Index.db
-rw-r--r--. 1 vchadoop vchadoop12
pairs” and “Timestamp overlap” sections might be
of use.
-B
On Jan 25, 2018, at 11:05 AM, Thakrar, Jayesh
mailto:jthak...@conversantmedia.com>> wrote:
Wondering if I can get some pointers to what's happening here and why sstables
that I think should be expired are not being dropped
Wondering if I can get some pointers to what's happening here and why sstables
that I think should be expired are not being dropped.
Here's the table's compaction property - note also set
"unchecked_tombstone_compaction" to true.
compaction = {'class':
'org.apache.cassandra.db.compaction.TimeW
Wondering what is the purpose - is it to get some insight into the cluster?
Besides the logs themselves, another approach that many and I have taken is to
pull the JMX metrics from Cassandra and push them to an appropriate
metrics/timeseries system.
Here's one approach of getting JMX metrics ou
Just curious - aside from the "sleep", is this all not part of the shutdown
command?
Is this an "opportunity" to improve C*?
Having worked with RDBMSes, Hadoop and HBase, stopping communication, flushing
memcache (HBase), and relinquishing ownership of data (HBase) is all part of
the shutdown pr
What you have is sequential and hence sequential processing.
Also Spark/Scala are not parallel programming languages.
But even if they were, statements are executed sequentially unless you exploit
the parallel/concurrent execution features.
Anyway, see if this works:
val (RDD1, RDD2) = (JavaFunc
Yep, similar symptoms - but no, there's no OOM killer
Also, if you look in the gc log around the time of failure, the heap memory was
much below the 16 GB limit.
And if I look at the 2nd last GC log before the crash, here’s what we see.
And you will notice that cleaning up the 4 GB Eden (along wi
executed are not needed anymore, so its ok for the
older prepared statements to be purged.
All the same, I will do some analysis on the prepared statements table.
Thanks for the tip/pointer!
On 8/22/17, 5:17 PM, "Alain Rastoul" wrote:
On 08/22/2017 05:39 PM, Thakrar, Jayesh wrote:
s.LZ4Compressor'}
AND crc_check_chance = 1.0
AND dclocal_read_repair_chance = 0.1
AND default_time_to_live = 0
AND gc_grace_seconds = 864000
AND max_index_interval = 2048
AND memtable_flush_period_in_ms = 0
AND min_index_interval = 128
AND read_repair_chance =
Hi All,
We are somewhat new users to Cassandra 3.10 on Linux and wanted to ping the
user group for their experiences.
Our usage profile is batch jobs that load millions of rows to Cassandra every
hour.
And there are similar period batch jobs that read millions of rows and do some
processing,
Can you provide more details.
E.g. table structure, the app used for the query, the query itself and the
error message.
Also get the output of the following commands from your cluster nodes (note
that one command uses "." and the other "space" between keyspace and tablename)
nodetool -h tables
Thanks Sander - this helps get a better understanding!
From: Fridtjof Sander
Date: Friday, June 30, 2017 at 4:19 AM
To: Vladimir Yudovin , "Thakrar, Jayesh"
Cc: Subroto Barua , Zhongxiang Zheng
, "user@cassandra.apache.org"
Subject: Re: Question: Behavior of inserting a
ECT * FROM test.test ;
k | v
---+-
1 | [3]
// = EXPECTED RESULT =
From: Subroto Barua
Date: Monday, June 19, 2017 at 11:09 PM
To: "Thakrar, Jayesh" , Subroto Barua
, Zhongxiang Zheng
Cc: "user@cassandra.apache.org"
Subject: Re: Question: Behavior of in
Subroto,
Cassandra docs say otherwise.
Writing list data is accomplished with a JSON-style syntax. To write a record
using INSERT, specify the entire list as a JSON array. Note: An INSERT will
always replace the entire list.
Maybe you can elaborate/shed some more light?
Thanks,
Jayesh
Lists
Asad,
The rest of the 42 GB of memory on your server is used by the filesystem buffer
cache - see the "cached" column and the -/+ buffers/cache line.
The OS (Linux) uses all free memory for filesystem buffer cache and if
applications need memory, will relinquish it appropriately.
To see the ac
Thank you Kurt - that makes sense.
Will certainly reduce it to 1024.
Greatly appreciate your quick reply.
Thanks,
Jayesh
From: kurt greaves
Sent: Wednesday, June 14, 5:53 PM
Subject: Re: Question: Large partition warning
To: Fay Hou [Data Pipeline & Real-time Analytics]
Cc: Thakrar, Ja
We are on Cassandra 2.2.5 and I am constantly seeing warning messages about
large partitions in system.log even though our setting for partition warning
threshold is set to 4096 (MB).
WARN [CompactionExecutor:43180] 2017-06-14 20:02:13,189
BigTableWriter.java:184 - Writing large partition
tsg
"polling" cycle overhead.
Furthermore, zk does not have the "overhead" of other things that Cassandra
does.
Honestly I am not familiar with Paxos and stuff, so can't speak to it.
On 5/25/17, 3:40 PM, "Jan Algermissen" wrote:
Hi Jayesh,
On
Hi Jan,
I would suggest looking at using Zookeeper for such a usecase.
See http://zookeeper.apache.org/doc/trunk/recipes.html for some examples.
Zookeeper is used for such purposes in Apache HBase (active master), Apache
Kafka (active controller), Apache Hadoop, etc.
Look for the "Leader Elect
me data everywhere increases.
C*heers,
---
Alain Rodriguez - @arodream -
al...@thelastpickle.com<mailto:al...@thelastpickle.com>
France
The Last Pickle - Apache Cassandra Consulting
http://www.thelastpickle.com
2017-04-21 15:54 GMT+02:00 Thakrar, Jayesh
mailto:jthak...@conver
e 2 failure cases I mentioned earlier, the only other way data
can become inconsistent is error when replicating the data in the background.
Does Cassandra have a retry policy for internal replication? Is there a setting
to change it?
On Thu, Apr 6, 2017 at 10:54 PM, Thakrar,
I had asked a similar/related question - on how to carry out repair, etc and
got some useful pointers.
I would highly recommend the youtube video or the slideshare link below (both
are for the same presentation).
https://www.youtube.com/watch?v=1Sz_K8UID6E
http://www.slideshare.net/DataStax/rea
To: "user@cassandra.apache.org"
Subject: RE: [Cassandra 3.0.9] Cannot allocate memory
JVM config is as below:
-Xms16G
-Xmx16G
-Xmn3000M
What I need to check in dmesg?
From: Thakrar, Jayesh [mailto:jthak...@conversantmedia.com]
Sent: 23 March 2017 03:39
To: Abhishek Kumar Maheshwari ;
user@
And what is the configured max heap?
Sometimes you may also be able to see some useful messages in "dmesg" output.
Jayesh
From: Abhishek Kumar Maheshwari
Sent: Wednesday, March 22, 2017 5:05:14 PM
To: Thakrar, Jayesh; user@cassandra.apache.org
S
Is/are the Cassandra server(s) shared?
E.g. do they run mesos + spark?
From: Abhishek Kumar Maheshwari
Date: Wednesday, March 22, 2017 at 12:45 AM
To: "user@cassandra.apache.org"
Subject: [Cassandra 3.0.9] Cannot allocate memory
Hi all,
I am using Cassandra 3.0.9. while I am adding new server
on threads being used during repair (according to compactionstats).
thank you also for your link recommendations. i will go through them.
On Sat, 2017-03-18 at 16:54 +0000, Thakrar, Jayesh wrote:
You changed compaction_throughput_mb_per_sec, but did you also increase
concurrent_compactors?
In re
You changed compaction_throughput_mb_per_sec, but did you also increase
concurrent_compactors?
In reference to the reaper and some other info I received on the user forum to
my question on "nodetool repair", here are some useful links/slides -
https://www.datastax.com/dev/blog/repair-in-cassa
Thank you Eric for helping out.
The reason I sent the question a second time is because I did not see my
question and the first reply from the usergroup.
After I sent the question a second time, I got a personal flame from somebody
else too and so examined my "spam" folders and that's where I fou
I understand that the nodetool command connects to a specific server and for
many of the commands, e.g. "info", "compactionstats", etc, the information is
for that specific node.
While for some other commands like "status", the info is for the whole cluster.
So is "nodetool repair" that operates
I understand that the nodetool command connects to a specific server and for
many of the commands, e.g. "info", "compactionstats", etc, the information is
for that specific node.
While for some other commands like "status", the info is for the whole cluster.
So is "nodetool repair" that operates
22:54, Thakrar, Jayesh
mailto:jthak...@conversantmedia.com>> wrote:
I have a situation where the off-heap memory is bloating the jvm process
memory, making it a candidate to be killed by the oom_killer.
My server has 256 GB RAM and Cassandra heap memory of 16 GB
Below is the output of "no
I have a situation where the off-heap memory is bloating the jvm process
memory, making it a candidate to be killed by the oom_killer.
My server has 256 GB RAM and Cassandra heap memory of 16 GB
Below is the output of "nodetool info" and nodetool compactionstats for a
culprit table which causes
s
, "user@cassandra.apache.org"
Subject: Re: OOM on Apache Cassandra on 30 Plus node at the same time
I was looking at nodetool info across all nodes. Consistently JVM heap used is
~ 12GB and off heap is ~ 4-5GB.
________
From: Thakrar, Jayesh
Sent: Saturday, M
STCS . LCS is not an
option for us as we have frequent updates.
Thanks,
Shravan
________
From: Thakrar, Jayesh
Sent: Friday, March 3, 2017 3:47:27 PM
To: Joaquin Casares; user@cassandra.apache.org
Subject: Re: OOM on Apache Cassandra on 30 Plus node at the same time
Had been fighting a similar battle, but am now over the hump for most part.
Get info on the server config (e.g. memory, cpu, free memory (free -g), etc)
Run "nodetool info" on the nodes to get heap and off-heap sizes
Run "nodetool tablestats" or "nodetool tablestats ." on the
key large tables
Ess
mamd
All are basically best effort.
Commit logs get corrupt and only flush periodically.
Bits rot on disk and while crossing networks network
Read repair is async and only happens randomly
Hinted handoff stops after some time and is not guarenteed.
On Monday, February 27, 2017, Thakrar, Jayesh
e gc_grace_seconds
after the data has been TTL'ed, you won't need an extra repair.
2017-02-27 18:29 GMT+01:00 Oskar Kjellin
mailto:oskar.kjel...@gmail.com>>:
Are you running multi dc?
Skickat från min iPad
27 feb. 2017 kl. 16:08 skrev Thakrar, Jayesh
mailto:jthak...@conversantmed
Suppose I have an application, where there are no deletes, only 5-10% of rows
being occasionally updated (and that too only once) and a lot of reads.
Furthermore, I have replication = 3 and both read and write are configured for
local_quorum.
Occasionally, servers do go into maintenance.
I und
41 matches
Mail list logo