When time series label values are substituted for new ones regularly and 
the old series becomes active with no new data points. Keeping ingestion 
rate constant, what is the performance impact of high churn rate on 
ingestion, querying, compaction, and other operations.

For example, container_memory_usage_bytes gives memory usage of a pod's 
containers in kubernetes. It contains a *pod* and *container *labels 
contains pod name and container name, respectively, these may change often 
due to change in deployment, auto scaling, restart of crashed pods or for 
load balancing for a large scale real time system.

I believe the inactive time series should be regularly flushed out of 
active memory. High cardinality caused by the high churn rate should not 
cause high RAM usage. 

Compaction operation runs in background packing time series blocks into 
bigger blocks. If ingestion rate is constant, I don't foresee an impact of 
compaction runtime or resource usage.

Querying for a specific pod/container (most commonly container) would touch 
blocks within the time horizon of query. Unless query spans a large time 
range, I don't think there should be a significant impact on query run 
time  and use too much CPU/RAM.

Is my understanding correct? Are there other performance considerations for 
a high churn rate?




 

 
 

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to prometheus-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/69beccfa-907d-459a-bbfa-4bd80b258ee3n%40googlegroups.com.

Reply via email to