Good day,

consider the following query:

WITH aggregation(
    SELECT
           a.*,
          (SELECT array_agg(b.*) FROM b WHERE b.a_id = a.id) as "bs",
          (SELECT array_agg(c.*) FROM c WHERE c.a_id = a.id) as "cs",
          (SELECT array_agg(d.*) FROM d WHERE d.a_id = a.id) as "ds",
          (SELECT array_agg(e.*) FROM d WHERE e.a_id = a.id) as "es"
    FROM a WHERE a.id IN (<some big list, ranging from 20-180 entries)
)
SELECT to_jsonb(aggregation.*) as "value" FROM aggregation;

Imagine that for each "a" there exists between 5-100 "b", "c", "d" and "e" which makes the result of this pretty big (worst case: around 300kb when saved to a text file). I noticed that adding the "to_jsonb" increases the query time by 100%, from 9-10ms to 17-23ms on average. This may not seem slow at all but this query has another issue: on an AWS Aurora Serverless V2 instance we are running into a RAM usage of around 30-50 GB compared to < 10 GB when using a simple LEFT JOINed query when under high load (> 1000 queries / sec). Furthermore the CPU usage is quite high.

Is there anything I could improve? I am open for other solutions but I am wondering if I ran into an edge case of "to_jsonb" for "anonymous records" (these are just rows without a defined UDT) - this is just a wild guess though. I am mostly looking to decrease the load (CPU and memory) on Postgres itself. Furthermore I would like to know why the memory usage is so significant. Any tips on how to analyze this issue are appreciated as well -  my knowledge is limited to being average at interpreting EXPLAIN ANALYZE results.

Here's a succinct list of the why's, what I have found out so far and solution I already tried/ don't want to consider:

- LEFT JOINing potentially creates a huge resultset because of the cartesian product, thats a nono - not using "to_jsonb" is sadly also not possible as Postgres' array + record syntax is very unfriendly and hard to parse (it's barely documented if at all and the quoting rules are cumbersome, furthermore I lack column names in the array which would make the parsing sensitive to future table changes and thus cumbersome to maintain) in my application - I know I could solve this with a separate query for a,b,c,d and e while "joinining" the result in my application, but I am looking for another way to do this (bear with me, treat this as an academic question :)) - I am using "to_jsonb" to simply map the result to my data model via a json mapper - EXPLAIN ANALYZE is not showing anything special when using "to_jsonb" vs. not using it, the outermost (hash) join just takes more time - is there a more granular EXPLAIN that shows me the runtime of functions like "to_jsonb"? - I tried an approach where b,c,d,e where array columns of UDTs: UDTs are not well supported by my application stack (JDBC) and are generally undesireable for me (because of a lack of migration possibilities) - I don't want to duplicate my data into another table (e.g. that has jsonb columns) - MATERIALIZED VIEWS are also undesirable as the manual update, its update is non-incremental which would make a refresh on a big data set take a long time - split the query into chunks to reduce the IN()-statement list size makes no measurable difference - I don't want to use JSONB columns for b,c,d and e because future changes of b,c,d or e's structure (e.g. new fields, changing a datatype) are harder to achieve with JSONB and it lacks constraint checks on insert (e.g. not null on column b.xy)

Kind regards and thank you for your time,
Nico Heller

P.S: Sorry for the long list of "I don't want to do this", some of them are not possible because of other requirements




Reply via email to