Hello all,

We are in the process of evaluating Kylin for use as an OLAP engine. To
that end, we are trying to get a minimum viable setup with a representative
sample of our data in order to gather performance metrics. We have kylin
running against a 10 node cluster, the provided cubes build successfully
and the system seems functional. Attempting to build a simple cube against
our data results in an OutOfMemoryError in the kylin server process (so far
we have given it up to a 46 gig heap). I was wondering if you could give me
some guidance as to likely causes, any configurations I'm likely to have
missed before I start diving into the source. I have changed the
"dictionary" setting to false, as recommended for high-cardinality
dimensions, but have not changed configuration significantly apart from
that.

For reference, the sizes of the hive tables we're building the cubes from
dimension table: 25,399,061 rows
fact table: 270,940,921 rows

(And as a note, there are no pertinent log messages except to indicate that
it is in the Build Dimension Dictionary step)

Thank you,
sam bock

Reply via email to