Liu Shaohui created KYLIN-4048: ---------------------------------- Summary: Too long spark cube building time for too many eviction and loading for dict slices Key: KYLIN-4048 URL: https://issues.apache.org/jira/browse/KYLIN-4048 Project: Kylin Issue Type: Improvement Affects Versions: v2.5.2 Reporter: Liu Shaohui
In our cluster, a cube building costs too long. In the log for spark, we found there are too many eviction and loading for dict slices in AppendTrieDictionary. {code:java} $ grep "read slice from" spark.log | wc -l 119721 $ grep "Evict slice with key" spark.log| wc -l 119634 {code} The reason is that the memory of spark executor(4G) is not enough to hold all the slices of dict(3.3G in hdfs) in memory, which cause the bad performance of cube building. -- This message was sent by Atlassian JIRA (v7.6.3#76005)