HI, I tried to build a cube on a 100 million data set. When I set 9 fields to build the cube with 10 cores. It nearly coast me a whole day to finish the job. At the same time, it generate almost 1”TB“ data in the "/tmp“ folder. Could we refer to the ”fast cube“ algorithm in apache Kylin To make the cube builder more quickly???
even run the group by first and generate the cube is more quilk. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Imporvement-the-cube-with-the-Fast-Cubing-In-apache-Kylin-tp26499.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org