NullPointerException in the first step of cube build
Hi, I have been using kylin 2.3.2 for a while but for few days I am getting the below error in the first step java.lang.NullPointerException at org.apache.kylin.source.hive.CreateFlatHiveTableStep.getCubeSpecificConfig(CreateFlatHiveTableStep.java:64) at org.apache.kylin.source.hive.CreateFlatHiveTableStep.doWork(CreateFlatHiveTableStep.java:69) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:125) at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:64) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:125) at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:144) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Even if I resume it multiple times the first step always fails. There is data in the synced hive table. Please help. Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: Need Suggestion for a usecase where my hive table schema changes periodically.
Hi, Thanks for your response. But in hybrid model, it says that the cube 1 will not be built anymore and will only be queried. But in my usecase the new measures and dimensions has to built for the historical period also. In that case, hybrid doesn't fit right? Please clarify Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: [DISCUSS] Change on Apache Kylin logo
Yes, the new logo looks like a fine improvement from the old. Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: Re:Query Problem
Hi, Even I get empty values in the metric column if I give select * and I am using kylin version 2.3.2. What will be the work around to get the raw metric values. I tried "select dimensions, sum(metric_col) group by dimension", but this gives the grouped data and not the raw data. Does that mean one cannot get the raw metric value from kylin? Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
CUBE_NOT_CONTAIN_ALL_COLUMN Error while Joining two cubes
Hi, After building the cube 1 for some time, there are few dimensions added to the fact table and so I create a second cube with the new dimensions and tried to join the two cubes on the common column. But I get the below error, No model found for OLAPContext, CUBE_NOT_CONTAIN_ALL_COLUMN. What does it mean? Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: Hive table design (multiple fact tables or rolled up)
Hi, I know this is the old post, but I don't think kylin 2.x has the facility to add more than 1 fact table. Is this correct. Please clarify. Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Need Suggestion for a usecase where my hive table schema changes periodically.
Hi, I have a use case and need suggestion. I have a hive table where in the future there can be new dimensions added. Currently I have a cube on that hive table with say 10 dimensions computed. Now as and when the new dimensions are added (say 3 dimensions are added), we have to refresh with the historical data. That means, I have to rebuild a new cube again with (10 "old"+ 3 "new") 13 dimensions. I feel this will be inefficient process, as again in the future new dimensions can added and also because we will refreshing again the whole set. How to handle this type of usecase? I have few thoughts and I need your suggestion in terms of time and performance of building the cube. 1. I have cube_1 with old dimensions computed. Now I will create cube_2 with the new dimensions and build for historical dataset. Then in the kylin query, I use these two cubes by joining and get the result. 2. Can I use hybrid concept in this case. Is it that hybrid can be used only when you don't need to refresh cube_2 with historical data? Please help. Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
The algorithm used is "layer", but why in-mem step while building the cube?
Hi, I have a confusion in the algorithms kylin has for different data. There are: 1. By Layer 2. Fast Cubing (MR) 3. Fast Cubing with spark >From what I have researched, fast cubing algorithm is the "in-mem" algorithm. Now, in my production, I am using "layer" cubing algorithm, but I see a step called "Build Cube In-Mem". 1. What does this step do? 2. So by-layer also uses in-mem at this step? Please clarify my doubt. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: What is the meaning of "executablePO is null"
Thanks for your response. We are using kylin 2.3.2 version. Is it has to do with the zookeeper because one of our zookeeper node is dead. Thanks, Neethu -- Sent from: http://apache-kylin.74782.x6.nabble.com/
What is the meaning of "executablePO is null"
Hi, I have been getting null pointer exception and the error says executablePO is null 2018-08-24 06:42:27,653 ERROR [http-bio-7070-exec-15] controller.JobController:146 : java.lang.NullPointerException... Can anyone please explain whats happening? Does that mean kylin does not find any task to execute and hence throw this error? If so, how can I handle this. If I resume the job build, it then starts running properly. Please help. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Kylin 2.3.2 - I am seeing this NPE error in the logs even if there is no job building, org.apache.kylin.rest.exception.InternalErrorException: java.lang.NullPointerException
Hi, I am using Kylin 2.3.2 and I see the following error appears in the log even if there is no job building org.apache.kylin.rest.exception.InternalErrorException: java.lang.NullPointerException. Is this the bug in Kylin? Please clarify. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Kylin 2.3.2 - Job Status is not getting updated even after the progress is 100 %
Hi, I am using kylin 2.3.2 and I see that kylin does not update the status even after the progress is 100%. Is this a bug or something else? The job runs continuously until I drop it. The results are saved to HBASE and I am able to query it. Please clarify. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: How to configure the default data type measure decimal(19, 4)?
So I have restored and restarted kylin and I can see that the data type has been changed to decimal(19,6). When I try to build the cube it says, it gives me an error saying that "inconsistent cubedesc_signature". Is there anything I should do for that? -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: How to configure the default data type measure decimal(19, 4)?
Hi, So I have changed the json file of my cube after taking the backup and now do I have to restore to see the change? -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: How to configure the default data type measure decimal(19, 4)?
One doubt, when I am creating the cube, I see the default type of the measure is decimal(19,4). Will it change after I build the cube? -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: How to configure the default data type measure decimal(19, 4)?
Yes, I have sync the hive table again and seems to take the default decimal(19,4). -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Re: How to configure the default data type measure decimal(19, 4)?
Hi, Thanks for your response. I have configured the kylin.properties file with kylin.source.hive.default-decimal-scale=6 kylin.source.hive.default-decimal-precision=19 and restarted kylin. But still I get the default as decimal(19,4). Please help. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
How to configure the default data type measure decimal(19, 4)?
Currently, I am not able to select the data type of the measure. I want to have decimal(19,6) but in my current version 2.2.0, kylin defaults the data type to decimal(19,4). How can I change this? Please respond. -- Sent from: http://apache-kylin.74782.x6.nabble.com/
Clarity on Dimensions and aggregate group
Hi, I am trying to use Apache Kylin for my work and need some clarity. Say, I have 10 dimensions but I have 2 AGG with 6 dimensions. From what I understand, the cuboids are created from the AGG. How does the row key will be created to be stored to HBASE? what about the rest 4 dimesions in the data model? Please clarify. -- Sent from: http://apache-kylin.74782.x6.nabble.com/