I suggest we have a parallel implementation by removing kettle and when it
stabilises we make kettle deprecated and then remove it
On Mon, 26 Dec 2016 at 1:25 PM, Liang Chen <chenliang6...@gmail.com> wrote:

> Hi
>
> Thanks you started a good discussion.
>
> For 1 and 2, i agree.  In 1.0.0 version, will support it.
> For 3 : Need keep the parameter, users can specify carbon's store location.
> If users don't specify the carbon store location, can use the default
> location what you suggested: "spark.sql.warehouse.dir"(spark2) or
> "hive.metastore.warehouse.dir"(spark1)
>
> Regards
> Liang
>
> QiangCai wrote
> > hi all,
> >
> >   I suggest to simplify deployment of CarbonData as following.
> >   1. remove kettle dependency completely, no need to deploy
> > "carbonplugins" folder on each node, no need to set "carbhon.kettle.home"
> >   2. remove carbon.properties file from executor side, pass CarbonData
> > configuration to executor side from driver side
> >   3. use "spark.sql.warehouse.dir"(spark2) or
> > "hive.metastore.warehouse.dir"(spark1) instead of "carbon.storelocation"
> >
> >   So we will just need to deploy CarbonData jars on cluster mode in the
> > future.
> >
> >   What's your opinion?
> >
> > Best Regards
> > David Cai
>
>
>
>
>
> --
> View this message in context:
> http://apache-carbondata-mailing-list-archive.1130556.n5.nabble.com/Discussion-Simplify-the-deployment-of-carbondata-tp5000p5006.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>

Reply via email to