+1

On Tue, Jul 2, 2024 at 7:19 AM yangjie01 <yangji...@baidu.com.invalid>
wrote:

> I have manually attempted to only modify the `assembly/pom.xml` and
> examined the results of executing `dev/make-distribution.sh --tgz`. The
> `spark-connect_2.13-4.0.0-SNAPSHOT.jar` is indeed included in the jars
> directory. However, if rearranging the directories would result in a
> clearer project structure, I believe that would also be a viable approach.
>
>
>
> *发件人**: *Hyukjin Kwon <gurwls...@apache.org>
> *日期**: *2024年7月2日 星期二 12:00
> *收件人**: *yangjie01 <yangji...@baidu.com.invalid>
> *抄送**: *dev <dev@spark.apache.org>
> *主题**: *Re: [外部邮件] [DISCUSS] Move Spark Connect server to builtin package
> (Client API layer stays external)
>
>
>
> My concern is that the `connector` directory is really for
> external/optional packages (and they aren't included in assembly IIRC).. so
> I am hesitant to just change the assembly.
> The actual changes are not quite large but it moves the files around.
>
>
>
> On Tue, 2 Jul 2024 at 12:23, yangjie01 <yangji...@baidu.com.invalid>
> wrote:
>
> I'm supportive of this initiative. However, if the purpose is just to
> avoid the additional `--packages` option, it seems that making some
> adjustments to the `assembly/pom.xml` could potentially meet our goal. Is
> it really necessary to restructure the code directory?
>
>
>
> Jie Yang
>
>
>
> *发件人**: *Hyukjin Kwon <gurwls...@apache.org>
> *日期**: *2024年7月2日 星期二 08:19
> *收件人**: *dev <dev@spark.apache.org>
> *主题**: *[外部邮件] [DISCUSS] Move Spark Connect server to builtin package
> (Client API layer stays external)
>
>
>
> Hi all,
>
> I would like to discuss moving Spark Connect server to builtin package.
> Right now, users have to specify —packages when they run Spark Connect
> server script, for example:
>
> ./sbin/start-connect-server.sh --jars `ls 
> connector/connect/server/target/**/spark-connect*SNAPSHOT.jar`
>
> or
>
> ./sbin/start-connect-server.sh --packages 
> org.apache.spark:spark-connect_2.12:3.5.1
>
> which is a little bit odd that sbin scripts should provide jars to start.
>
> Moving it to builtin package is pretty straightforward because most of
> jars are shaded, and the impact would be minimal, I have a prototype here
> apache/spark/#47157
> <https://mailshield.baidu.com/check?q=%2fcy8z7%2fYZNJHeBln9msQEVrqe7rk5HGMbpa7h9lbvgUO6rNqy1kPDW2sYiQgCGHS>.
> This also simplifies Python local running logic a lot.
>
> User facing API layer, Spark Connect Client, stays external but I would
> like the internal/admin server layer, Spark Connect Server, implementation
> to be built in Spark.
>
> Please let me know if you have thoughts on this!
>
>
>
>

Reply via email to