GitHub user krisgeus opened a pull request:

    https://github.com/apache/incubator-zeppelin/pull/868

    Make zeppelin work with CDH5.7.0

    ### What is this PR for?
    The downloadable zeppelin install wasn;t working with CDH 5.7.0. I had to 
change a few things to make it work nicely together.
    
    ### What type of PR is it?
    Bug Fix
    
    ### Todos
    
    ### What is the Jira issue?
    
https://www.mail-archive.com/users@zeppelin.incubator.apache.org/msg03471.html
    
    ### How should this be tested?
    Install local cdh5.7.0 (see 
https://github.com/krisgeus/ansible_local_cdh_hadoop)
    build zeppeling (with this patch) with the following options:
    mvn clean package -DskipTests -Pspark-1.6 -Phadoop-2.6 
-Dspark.version=1.6.0-cdh5.7.0 -Dhadoop.version=2.6.0-cdh5.7.0 -Pvendor-repo
    
    ### Screenshots (if appropriate)
    
    ### Questions:
    * Does the licenses files need update? No
    * Is there breaking changes for older versions? Didn't verify
    * Does this needs documentation? Don't think so


You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/krisgeus/incubator-zeppelin master

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/incubator-zeppelin/pull/868.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #868
    
----
commit 50717dd340082545c1194134cca91d19169f97b6
Author: Kris Geusebroek <kgeusebroek@kgmbp2015.local>
Date:   2016-04-29T15:24:34Z

    Use slf4j instead of parquet bundled one since parquet doesn't bundle it 
anymore

commit 24ea5843c7408bfd73b5c496533c38249f2e9720
Author: Kris Geusebroek <kgeusebroek@kgmbp2015.local>
Date:   2016-04-29T15:24:48Z

    method classServerUri not available in cdh5.7.0 Spark version. Only set 
config if variable is filled

commit 146b524ea9dac87733f7cfa4c7d1df6e82e917f6
Author: Kris Geusebroek <kgeusebroek@kgmbp2015.local>
Date:   2016-04-29T15:25:08Z

    Don't use extra driver classpath option since I use SPARK_CLASSPATH in 
spark-env.sh and using both is not supported

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to