Repository: ambari
Updated Branches:
  refs/heads/branch-2.5 548a0b6a3 -> 7fc0a7834
  refs/heads/trunk 5567b7f0d -> 7df6416e4


AMBARI-19626. LLAP Startup: slider dependencies are not deployed by Ambari 
(aonishuk)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/7df6416e
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/7df6416e
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/7df6416e

Branch: refs/heads/trunk
Commit: 7df6416e47cdbfa08b28aa673c7a23d83f73aff0
Parents: 5567b7f
Author: Andrew Onishuk <aonis...@hortonworks.com>
Authored: Thu Jan 19 15:51:24 2017 +0200
Committer: Andrew Onishuk <aonis...@hortonworks.com>
Committed: Thu Jan 19 15:51:24 2017 +0200

----------------------------------------------------------------------
 ambari-server/src/main/resources/scripts/Ambaripreupload.py | 1 +
 1 file changed, 1 insertion(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/7df6416e/ambari-server/src/main/resources/scripts/Ambaripreupload.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/scripts/Ambaripreupload.py 
b/ambari-server/src/main/resources/scripts/Ambaripreupload.py
index bedf859..0e25a2a 100644
--- a/ambari-server/src/main/resources/scripts/Ambaripreupload.py
+++ b/ambari-server/src/main/resources/scripts/Ambaripreupload.py
@@ -413,6 +413,7 @@ with Environment() as env:
   copy_tarballs_to_hdfs(format("/usr/hdp/{stack_version}/pig/pig.tar.gz"), 
hdfs_path_prefix+"/hdp/apps/{{ stack_version_formatted }}/pig/", 
'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, 
params.user_group)
   
copy_tarballs_to_hdfs(format("/usr/hdp/{stack_version}/hadoop-mapreduce/hadoop-streaming.jar"),
 hdfs_path_prefix+"/hdp/apps/{{ stack_version_formatted }}/mapreduce/", 
'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, 
params.user_group)
   copy_tarballs_to_hdfs(format("/usr/hdp/{stack_version}/sqoop/sqoop.tar.gz"), 
hdfs_path_prefix+"/hdp/apps/{{ stack_version_formatted }}/sqoop/", 
'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, 
params.user_group)
+  
copy_tarballs_to_hdfs(format("/usr/hdp/{stack_version}/slider/lib/slider.tar.gz"),
 hdfs_path_prefix+"/hdp/apps/{{ stack_version_formatted }}/slider/", 
'hadoop-mapreduce-historyserver', params.hdfs_user, params.hdfs_user, 
params.user_group)
   
   createHdfsResources()
   
copy_zeppelin_dependencies_to_hdfs(format("/usr/hdp/{stack_version}/zeppelin/interpreter/spark/dep/zeppelin-spark-dependencies*.jar"))

Reply via email to