I looks like it's still trying to grab the 2.1.7 RPMs.

1) Can you confirm what is in /etc/yum.repos.d/HDP.repo ? This file is
generated by Ambari based on the Base URLs you enter and should reflect the
2.1.3 urls that you entered during the wizard.
2) "yum clean all"
3) "yum info hadoop" and see what version it returns.


On Wed, Nov 5, 2014 at 4:16 AM, guxiaobo1982 <guxiaobo1...@qq.com> wrote:

> I tried
> http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.1.3.0 and
> http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.1.5.0 for
> CENTOS6 both,
>
> can with error like this
>
>
> stderr:   /var/lib/ambari-agent/data/errors-302.txt
>
> 2014-11-05 17:12:21,987 - Error while executing command 'install':
> Traceback (most recent call last):
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 111, in execute
>     method(env)
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.1/services/FALCON/package/scripts/falcon_client.py",
>  line 25, in install
>     self.install_packages(env)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 167, in install_packages
>     Package(name)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
> line 148, in __init__
>     self.env.run()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 149, in run
>     self.run_action(resource, action)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 115, in run_action
>     provider_action()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/__init__.py",
>  line 40, in action_install
>     self.install_package(package_name)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/yumrpm.py",
>  line 36, in install_package
>     shell.checked_call(cmd)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", 
> line 35, in checked_call
>     return _call(command, logoutput, True, cwd, env, preexec_fn, user, 
> wait_for_finish, timeout)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", 
> line 90, in _call
>     raise Fail(err_msg)
> Fail: Execution of '/usr/bin/yum -d 0 -e 0 -y install falcon' returned 1. 
> Error Downloading Packages:
>   hadoop-yarn-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-yarn-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] 
> No more mirrors to try.
>   bigtop-jsvc-1.0.10-1.el6.x86_64: failure: 
> bigtop-jsvc/bigtop-jsvc-1.0.10-1.el6.x86_64.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>   hadoop-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>   hadoop-client-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-client-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   hadoop-hdfs-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-hdfs-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] 
> No more mirrors to try.
>   hadoop-mapreduce-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-mapreduce-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   zookeeper-3.4.5.2.1.7.0-784.el6.noarch: failure: 
> zookeeper/zookeeper-3.4.5.2.1.7.0-784.el6.noarch.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   falcon-0.5.0.2.1.7.0-784.el6.noarch: failure: 
> falcon/falcon-0.5.0.2.1.7.0-784.el6.noarch.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>
> stdout:   /var/lib/ambari-agent/data/output-302.txt
>
> 2014-11-05 17:12:13,977 - Execute['mkdir -p /tmp/HDP-artifacts/;     curl -kf 
> -x "" --retry 10     
> http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o 
> /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] {'environment': ..., 
> 'not_if': 'test -e /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip', 
> 'ignore_failures': True, 'path': ['/bin', '/usr/bin/']}
> 2014-11-05 17:12:13,987 - Skipping Execute['mkdir -p /tmp/HDP-artifacts/;     
> curl -kf -x "" --retry 10     
> http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o 
> /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] due to not_if
> 2014-11-05 17:12:13,998 - Repository['HDP-2.1'] {'base_url': 
> 'http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.1.3.0', 
> 'action': ['create'], 'components': ['HDP', 'main'], 'repo_template': 
> 'repo_suse_rhel.j2', 'repo_file_name': 'HDP', 'mirror_list': None}
> 2014-11-05 17:12:14,003 - File['/etc/yum.repos.d/HDP.repo'] {'content': 
> Template('repo_suse_rhel.j2')}
> 2014-11-05 17:12:14,004 - Repository['HDP-UTILS-1.1.0.17'] {'base_url': 
> 'http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.17/repos/centos6', 
> 'action': ['create'], 'components': ['HDP', 'main'], 'repo_template': 
> 'repo_suse_rhel.j2', 'repo_file_name': 'HDP-UTILS', 'mirror_list': None}
> 2014-11-05 17:12:14,006 - File['/etc/yum.repos.d/HDP-UTILS.repo'] {'content': 
> Template('repo_suse_rhel.j2')}
> 2014-11-05 17:12:14,007 - Package['unzip'] {}
> 2014-11-05 17:12:14,026 - Skipping installing existent package unzip
> 2014-11-05 17:12:14,026 - Package['curl'] {}
> 2014-11-05 17:12:14,048 - Skipping installing existent package curl
> 2014-11-05 17:12:14,048 - Execute['mkdir -p /tmp/HDP-artifacts/ ;   curl -kf 
> -x ""   --retry 10 
> http://ambari.bh.com:8080/resources//jdk-7u45-linux-x64.tar.gz -o 
> /tmp/HDP-artifacts//jdk-7u45-linux-x64.tar.gz'] {'environment': ..., 
> 'not_if': 'test -e /usr/jdk64/jdk1.7.0_45/bin/java', 'path': ['/bin', 
> '/usr/bin/']}
> 2014-11-05 17:12:14,057 - Skipping Execute['mkdir -p /tmp/HDP-artifacts/ ;   
> curl -kf -x ""   --retry 10 
> http://ambari.bh.com:8080/resources//jdk-7u45-linux-x64.tar.gz -o 
> /tmp/HDP-artifacts//jdk-7u45-linux-x64.tar.gz'] due to not_if
> 2014-11-05 17:12:14,057 - Execute['mkdir -p /usr/jdk64 ; cd /usr/jdk64 ; tar 
> -xf /tmp/HDP-artifacts//jdk-7u45-linux-x64.tar.gz > /dev/null 2>&1'] 
> {'not_if': 'test -e /usr/jdk64/jdk1.7.0_45/bin/java', 'path': ['/bin', 
> '/usr/bin/']}
> 2014-11-05 17:12:14,066 - Skipping Execute['mkdir -p /usr/jdk64 ; cd 
> /usr/jdk64 ; tar -xf /tmp/HDP-artifacts//jdk-7u45-linux-x64.tar.gz > 
> /dev/null 2>&1'] due to not_if
> 2014-11-05 17:12:14,066 - Group['hadoop'] {'ignore_failures': False}
> 2014-11-05 17:12:14,067 - Modifying group hadoop
> 2014-11-05 17:12:14,090 - Group['users'] {'ignore_failures': False}
> 2014-11-05 17:12:14,091 - Modifying group users
> 2014-11-05 17:12:14,116 - Group['users'] {'ignore_failures': False}
> 2014-11-05 17:12:14,117 - Modifying group users
> 2014-11-05 17:12:14,141 - User['ambari-qa'] {'gid': 'hadoop', 
> 'ignore_failures': False, 'groups': [u'users']}
> 2014-11-05 17:12:14,141 - Modifying user ambari-qa
> 2014-11-05 17:12:14,158 - File['/tmp/changeUid.sh'] {'content': 
> StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2014-11-05 17:12:14,159 - Execute['/tmp/changeUid.sh ambari-qa 
> /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
>  2>/dev/null'] {'not_if': 'test $(id -u ambari-qa) -gt 1000'}
> 2014-11-05 17:12:14,175 - Skipping Execute['/tmp/changeUid.sh ambari-qa 
> /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
>  2>/dev/null'] due to not_if
> 2014-11-05 17:12:14,177 - User['hbase'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'hadoop']}
> 2014-11-05 17:12:14,177 - Modifying user hbase
> 2014-11-05 17:12:14,188 - File['/tmp/changeUid.sh'] {'content': 
> StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2014-11-05 17:12:14,189 - Execute['/tmp/changeUid.sh hbase 
> /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/hadoop/hbase 
> 2>/dev/null'] {'not_if': 'test $(id -u hbase) -gt 1000'}
> 2014-11-05 17:12:14,200 - Skipping Execute['/tmp/changeUid.sh hbase 
> /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/hadoop/hbase 
> 2>/dev/null'] due to not_if
> 2014-11-05 17:12:14,200 - Group['nagios'] {'ignore_failures': False}
> 2014-11-05 17:12:14,200 - Modifying group nagios
> 2014-11-05 17:12:14,221 - User['nagios'] {'gid': 'nagios', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,221 - Modifying user nagios
> 2014-11-05 17:12:14,232 - User['oozie'] {'gid': 'hadoop', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,232 - Modifying user oozie
> 2014-11-05 17:12:14,242 - User['hcat'] {'gid': 'hadoop', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,243 - Modifying user hcat
> 2014-11-05 17:12:14,252 - User['hcat'] {'gid': 'hadoop', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,253 - Modifying user hcat
> 2014-11-05 17:12:14,262 - User['hive'] {'gid': 'hadoop', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,262 - Modifying user hive
> 2014-11-05 17:12:14,272 - User['yarn'] {'gid': 'hadoop', 'ignore_failures': 
> False}
> 2014-11-05 17:12:14,272 - Modifying user yarn
> 2014-11-05 17:12:14,282 - Group['nobody'] {'ignore_failures': False}
> 2014-11-05 17:12:14,282 - Modifying group nobody
> 2014-11-05 17:12:14,305 - Group['nobody'] {'ignore_failures': False}
> 2014-11-05 17:12:14,305 - Modifying group nobody
> 2014-11-05 17:12:14,326 - User['nobody'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'nobody']}
> 2014-11-05 17:12:14,326 - Modifying user nobody
> 2014-11-05 17:12:14,337 - User['nobody'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'nobody']}
> 2014-11-05 17:12:14,337 - Modifying user nobody
> 2014-11-05 17:12:14,350 - User['hdfs'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'hadoop']}
> 2014-11-05 17:12:14,350 - Modifying user hdfs
> 2014-11-05 17:12:14,366 - User['mapred'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'hadoop']}
> 2014-11-05 17:12:14,368 - Modifying user mapred
> 2014-11-05 17:12:14,387 - User['zookeeper'] {'gid': 'hadoop', 
> 'ignore_failures': False}
> 2014-11-05 17:12:14,388 - Modifying user zookeeper
> 2014-11-05 17:12:14,405 - User['storm'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'hadoop']}
> 2014-11-05 17:12:14,405 - Modifying user storm
> 2014-11-05 17:12:14,425 - User['falcon'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'hadoop']}
> 2014-11-05 17:12:14,426 - Modifying user falcon
> 2014-11-05 17:12:14,446 - User['tez'] {'gid': 'hadoop', 'ignore_failures': 
> False, 'groups': [u'users']}
> 2014-11-05 17:12:14,446 - Modifying user tez
> 2014-11-05 17:12:14,576 - Package['falcon'] {}
> 2014-11-05 17:12:14,610 - Installing package falcon ('/usr/bin/yum -d 0 -e 0 
> -y install falcon')
> 2014-11-05 17:12:21,987 - Error while executing command 'install':
> Traceback (most recent call last):
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 111, in execute
>     method(env)
>   File 
> "/var/lib/ambari-agent/cache/stacks/HDP/2.1/services/FALCON/package/scripts/falcon_client.py",
>  line 25, in install
>     self.install_packages(env)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
>  line 167, in install_packages
>     Package(name)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", 
> line 148, in __init__
>     self.env.run()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 149, in run
>     self.run_action(resource, action)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
> line 115, in run_action
>     provider_action()
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/__init__.py",
>  line 40, in action_install
>     self.install_package(package_name)
>   File 
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/yumrpm.py",
>  line 36, in install_package
>     shell.checked_call(cmd)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", 
> line 35, in checked_call
>     return _call(command, logoutput, True, cwd, env, preexec_fn, user, 
> wait_for_finish, timeout)
>   File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", 
> line 90, in _call
>     raise Fail(err_msg)
> Fail: Execution of '/usr/bin/yum -d 0 -e 0 -y install falcon' returned 1. 
> Error Downloading Packages:
>   hadoop-yarn-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-yarn-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] 
> No more mirrors to try.
>   bigtop-jsvc-1.0.10-1.el6.x86_64: failure: 
> bigtop-jsvc/bigtop-jsvc-1.0.10-1.el6.x86_64.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>   hadoop-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>   hadoop-client-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-client-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   hadoop-hdfs-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-hdfs-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 256] 
> No more mirrors to try.
>   hadoop-mapreduce-2.4.0.2.1.7.0-784.el6.x86_64: failure: 
> hadoop/hadoop-mapreduce-2.4.0.2.1.7.0-784.el6.x86_64.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   zookeeper-3.4.5.2.1.7.0-784.el6.noarch: failure: 
> zookeeper/zookeeper-3.4.5.2.1.7.0-784.el6.noarch.rpm from HDP-2.1: [Errno 
> 256] No more mirrors to try.
>   falcon-0.5.0.2.1.7.0-784.el6.noarch: failure: 
> falcon/falcon-0.5.0.2.1.7.0-784.el6.noarch.rpm from HDP-2.1: [Errno 256] No 
> more mirrors to try.
>
>
>
>
> ------------------ Original ------------------
> *From: * "guxiaobo1982";<guxiaobo1...@qq.com>;
> *Send time:* Wednesday, Nov 5, 2014 3:37 PM
> *To:* "user"<user@ambari.apache.org>;
> *Subject: * Re: how to install a specific version of HDP using Ambari
>
> Is there a mapping for HDP and HDP-UTILS minor versions, if I choose to
> install HDP 2.1.5, which version of HDP-UTILS should I use?
>
>
> ------------------ Original ------------------
> *From: * "Jeff Sposetti";<j...@hortonworks.com>;
> *Send time:* Wednesday, Nov 5, 2014 9:10 AM
> *To:* "user@ambari.apache.org"<user@ambari.apache.org>;
> *Subject: * Re: how to install a specific version of HDP using Ambari
>
> You are correct that Ambari will grab the latest HDP 2.1.x maintenance
> release repos if you are connected to the internet (for it to check for the
> latest) and you select stack HDP 2.1.
>
> But if you want to install an older version of HDP 2.1.x, do the following:
>
> 1) During install, on the Select Stack page, select HDP 2.1
> 2) Expand the Advanced Repository Options section
> 3) Enter the Base URL for the HDP 2.1.x version you wish to install
> (overwriting the 2.1.7.0 repo entries that show up by default)
>
> Looking at the docs here:
>
> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.3/index.html
>
> The Base URL for HDP is
> http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.1.3.0 and
> HDP-UTILS is
> http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.17/repos/centos6
>
>
>
> On Tue, Nov 4, 2014 at 8:04 PM, guxiaobo1982 <guxiaobo1...@qq.com> wrote:
>
>> Hi,
>>
>> The current GUI can let user choose major versions of HDP to install,
>> such as 2.1, 2.0, and will install the latest minor version, such as 2.1.7,
>> but how can I choose to install a specific minor version such as 2.1.3,
>> since I found 2.1.7 may have some bugs about hive.
>>
>> Regards,
>>
>> Xiaobo gu
>>
>
>
> CONFIDENTIALITY NOTICE
> NOTICE: This message is intended for the use of the individual or entity
> to which it is addressed and may contain information that is confidential,
> privileged and exempt from disclosure under applicable law. If the reader
> of this message is not the intended recipient, you are hereby notified that
> any printing, copying, dissemination, distribution, disclosure or
> forwarding of this communication is strictly prohibited. If you have
> received this communication in error, please contact the sender immediately
> and delete it from your system. Thank You.
>

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Reply via email to