[
https://issues.apache.org/jira/browse/AMBARI-7484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Dmytro Sen updated AMBARI-7484:
-------------------------------
Description:
Versioned RPMS install fails on after-install Hook.
{code}
': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin',
'/usr/bin']}
2014-09-24 23:35:34,085 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c
HDPFlumeServer -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia',
'/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,119 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c
HDPSlaves -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia',
'/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,154 - Execute['chkconfig gmond off'] {'path':
['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,165 - Execute['chkconfig gmetad off'] {'path':
['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,284 - Directory['/etc/hadoop/conf.empty'] {'owner': 'root',
'group': 'root', 'recursive': True}
2014-09-24 23:35:34,285 - Creating directory Directory['/etc/hadoop/conf.empty']
2014-09-24 23:35:34,287 - Link['/etc/hadoop/conf'] {'not_if': 'ls
/etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'}
2014-09-24 23:35:34,297 - Creating symbolic Link['/etc/hadoop/conf']
2014-09-24 23:35:34,309 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content':
InlineTemplate(...), 'owner': 'hdfs'}
2014-09-24 23:35:34,309 - Writing File['/etc/hadoop/conf/hadoop-env.sh']
because it doesn't exist
2014-09-24 23:35:34,310 - Changing owner for /etc/hadoop/conf/hadoop-env.sh
from 0 to hdfs
2014-09-24 23:35:34,310 - Execute['ln -s /usr/hdp/2.9.9.9* /usr/hdp/current']
{'not_if': 'ls /usr/hdp/current'}
2014-09-24 23:35:34,329 - Error while executing command 'install':
Traceback (most recent call last):
File
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
line 122, in execute
method(env)
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
line 31, in hook
setup_hadoop_env()
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
line 44, in setup_hadoop_env
not_if=format('ls {versioned_hdp_root}')
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py",
line 148, in __init__
self.env.run()
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 149, in run
self.run_action(resource, action)
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 115, in run_action
provider_action()
File
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
line 237, in action_run
raise ex
Fail: Execution of 'ln -s /usr/hdp/2.9.9.9* /usr/hdp/current' returned 1. ln:
creating symbolic link `/usr/hdp/current': No such file or directory
{code}
was:
Versioned RPMS install fails on preinstall Hook.
{code}
': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin',
'/usr/bin']}
2014-09-24 23:35:34,085 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c
HDPFlumeServer -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia',
'/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,119 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c
HDPSlaves -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia',
'/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,154 - Execute['chkconfig gmond off'] {'path':
['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,165 - Execute['chkconfig gmetad off'] {'path':
['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,284 - Directory['/etc/hadoop/conf.empty'] {'owner': 'root',
'group': 'root', 'recursive': True}
2014-09-24 23:35:34,285 - Creating directory Directory['/etc/hadoop/conf.empty']
2014-09-24 23:35:34,287 - Link['/etc/hadoop/conf'] {'not_if': 'ls
/etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'}
2014-09-24 23:35:34,297 - Creating symbolic Link['/etc/hadoop/conf']
2014-09-24 23:35:34,309 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content':
InlineTemplate(...), 'owner': 'hdfs'}
2014-09-24 23:35:34,309 - Writing File['/etc/hadoop/conf/hadoop-env.sh']
because it doesn't exist
2014-09-24 23:35:34,310 - Changing owner for /etc/hadoop/conf/hadoop-env.sh
from 0 to hdfs
2014-09-24 23:35:34,310 - Execute['ln -s /usr/hdp/2.9.9.9* /usr/hdp/current']
{'not_if': 'ls /usr/hdp/current'}
2014-09-24 23:35:34,329 - Error while executing command 'install':
Traceback (most recent call last):
File
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
line 122, in execute
method(env)
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
line 31, in hook
setup_hadoop_env()
File
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
line 44, in setup_hadoop_env
not_if=format('ls {versioned_hdp_root}')
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py",
line 148, in __init__
self.env.run()
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 149, in run
self.run_action(resource, action)
File
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
line 115, in run_action
provider_action()
File
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
line 237, in action_run
raise ex
Fail: Execution of 'ln -s /usr/hdp/2.9.9.9* /usr/hdp/current' returned 1. ln:
creating symbolic link `/usr/hdp/current': No such file or directory
{code}
> Versioned RPMS install fails on preinstall Hook.
> ------------------------------------------------
>
> Key: AMBARI-7484
> URL: https://issues.apache.org/jira/browse/AMBARI-7484
> Project: Ambari
> Issue Type: Bug
> Components: stacks
> Affects Versions: 1.7.0
> Environment: HDP2.2
> Reporter: Dmytro Sen
> Assignee: Dmytro Sen
> Priority: Blocker
> Fix For: 1.7.0
>
>
> Versioned RPMS install fails on after-install Hook.
> {code}
> ': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin',
> '/usr/bin']}
> 2014-09-24 23:35:34,085 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh
> -c HDPFlumeServer -m -o root -g hadoop'] {'path':
> ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin',
> '/usr/bin']}
> 2014-09-24 23:35:34,119 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh
> -c HDPSlaves -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia',
> '/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
> 2014-09-24 23:35:34,154 - Execute['chkconfig gmond off'] {'path':
> ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
> 2014-09-24 23:35:34,165 - Execute['chkconfig gmetad off'] {'path':
> ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
> 2014-09-24 23:35:34,284 - Directory['/etc/hadoop/conf.empty'] {'owner':
> 'root', 'group': 'root', 'recursive': True}
> 2014-09-24 23:35:34,285 - Creating directory
> Directory['/etc/hadoop/conf.empty']
> 2014-09-24 23:35:34,287 - Link['/etc/hadoop/conf'] {'not_if': 'ls
> /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'}
> 2014-09-24 23:35:34,297 - Creating symbolic Link['/etc/hadoop/conf']
> 2014-09-24 23:35:34,309 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content':
> InlineTemplate(...), 'owner': 'hdfs'}
> 2014-09-24 23:35:34,309 - Writing File['/etc/hadoop/conf/hadoop-env.sh']
> because it doesn't exist
> 2014-09-24 23:35:34,310 - Changing owner for /etc/hadoop/conf/hadoop-env.sh
> from 0 to hdfs
> 2014-09-24 23:35:34,310 - Execute['ln -s /usr/hdp/2.9.9.9* /usr/hdp/current']
> {'not_if': 'ls /usr/hdp/current'}
> 2014-09-24 23:35:34,329 - Error while executing command 'install':
> Traceback (most recent call last):
> File
> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
> line 122, in execute
> method(env)
> File
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py",
> line 31, in hook
> setup_hadoop_env()
> File
> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py",
> line 44, in setup_hadoop_env
> not_if=format('ls {versioned_hdp_root}')
> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py",
> line 148, in __init__
> self.env.run()
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
> line 149, in run
> self.run_action(resource, action)
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py",
> line 115, in run_action
> provider_action()
> File
> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
> line 237, in action_run
> raise ex
> Fail: Execution of 'ln -s /usr/hdp/2.9.9.9* /usr/hdp/current' returned 1. ln:
> creating symbolic link `/usr/hdp/current': No such file or directory
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)