[ 
https://issues.apache.org/jira/browse/AMBARI-25885?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

LiJie2023 updated AMBARI-25885:
-------------------------------
    Description: 
When I use ambari to install, some warning messages will appear and prompt 
"Command aborted. Reason: 'Server consistent task failed and automatically 
aborted it'". But when I click "Retry", the installation will succeed. What is 
the reason? My error reason is inconsistent with 
+https://issues.apache.org/jira/browse/AMBARI-25069+

 

!image-2023-03-08-11-12-27-370.png|width=1525,height=348!

Detailed log:

stderr: 

Command aborted. Reason: 'Server considered task failed and automatically 
aborted it'
 stdout:
2023-03-06 17:23:16,264 - Stack Feature Version Info: Cluster Stack=1.0, 
Command Stack=None, Command Version=None -> 1.0
2023-03-06 17:23:16,267 - Group['flink'] {}
2023-03-06 17:23:16,275 - Adding group Group['flink']
2023-03-06 17:23:16,346 - Group['elasticsearch'] {}
2023-03-06 17:23:16,347 - Adding group Group['elasticsearch']
2023-03-06 17:23:16,362 - Group['spark'] {}
2023-03-06 17:23:16,363 - Adding group Group['spark']
2023-03-06 17:23:16,376 - Group['hdfs'] {}
2023-03-06 17:23:16,376 - Adding group Group['hdfs']
2023-03-06 17:23:16,389 - Group['hadoop'] {}
2023-03-06 17:23:16,390 - Adding group Group['hadoop']
2023-03-06 17:23:16,403 - Group['kibana'] {}
2023-03-06 17:23:16,404 - Adding group Group['kibana']
2023-03-06 17:23:16,418 - User['hive'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:16,418 - Adding user User['hive']
2023-03-06 17:23:17,167 - User['zookeeper'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,168 - Adding user User['zookeeper']
2023-03-06 17:23:17,192 - User['efak'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,193 - Adding user User['efak']
2023-03-06 17:23:17,217 - User['ams'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,217 - Adding user User['ams']
2023-03-06 17:23:17,241 - User['hubble'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,241 - Adding user User['hubble']
2023-03-06 17:23:17,266 - User['flink'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['flink', 'hadoop'], 'uid': None}
2023-03-06 17:23:17,266 - Adding user User['flink']
2023-03-06 17:23:17,290 - User['hugegraph'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,290 - Adding user User['hugegraph']
2023-03-06 17:23:17,316 - User['elasticsearch'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,316 - Adding user User['elasticsearch']
2023-03-06 17:23:17,389 - User['spark'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['spark', 'hadoop'], 'uid': None}
2023-03-06 17:23:17,389 - Adding user User['spark']
2023-03-06 17:23:17,452 - User['ambari-qa'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,452 - Adding user User['ambari-qa']
2023-03-06 17:23:17,595 - User['kafka'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,595 - Adding user User['kafka']
2023-03-06 17:23:17,622 - User['hdfs'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop'], 'uid': None}
2023-03-06 17:23:17,622 - Adding user User['hdfs']
2023-03-06 17:23:17,645 - User['yarn'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,645 - Adding user User['yarn']
2023-03-06 17:23:17,669 - User['kibana'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'kibana'], 'uid': None}
2023-03-06 17:23:17,669 - Adding user User['kibana']
2023-03-06 17:23:17,692 - User['mapred'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,693 - Adding user User['mapred']
2023-03-06 17:23:17,881 - User['hbase'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
2023-03-06 17:23:17,882 - Adding user User['hbase']
2023-03-06 17:23:18,436 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
\{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2023-03-06 17:23:18,440 - Writing 
File['/var/lib/ambari-agent/tmp/changeUid.sh'] because it doesn't exist
2023-03-06 17:23:18,440 - Changing permission for 
/var/lib/ambari-agent/tmp/changeUid.sh from 644 to 555
2023-03-06 17:23:18,441 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh 
ambari-qa 
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
 0'] \{'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2023-03-06 17:23:18,446 - Skipping 
Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa 
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
 0'] due to not_if
2023-03-06 17:23:18,446 - Directory['/tmp/hbase-hbase'] \{'owner': 'hbase', 
'create_parents': True, 'mode': 0775, 'cd_access': 'a'}
2023-03-06 17:23:18,446 - Creating directory Directory['/tmp/hbase-hbase'] 
since it doesn't exist.
2023-03-06 17:23:18,447 - Changing owner for /tmp/hbase-hbase from 0 to hbase
2023-03-06 17:23:18,447 - Changing permission for /tmp/hbase-hbase from 755 to 
775
2023-03-06 17:23:18,447 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
\{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2023-03-06 17:23:18,448 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
\{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2023-03-06 17:23:18,449 - call['/var/lib/ambari-agent/tmp/changeUid.sh hbase'] 
{}
2023-03-06 17:23:18,457 - call returned (0, '1015')
2023-03-06 17:23:18,458 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase 
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1015'] 
\{'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2023-03-06 17:23:18,463 - Skipping 
Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase 
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1015'] 
due to not_if
2023-03-06 17:23:18,463 - Group['hdfs'] {}
2023-03-06 17:23:18,464 - User['hdfs'] \{'fetch_nonlocal_groups': True, 
'groups': ['hdfs', 'hadoop', u'hdfs']}
2023-03-06 17:23:18,465 - FS Type: HDFS
2023-03-06 17:23:18,465 - Directory['/etc/hadoop'] \{'mode': 0755}
2023-03-06 17:23:18,465 - Creating directory Directory['/etc/hadoop'] since it 
doesn't exist.
2023-03-06 17:23:18,465 - 
Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] \{'owner': 'hdfs', 
'group': 'hadoop', 'mode': 01777}
2023-03-06 17:23:18,466 - Creating directory 
Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] since it doesn't 
exist.
2023-03-06 17:23:18,466 - Changing owner for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hdfs
2023-03-06 17:23:18,466 - Changing group for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hadoop
2023-03-06 17:23:18,466 - Changing permission for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 755 to 1777
2023-03-06 17:23:18,478 - Repository['BGTP-1.0-repo-1'] \{'base_url': 
'http://master.bigdata.repo:5376/bigtop', 'action': ['prepare'], 'components': 
[u'BGTP', 'main'], 'repo_template': '[{{repo_id}}]\nname=\{{repo_id}}\n\{% if 
mirror_list %}mirrorlist=\{{mirror_list}}\{% else %}baseurl=\{{base_url}}\{% 
endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-bgtp-1', 
'mirror_list': None}
2023-03-06 17:23:18,486 - Repository[None] \{'action': ['create']}
2023-03-06 17:23:18,487 - File['/tmp/tmpFcKB36'] \{'content': 
'[BGTP-1.0-repo-1]\nname=BGTP-1.0-repo-1\nbaseurl=http://master.bigdata.repo:5376/bigtop\n\npath=/\nenabled=1\ngpgcheck=0'}
2023-03-06 17:23:18,488 - Writing File['/tmp/tmpFcKB36'] because contents don't 
match
2023-03-06 17:23:18,488 - Rewriting /etc/yum.repos.d/ambari-bgtp-1.repo since 
it has changed.
2023-03-06 17:23:18,488 - File['/etc/yum.repos.d/ambari-bgtp-1.repo'] 
\{'content': StaticFile('/tmp/tmpFcKB36')}
2023-03-06 17:23:18,489 - Writing File['/etc/yum.repos.d/ambari-bgtp-1.repo'] 
because it doesn't exist
2023-03-06 17:23:18,489 - Package['unzip'] \{'retry_on_repo_unavailability': 
False, 'retry_count': 5}
Command aborted. Reason: 'Server considered task failed and automatically 
aborted it'

Command failed after 1 tries

 

  was:
When I use ambari to install, some warning messages will appear and prompt 
"Command aborted. Reason: 'Server consistent task failed and automatically 
aborted it'". But when I click "Retry", the installation will succeed. What is 
the reason? My error reason is inconsistent with 
+https://issues.apache.org/jira/browse/AMBARI-25069+

 

!image-2023-03-08-11-12-27-370.png|width=1525,height=348!

Detailed log:

stderr: Command aborted. Reason: 'Server considered task failed and 
automatically aborted it' stdout: 2023-03-06 17:23:16,264 - Stack Feature 
Version Info: Cluster Stack=1.0, Command Stack=None, Command Version=None -> 
1.0 2023-03-06 17:23:16,267 - Group['flink'] {} 2023-03-06 17:23:16,275 - 
Adding group Group['flink'] 2023-03-06 17:23:16,346 - Group['elasticsearch'] {} 
2023-03-06 17:23:16,347 - Adding group Group['elasticsearch'] 2023-03-06 
17:23:16,362 - Group['spark'] {} 2023-03-06 17:23:16,363 - Adding group 
Group['spark'] 2023-03-06 17:23:16,376 - Group['hdfs'] {} 2023-03-06 
17:23:16,376 - Adding group Group['hdfs'] 2023-03-06 17:23:16,389 - 
Group['hadoop'] {} 2023-03-06 17:23:16,390 - Adding group Group['hadoop'] 
2023-03-06 17:23:16,403 - Group['kibana'] {} 2023-03-06 17:23:16,404 - Adding 
group Group['kibana'] 2023-03-06 17:23:16,418 - User['hive'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2023-03-06 
17:23:16,418 - Adding user User['hive'] 2023-03-06 17:23:17,167 - 
User['zookeeper'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,168 - Adding user 
User['zookeeper'] 2023-03-06 17:23:17,192 - User['efak'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2023-03-06 
17:23:17,193 - Adding user User['efak'] 2023-03-06 17:23:17,217 - User['ams'] 
\{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': 
None} 2023-03-06 17:23:17,217 - Adding user User['ams'] 2023-03-06 17:23:17,241 
- User['hubble'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,241 - Adding user User['hubble'] 
2023-03-06 17:23:17,266 - User['flink'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['flink', 'hadoop'], 'uid': None} 
2023-03-06 17:23:17,266 - Adding user User['flink'] 2023-03-06 17:23:17,290 - 
User['hugegraph'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,290 - Adding user 
User['hugegraph'] 2023-03-06 17:23:17,316 - User['elasticsearch'] \{'gid': 
'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 
2023-03-06 17:23:17,316 - Adding user User['elasticsearch'] 2023-03-06 
17:23:17,389 - User['spark'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 
'groups': ['spark', 'hadoop'], 'uid': None} 2023-03-06 17:23:17,389 - Adding 
user User['spark'] 2023-03-06 17:23:17,452 - User['ambari-qa'] \{'gid': 
'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 
2023-03-06 17:23:17,452 - Adding user User['ambari-qa'] 2023-03-06 17:23:17,595 
- User['kafka'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,595 - Adding user User['kafka'] 
2023-03-06 17:23:17,622 - User['hdfs'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop'], 'uid': None} 
2023-03-06 17:23:17,622 - Adding user User['hdfs'] 2023-03-06 17:23:17,645 - 
User['yarn'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,645 - Adding user User['yarn'] 
2023-03-06 17:23:17,669 - User['kibana'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'kibana'], 'uid': None} 
2023-03-06 17:23:17,669 - Adding user User['kibana'] 2023-03-06 17:23:17,692 - 
User['mapred'] \{'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': 
['hadoop'], 'uid': None} 2023-03-06 17:23:17,693 - Adding user User['mapred'] 
2023-03-06 17:23:17,881 - User['hbase'] \{'gid': 'hadoop', 
'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2023-03-06 
17:23:17,882 - Adding user User['hbase'] 2023-03-06 17:23:18,436 - 
File['/var/lib/ambari-agent/tmp/changeUid.sh'] \{'content': 
StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2023-03-06 17:23:18,440 - 
Writing File['/var/lib/ambari-agent/tmp/changeUid.sh'] because it doesn't exist 
2023-03-06 17:23:18,440 - Changing permission for 
/var/lib/ambari-agent/tmp/changeUid.sh from 644 to 555 2023-03-06 17:23:18,441 
- Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa 
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
 0'] \{'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'} 2023-03-06 
17:23:18,446 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh 
ambari-qa 
/tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
 0'] due to not_if 2023-03-06 17:23:18,446 - Directory['/tmp/hbase-hbase'] 
\{'owner': 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'} 
2023-03-06 17:23:18,446 - Creating directory Directory['/tmp/hbase-hbase'] 
since it doesn't exist. 2023-03-06 17:23:18,447 - Changing owner for 
/tmp/hbase-hbase from 0 to hbase 2023-03-06 17:23:18,447 - Changing permission 
for /tmp/hbase-hbase from 755 to 775 2023-03-06 17:23:18,447 - 
File['/var/lib/ambari-agent/tmp/changeUid.sh'] \{'content': 
StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2023-03-06 17:23:18,448 - 
File['/var/lib/ambari-agent/tmp/changeUid.sh'] \{'content': 
StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2023-03-06 17:23:18,449 - 
call['/var/lib/ambari-agent/tmp/changeUid.sh hbase'] {} 2023-03-06 17:23:18,457 
- call returned (0, '1015') 2023-03-06 17:23:18,458 - 
Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase 
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1015'] 
\{'not_if': '(test $(id -u hbase) -gt 1000) || (false)'} 2023-03-06 
17:23:18,463 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase 
/home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1015'] 
due to not_if 2023-03-06 17:23:18,463 - Group['hdfs'] {} 2023-03-06 
17:23:18,464 - User['hdfs'] \{'fetch_nonlocal_groups': True, 'groups': ['hdfs', 
'hadoop', u'hdfs']} 2023-03-06 17:23:18,465 - FS Type: HDFS 2023-03-06 
17:23:18,465 - Directory['/etc/hadoop'] \{'mode': 0755} 2023-03-06 17:23:18,465 
- Creating directory Directory['/etc/hadoop'] since it doesn't exist. 
2023-03-06 17:23:18,465 - 
Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] \{'owner': 'hdfs', 
'group': 'hadoop', 'mode': 01777} 2023-03-06 17:23:18,466 - Creating directory 
Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] since it doesn't 
exist. 2023-03-06 17:23:18,466 - Changing owner for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hdfs 2023-03-06 
17:23:18,466 - Changing group for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hadoop 2023-03-06 
17:23:18,466 - Changing permission for 
/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 755 to 1777 2023-03-06 
17:23:18,478 - Repository['BGTP-1.0-repo-1'] \{'base_url': 
'http://master.bigdata.repo:5376/bigtop', 'action': ['prepare'], 'components': 
[u'BGTP', 'main'], 'repo_template': '[{{repo_id}}]\nname=\{{repo_id}}\n\{% if 
mirror_list %}mirrorlist=\{{mirror_list}}\{% else %}baseurl=\{{base_url}}\{% 
endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-bgtp-1', 
'mirror_list': None} 2023-03-06 17:23:18,486 - Repository[None] \{'action': 
['create']} 2023-03-06 17:23:18,487 - File['/tmp/tmpFcKB36'] \{'content': 
'[BGTP-1.0-repo-1]\nname=BGTP-1.0-repo-1\nbaseurl=http://master.bigdata.repo:5376/bigtop\n\npath=/\nenabled=1\ngpgcheck=0'}
 2023-03-06 17:23:18,488 - Writing File['/tmp/tmpFcKB36'] because contents 
don't match 2023-03-06 17:23:18,488 - Rewriting 
/etc/yum.repos.d/ambari-bgtp-1.repo since it has changed. 2023-03-06 
17:23:18,488 - File['/etc/yum.repos.d/ambari-bgtp-1.repo'] \{'content': 
StaticFile('/tmp/tmpFcKB36')} 2023-03-06 17:23:18,489 - Writing 
File['/etc/yum.repos.d/ambari-bgtp-1.repo'] because it doesn't exist 2023-03-06 
17:23:18,489 - Package['unzip'] \{'retry_on_repo_unavailability': False, 
'retry_count': 5} Command aborted. Reason: 'Server considered task failed and 
automatically aborted it' Command failed after 1 tries

 


> “Server considered task failed and automatically aborted it”
> ------------------------------------------------------------
>
>                 Key: AMBARI-25885
>                 URL: https://issues.apache.org/jira/browse/AMBARI-25885
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 2.7.5
>            Reporter: LiJie2023
>            Priority: Blocker
>         Attachments: image-2023-03-08-11-12-27-370.png
>
>
> When I use ambari to install, some warning messages will appear and prompt 
> "Command aborted. Reason: 'Server consistent task failed and automatically 
> aborted it'". But when I click "Retry", the installation will succeed. What 
> is the reason? My error reason is inconsistent with 
> +https://issues.apache.org/jira/browse/AMBARI-25069+
>  
> !image-2023-03-08-11-12-27-370.png|width=1525,height=348!
> Detailed log:
> stderr: 
> Command aborted. Reason: 'Server considered task failed and automatically 
> aborted it'
>  stdout:
> 2023-03-06 17:23:16,264 - Stack Feature Version Info: Cluster Stack=1.0, 
> Command Stack=None, Command Version=None -> 1.0
> 2023-03-06 17:23:16,267 - Group['flink'] {}
> 2023-03-06 17:23:16,275 - Adding group Group['flink']
> 2023-03-06 17:23:16,346 - Group['elasticsearch'] {}
> 2023-03-06 17:23:16,347 - Adding group Group['elasticsearch']
> 2023-03-06 17:23:16,362 - Group['spark'] {}
> 2023-03-06 17:23:16,363 - Adding group Group['spark']
> 2023-03-06 17:23:16,376 - Group['hdfs'] {}
> 2023-03-06 17:23:16,376 - Adding group Group['hdfs']
> 2023-03-06 17:23:16,389 - Group['hadoop'] {}
> 2023-03-06 17:23:16,390 - Adding group Group['hadoop']
> 2023-03-06 17:23:16,403 - Group['kibana'] {}
> 2023-03-06 17:23:16,404 - Adding group Group['kibana']
> 2023-03-06 17:23:16,418 - User['hive'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:16,418 - Adding user User['hive']
> 2023-03-06 17:23:17,167 - User['zookeeper'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,168 - Adding user User['zookeeper']
> 2023-03-06 17:23:17,192 - User['efak'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,193 - Adding user User['efak']
> 2023-03-06 17:23:17,217 - User['ams'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,217 - Adding user User['ams']
> 2023-03-06 17:23:17,241 - User['hubble'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,241 - Adding user User['hubble']
> 2023-03-06 17:23:17,266 - User['flink'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['flink', 'hadoop'], 'uid': None}
> 2023-03-06 17:23:17,266 - Adding user User['flink']
> 2023-03-06 17:23:17,290 - User['hugegraph'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,290 - Adding user User['hugegraph']
> 2023-03-06 17:23:17,316 - User['elasticsearch'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,316 - Adding user User['elasticsearch']
> 2023-03-06 17:23:17,389 - User['spark'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['spark', 'hadoop'], 'uid': None}
> 2023-03-06 17:23:17,389 - Adding user User['spark']
> 2023-03-06 17:23:17,452 - User['ambari-qa'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,452 - Adding user User['ambari-qa']
> 2023-03-06 17:23:17,595 - User['kafka'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,595 - Adding user User['kafka']
> 2023-03-06 17:23:17,622 - User['hdfs'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop'], 'uid': None}
> 2023-03-06 17:23:17,622 - Adding user User['hdfs']
> 2023-03-06 17:23:17,645 - User['yarn'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,645 - Adding user User['yarn']
> 2023-03-06 17:23:17,669 - User['kibana'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'kibana'], 'uid': None}
> 2023-03-06 17:23:17,669 - Adding user User['kibana']
> 2023-03-06 17:23:17,692 - User['mapred'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,693 - Adding user User['mapred']
> 2023-03-06 17:23:17,881 - User['hbase'] \{'gid': 'hadoop', 
> 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None}
> 2023-03-06 17:23:17,882 - Adding user User['hbase']
> 2023-03-06 17:23:18,436 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
> \{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2023-03-06 17:23:18,440 - Writing 
> File['/var/lib/ambari-agent/tmp/changeUid.sh'] because it doesn't exist
> 2023-03-06 17:23:18,440 - Changing permission for 
> /var/lib/ambari-agent/tmp/changeUid.sh from 644 to 555
> 2023-03-06 17:23:18,441 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh 
> ambari-qa 
> /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
>  0'] \{'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
> 2023-03-06 17:23:18,446 - Skipping 
> Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa 
> /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa
>  0'] due to not_if
> 2023-03-06 17:23:18,446 - Directory['/tmp/hbase-hbase'] \{'owner': 'hbase', 
> 'create_parents': True, 'mode': 0775, 'cd_access': 'a'}
> 2023-03-06 17:23:18,446 - Creating directory Directory['/tmp/hbase-hbase'] 
> since it doesn't exist.
> 2023-03-06 17:23:18,447 - Changing owner for /tmp/hbase-hbase from 0 to hbase
> 2023-03-06 17:23:18,447 - Changing permission for /tmp/hbase-hbase from 755 
> to 775
> 2023-03-06 17:23:18,447 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
> \{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2023-03-06 17:23:18,448 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] 
> \{'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2023-03-06 17:23:18,449 - call['/var/lib/ambari-agent/tmp/changeUid.sh 
> hbase'] {}
> 2023-03-06 17:23:18,457 - call returned (0, '1015')
> 2023-03-06 17:23:18,458 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh 
> hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 
> 1015'] \{'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
> 2023-03-06 17:23:18,463 - Skipping 
> Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase 
> /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase 1015'] 
> due to not_if
> 2023-03-06 17:23:18,463 - Group['hdfs'] {}
> 2023-03-06 17:23:18,464 - User['hdfs'] \{'fetch_nonlocal_groups': True, 
> 'groups': ['hdfs', 'hadoop', u'hdfs']}
> 2023-03-06 17:23:18,465 - FS Type: HDFS
> 2023-03-06 17:23:18,465 - Directory['/etc/hadoop'] \{'mode': 0755}
> 2023-03-06 17:23:18,465 - Creating directory Directory['/etc/hadoop'] since 
> it doesn't exist.
> 2023-03-06 17:23:18,465 - 
> Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] \{'owner': 
> 'hdfs', 'group': 'hadoop', 'mode': 01777}
> 2023-03-06 17:23:18,466 - Creating directory 
> Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] since it doesn't 
> exist.
> 2023-03-06 17:23:18,466 - Changing owner for 
> /var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hdfs
> 2023-03-06 17:23:18,466 - Changing group for 
> /var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 0 to hadoop
> 2023-03-06 17:23:18,466 - Changing permission for 
> /var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir from 755 to 1777
> 2023-03-06 17:23:18,478 - Repository['BGTP-1.0-repo-1'] \{'base_url': 
> 'http://master.bigdata.repo:5376/bigtop', 'action': ['prepare'], 
> 'components': [u'BGTP', 'main'], 'repo_template': 
> '[{{repo_id}}]\nname=\{{repo_id}}\n\{% if mirror_list 
> %}mirrorlist=\{{mirror_list}}\{% else %}baseurl=\{{base_url}}\{% endif 
> %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'ambari-bgtp-1', 
> 'mirror_list': None}
> 2023-03-06 17:23:18,486 - Repository[None] \{'action': ['create']}
> 2023-03-06 17:23:18,487 - File['/tmp/tmpFcKB36'] \{'content': 
> '[BGTP-1.0-repo-1]\nname=BGTP-1.0-repo-1\nbaseurl=http://master.bigdata.repo:5376/bigtop\n\npath=/\nenabled=1\ngpgcheck=0'}
> 2023-03-06 17:23:18,488 - Writing File['/tmp/tmpFcKB36'] because contents 
> don't match
> 2023-03-06 17:23:18,488 - Rewriting /etc/yum.repos.d/ambari-bgtp-1.repo since 
> it has changed.
> 2023-03-06 17:23:18,488 - File['/etc/yum.repos.d/ambari-bgtp-1.repo'] 
> \{'content': StaticFile('/tmp/tmpFcKB36')}
> 2023-03-06 17:23:18,489 - Writing File['/etc/yum.repos.d/ambari-bgtp-1.repo'] 
> because it doesn't exist
> 2023-03-06 17:23:18,489 - Package['unzip'] \{'retry_on_repo_unavailability': 
> False, 'retry_count': 5}
> Command aborted. Reason: 'Server considered task failed and automatically 
> aborted it'
> Command failed after 1 tries
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@ambari.apache.org
For additional commands, e-mail: issues-h...@ambari.apache.org

Reply via email to