Hello!

I am trying to set up an OpenStack cluster but I am not having much
success. I have been fighting it for several weeks thus my decision
to join this list.

I have 4 hosts all running CentOS 7:

   infra1 (Celeron J1900 CPU [4core, 2Ghz], 8GB RAM, 120GB SSD)
   compute1 (Core i7 CPU [4core, 4Ghz], 16GB RAM, 120GB SSD)
   log1 (AMD Athlon CPU [2core, 2Ghz], 3GB RAM, 120GB HDD)
   storage1 (Xeon E3 CPU [4core, 2Ghz], 8GB RAM, 8TB RAID10)

Considering the none-too-powerful specs of infra1 and log1 I have
set the following services to run on metal:

  aodh_container
  ceilometer_central_container
  cinder_api_container
  cinder_scheduler_container
  galera_container
  glance_container
  gnocchi_container
  heat_apis_container
  heat_engine_container
  horizon_container
  keystone_container
  memcached_container
  neutron_agents_container
  neutron_server_container
  nova_api_metadata_container
  nova_api_os_compute_container
  nova_api_placement_container
  nova_conductor_container
  nova_console_container
  nova_scheduler_container
  rabbit_mq_container
  repo_container
  rsyslog_container

When I run setup_hosts.yml I get errors on infra1. The specific
errors vary each time but it generally seems to fail with container
creation.

8<---8<---8<---8<---

TASK [lxc_container_create : LXC autodev setup] ******************************** Thursday 19 October 2017 17:00:28 -0700 (0:01:36.565) 0:57:32.820 ****** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: OSError: [Errno 12] Cannot allocate memory fatal: [infra1_aodh_container-3ef9fdf1]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""} An exception occurred during task execution. To see the full traceback, use -vvv. The error was: OSError: [Errno 12] Cannot allocate memory fatal: [infra1_utility_container-7578d165]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""} An exception occurred during task execution. To see the full traceback, use -vvv. The error was: OSError: [Errno 12] Cannot allocate memory fatal: [infra1_horizon_container-4056733b]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""}
ok: [infra1_nova_scheduler_container-752fb34b -> 172.29.236.11]
ok: [infra1_keystone_container-23bb4cba -> 172.29.236.11]
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: OSError: [Errno 12] Cannot allocate memory fatal: [infra1_glance_container-299bd597]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""}
changed: [infra1_neutron_agents_container-e319526f -> 172.29.236.11]
changed: [infra1_cinder_scheduler_container-84442b11 -> 172.29.236.11]
changed: [infra1_neutron_server_container-d19ab320 -> 172.29.236.11]
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: OSError: [Errno 12] Cannot allocate memory fatal: [infra1_repo_container-9b73f4cd]: FAILED! => {"failed": true, "msg": "Unexpected failure during module execution.", "stdout": ""}
ok: [infra1_cinder_api_container-05fbf13a -> 172.29.236.11]
ok: [infra1_nova_api_os_compute_container-99a9a1e0 -> 172.29.236.11]
ok: [infra1_nova_api_metadata_container-0a10aa4a -> 172.29.236.11]
ok: [infra1_galera_container-a3be12a1 -> 172.29.236.11]
ok: [infra1_nova_conductor_container-d8c2040f -> 172.29.236.11]
ok: [infra1_nova_console_container-e4a8d3ae -> 172.29.236.11]
ok: [infra1_gnocchi_container-e83732f5 -> 172.29.236.11]
ok: [infra1_rabbit_mq_container-4c8a4541 -> 172.29.236.11]
ok: [infra1_ceilometer_central_container-fe8f973b -> 172.29.236.11]
ok: [infra1_memcached_container-895a7ccf -> 172.29.236.11]
ok: [infra1_nova_api_placement_container-ec10eadb -> 172.29.236.11]
ok: [infra1_heat_apis_container-7579f33e -> 172.29.236.11]
ok: [infra1_heat_engine_container-2a26e880 -> 172.29.236.11]

8<---8<---8<---8<---

In case it is useful here is my user config:

---
cidr_networks:
 container: 172.29.236.0/22
 tunnel: 172.29.240.0/22
 storage: 172.29.244.0/22

used_ips:
 - 172.29.236.1
 - "172.29.236.100,172.29.236.200"
 - "172.29.240.100,172.29.240.200"
 - "172.29.244.100,172.29.244.200"

global_overrides:
 internal_lb_vip_address: 172.29.236.9
 #
 # The below domain name must resolve to an IP address
 # in the CIDR specified in haproxy_keepalived_external_vip_cidr.
 # If using different protocols (https/http) for the public/internal
 # endpoints the two addresses must be different.
 #
 external_lb_vip_address: 172.29.236.10
 tunnel_bridge: "br-vxlan"
 management_bridge: "br-mgmt"
 provider_networks:
   - network:
       container_bridge: "br-mgmt"
       container_type: "veth"
       container_interface: "eth1"
       ip_from_q: "container"
       type: "raw"
       group_binds:
         - all_containers
         - hosts
       is_container_address: true
       is_ssh_address: true
   - network:
       container_bridge: "br-vxlan"
       container_type: "veth"
       container_interface: "eth10"
       ip_from_q: "tunnel"
       type: "vxlan"
       range: "1:1000"
       net_name: "vxlan"
       group_binds:
         - neutron_linuxbridge_agent
   - network:
       container_bridge: "br-vlan"
       container_type: "veth"
       container_interface: "eth12"
       host_bind_over: "br-vlan"
       type: "flat"
       net_name: "flat"
       group_binds:
         - neutron_linuxbridge_agent
   - network:
       container_bridge: "br-vlan"
       container_type: "veth"
       container_interface: "eth11"
       type: "vlan"
       range: "1:1"
       net_name: "vlan"
       group_binds:
         - neutron_linuxbridge_agent
   - network:
       container_bridge: "br-storage"
       container_type: "veth"
       container_interface: "eth2"
       ip_from_q: "storage"
       type: "raw"
       group_binds:
         - glance_api
         - cinder_api
         - cinder_volume
         - nova_compute

###
### Infrastructure
###

# galera, memcache, rabbitmq, utility
shared-infra_hosts:
 infra1:
   ip: 172.29.236.11

# repository (apt cache, python packages, etc)
repo-infra_hosts:
 infra1:
   ip: 172.29.236.11

# load balancer
# Ideally the load balancer should not use the Infrastructure hosts.
# Dedicated hardware is best for improved performance and security.
haproxy_hosts:
 infra1:
   ip: 172.29.236.11

# rsyslog server
log_hosts:
 log1:
   ip: 172.29.236.14

###
### OpenStack
###

# keystone
identity_hosts:
 infra1:
   ip: 172.29.236.11

# cinder api services
storage-infra_hosts:
 infra1:
   ip: 172.29.236.11

# glance
# The settings here are repeated for each infra host.
# They could instead be applied as global settings in
# user_variables, but are left here to illustrate that
# each container could have different storage targets.
image_hosts:
 infra1:
   ip: 172.29.236.11
   container_vars:
     limit_container_types: glance
     glance_nfs_client:
       - server: "172.29.244.15"
         remote_path: "/images"
         local_path: "/var/lib/glance/images"
         type: "nfs"
         options: "_netdev,auto"

# nova api, conductor, etc services
compute-infra_hosts:
 infra1:
   ip: 172.29.236.11

# heat
orchestration_hosts:
 infra1:
   ip: 172.29.236.11

# horizon
dashboard_hosts:
 infra1:
   ip: 172.29.236.11

# neutron server, agents (L3, etc)
network_hosts:
 infra1:
   ip: 172.29.236.11

# ceilometer (telemetry data collection)
metering-infra_hosts:
 infra1:
   ip: 172.29.236.11

# aodh (telemetry alarm service)
metering-alarm_hosts:
 infra1:
   ip: 172.29.236.11

# gnocchi (telemetry metrics storage)
metrics_hosts:
 infra1:
   ip: 172.29.236.11

# nova hypervisors
compute_hosts:
 compute1:
   ip: 172.29.236.12

# ceilometer compute agent (telemetry data collection)
metering-compute_hosts:
 compute1:
   ip: 172.29.236.12

---

Thanks for any ideas!

FV

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to