Hey, Thanks, That seems to have done it!
Now I have some whole new errors to work on :) 2012-12-10 20:08:44 ERROR keystone.middleware.auth_token [-] HTTP connection exception: [Errno 1] _ssl.c:490: error:140770FC:SSL routines:SSL23_GET_SERVER_HELLO:unknown protocol 2012-12-10 20:08:44 WARNING keystone.middleware.auth_token [-] Authorization failed for token fbb8f2fe3fcb4155b9428b862b5bb943 2012-12-10 20:08:44 INFO keystone.middleware.auth_token [-] Invalid user token - rejecting request 2012-12-10 20:08:44 INFO nova.osapi_compute.wsgi.server [-] 127.0.0.1 - - [10/Dec/2012 20:08:44] "GET /v2/3efa0ffe282c4a0c8b0b3a2812b1b4d0/servers/detail HTTP/1.1" 401 462 0.012388 2012-12-10 20:08:44 ERROR keystone.middleware.auth_token [-] HTTP connection exception: [Errno 1] _ssl.c:490: error:140770FC:SSL routines:SSL23_GET_SERVER_HELLO:unknown protocol 2012-12-10 20:08:44 WARNING keystone.middleware.auth_token [-] Authorization failed for token e6f30024fc1142638452854f03a7735c 2012-12-10 20:08:44 INFO keystone.middleware.auth_token [-] Invalid user token - rejecting request 2012-12-10 20:08:44 INFO nova.osapi_compute.wsgi.server [-] 127.0.0.1 - - [10/Dec/2012 20:08:44] "GET /v2/3efa0ffe282c4a0c8b0b3a2812b1b4d0/servers/detail HTTP/1.1" 401 462 0.002427 Take care, Andrew On Dec 10, 2012, at 8:01 PM, Vishvananda Ishaya wrote: > I just realized the problem. Your issue is actually the metadata api since > you have something listening on 8775. If you are running nova-api-metadata > separately then you can remove it from your list of enabled apis: > > enabled_apis=ec2,osapi_compute > > Alternatively just kill nova-api-metadata and allow it to run as wone of the > nova-api components. > > Just for your reference, nova-api is an easy way to run all of the apis as > one service. In this case it uses the enabled_apis config option. You can > also run all of the apis separately by using the individual binaries: > > nova-api-ec2 > nova-api-metadata > nova-api-os-compute > nova-api-os-volume > > Vish > > On Dec 10, 2012, at 10:42 AM, Andrew Holway <a.hol...@syseleven.de> wrote: > >> Hi, >> >> I have actually no idea how do do that. But the service opts look vaguely >> relevant: >> >> Does anyone have a working installation on Centos 6.3? >> >> Thanks, >> >> Andrew >> >> >> >> service_opts = [ >> cfg.IntOpt('report_interval', >> default=10, >> help='seconds between nodes reporting state to datastore'), >> cfg.IntOpt('periodic_interval', >> default=60, >> help='seconds between running periodic tasks'), >> cfg.IntOpt('periodic_fuzzy_delay', >> default=60, >> help='range of seconds to randomly delay when starting the' >> ' periodic task scheduler to reduce stampeding.' >> ' (Disable by setting to 0)'), >> cfg.StrOpt('ec2_listen', >> default="0.0.0.0", >> help='IP address for EC2 API to listen'), >> cfg.IntOpt('ec2_listen_port', >> default=8773, >> help='port for ec2 api to listen'), >> cfg.IntOpt('ec2_workers', >> default=None, >> help='Number of workers for EC2 API service'), >> cfg.StrOpt('osapi_compute_listen', >> default="0.0.0.0", >> help='IP address for OpenStack API to listen'), >> cfg.IntOpt('osapi_compute_listen_port', >> default=8774, >> help='list port for osapi compute'), >> cfg.IntOpt('osapi_compute_workers', >> default=None, >> help='Number of workers for OpenStack API service'), >> cfg.StrOpt('metadata_manager', >> default='nova.api.manager.MetadataManager', >> help='OpenStack metadata service manager'), >> cfg.StrOpt('metadata_listen', >> default="0.0.0.0", >> help='IP address for metadata api to listen'), >> cfg.IntOpt('metadata_listen_port', >> default=8775, >> help='port for metadata api to listen'), >> cfg.IntOpt('metadata_workers', >> default=None, >> help='Number of workers for metadata service'), >> cfg.StrOpt('osapi_volume_listen', >> default="0.0.0.0", >> help='IP address for OpenStack Volume API to listen'), >> cfg.IntOpt('osapi_volume_listen_port', >> default=8776, >> help='port for os volume api to listen'), >> cfg.IntOpt('osapi_volume_workers', >> default=None, >> help='Number of workers for OpenStack Volume API service'), >> ] >> >> On Dec 10, 2012, at 7:29 PM, Vishvananda Ishaya wrote: >> >>> Nope. Best i can think of is to throw some log statements into >>> nova/service.py right before the exception gets thrown. See which api it is >>> trying to start and what it thinks the value of enabled_apis is. Etc. >>> >>> Vish >>> >>> On Dec 10, 2012, at 10:24 AM, Andrew Holway <a.hol...@syseleven.de> wrote: >>> >>>> Hi, >>>> >>>> maybe this will shed some light on it..? >>>> >>>> Thanks, >>>> >>>> Andrew >>>> >>>> [root@blade02 init.d]# cat /etc/nova/api-paste.ini >>>> ############ >>>> # Metadata # >>>> ############ >>>> [composite:metadata] >>>> use = egg:Paste#urlmap >>>> /: meta >>>> >>>> [pipeline:meta] >>>> pipeline = ec2faultwrap logrequest metaapp >>>> >>>> [app:metaapp] >>>> paste.app_factory = >>>> nova.api.metadata.handler:MetadataRequestHandler.factory >>>> >>>> ####### >>>> # EC2 # >>>> ####### >>>> >>>> [composite:ec2] >>>> use = egg:Paste#urlmap >>>> /services/Cloud: ec2cloud >>>> >>>> [composite:ec2cloud] >>>> use = call:nova.api.auth:pipeline_factory >>>> noauth = ec2faultwrap logrequest ec2noauth cloudrequest validator >>>> ec2executor >>>> keystone = ec2faultwrap logrequest ec2keystoneauth cloudrequest validator >>>> ec2executor >>>> >>>> [filter:ec2faultwrap] >>>> paste.filter_factory = nova.api.ec2:FaultWrapper.factory >>>> >>>> [filter:logrequest] >>>> paste.filter_factory = nova.api.ec2:RequestLogging.factory >>>> >>>> [filter:ec2lockout] >>>> paste.filter_factory = nova.api.ec2:Lockout.factory >>>> >>>> [filter:ec2keystoneauth] >>>> paste.filter_factory = nova.api.ec2:EC2KeystoneAuth.factory >>>> >>>> [filter:ec2noauth] >>>> paste.filter_factory = nova.api.ec2:NoAuth.factory >>>> >>>> [filter:cloudrequest] >>>> controller = nova.api.ec2.cloud.CloudController >>>> paste.filter_factory = nova.api.ec2:Requestify.factory >>>> >>>> [filter:authorizer] >>>> paste.filter_factory = nova.api.ec2:Authorizer.factory >>>> >>>> [filter:validator] >>>> paste.filter_factory = nova.api.ec2:Validator.factory >>>> >>>> [app:ec2executor] >>>> paste.app_factory = nova.api.ec2:Executor.factory >>>> >>>> ############# >>>> # Openstack # >>>> ############# >>>> >>>> [composite:osapi_compute] >>>> use = call:nova.api.openstack.urlmap:urlmap_factory >>>> /: oscomputeversions >>>> /v1.1: openstack_compute_api_v2 >>>> /v2: openstack_compute_api_v2 >>>> >>>> [composite:osapi_volume] >>>> use = call:nova.api.openstack.urlmap:urlmap_factory >>>> /: osvolumeversions >>>> /v1: openstack_volume_api_v1 >>>> >>>> [composite:openstack_compute_api_v2] >>>> use = call:nova.api.auth:pipeline_factory >>>> noauth = faultwrap sizelimit noauth ratelimit osapi_compute_app_v2 >>>> keystone = faultwrap sizelimit authtoken keystonecontext ratelimit >>>> osapi_compute_app_v2 >>>> keystone_nolimit = faultwrap sizelimit authtoken keystonecontext >>>> osapi_compute_app_v2 >>>> >>>> [composite:openstack_volume_api_v1] >>>> use = call:nova.api.auth:pipeline_factory >>>> noauth = faultwrap sizelimit noauth ratelimit osapi_volume_app_v1 >>>> keystone = faultwrap sizelimit authtoken keystonecontext ratelimit >>>> osapi_volume_app_v1 >>>> keystone_nolimit = faultwrap sizelimit authtoken keystonecontext >>>> osapi_volume_app_v1 >>>> >>>> [filter:faultwrap] >>>> paste.filter_factory = nova.api.openstack:FaultWrapper.factory >>>> >>>> [filter:noauth] >>>> paste.filter_factory = nova.api.openstack.auth:NoAuthMiddleware.factory >>>> >>>> [filter:ratelimit] >>>> paste.filter_factory = >>>> nova.api.openstack.compute.limits:RateLimitingMiddleware.factory >>>> >>>> [filter:sizelimit] >>>> paste.filter_factory = nova.api.sizelimit:RequestBodySizeLimiter.factory >>>> >>>> [app:osapi_compute_app_v2] >>>> paste.app_factory = nova.api.openstack.compute:APIRouter.factory >>>> >>>> [pipeline:oscomputeversions] >>>> pipeline = faultwrap oscomputeversionapp >>>> >>>> [app:osapi_volume_app_v1] >>>> paste.app_factory = nova.api.openstack.volume:APIRouter.factory >>>> >>>> [app:oscomputeversionapp] >>>> paste.app_factory = nova.api.openstack.compute.versions:Versions.factory >>>> >>>> [pipeline:osvolumeversions] >>>> pipeline = faultwrap osvolumeversionapp >>>> >>>> [app:osvolumeversionapp] >>>> paste.app_factory = nova.api.openstack.volume.versions:Versions.factory >>>> >>>> ########## >>>> # Shared # >>>> ########## >>>> >>>> [filter:keystonecontext] >>>> paste.filter_factory = nova.api.auth:NovaKeystoneContext.factory >>>> >>>> [filter:authtoken] >>>> paste.filter_factory = keystone.middleware.auth_token:filter_factory >>>> admin_tenant_name = service >>>> admin_user = nova >>>> admin_password = x7deix7dei >>>> auth_uri = http://controller:5000/ >>>> On Dec 10, 2012, at 7:10 PM, Vishvananda Ishaya wrote: >>>> >>>>> Odd. This looks remarkably like it is trying to start osapi_volume even >>>>> though you don't have it specified in enabled apis. Your enabled_apis >>>>> setting looks correct to me. >>>>> >>>>> Vish >>>>> >>>>> >>>>> On Dec 10, 2012, at 9:24 AM, Andrew Holway <a.hol...@syseleven.de> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I cannot start the nova-api service. >>>>>> >>>>>> [root@blade02 07-openstack-controller]# nova list >>>>>> ERROR: ConnectionRefused: '[Errno 111] Connection refused' >>>>>> >>>>>> I followed this guide very carefully: >>>>>> >>>>>> https://github.com/beloglazov/openstack-centos-kvm-glusterfs/#07-openstack-controller-controller >>>>>> >>>>>> Here is api.log >>>>>> >>>>>> 2012-12-10 17:51:31 DEBUG nova.wsgi [-] Loading app metadata from >>>>>> /etc/nova/api-paste.ini from (pid=2536) load_app >>>>>> /usr/lib/python2.6/site-packages/nova/wsgi.py:371 >>>>>> 2012-12-10 17:51:31 CRITICAL nova [-] [Errno 98] Address already in use >>>>>> 2012-12-10 17:51:31 TRACE nova Traceback (most recent call last): >>>>>> 2012-12-10 17:51:31 TRACE nova File "/usr/bin/nova-api", line 50, in >>>>>> <module> >>>>>> 2012-12-10 17:51:31 TRACE nova server = service.WSGIService(api) >>>>>> 2012-12-10 17:51:31 TRACE nova File >>>>>> "/usr/lib/python2.6/site-packages/nova/service.py", line 584, in __init__ >>>>>> 2012-12-10 17:51:31 TRACE nova port=self.port) >>>>>> 2012-12-10 17:51:31 TRACE nova File >>>>>> "/usr/lib/python2.6/site-packages/nova/wsgi.py", line 72, in __init__ >>>>>> 2012-12-10 17:51:31 TRACE nova self._socket = eventlet.listen((host, >>>>>> port), backlog=backlog) >>>>>> 2012-12-10 17:51:31 TRACE nova File >>>>>> "/usr/lib/python2.6/site-packages/eventlet/convenience.py", line 38, in >>>>>> listen >>>>>> 2012-12-10 17:51:31 TRACE nova sock.bind(addr) >>>>>> 2012-12-10 17:51:31 TRACE nova File "<string>", line 1, in bind >>>>>> 2012-12-10 17:51:31 TRACE nova error: [Errno 98] Address already in use >>>>>> 2012-12-10 17:51:31 TRACE nova >>>>>> 2012-12-10 17:51:31 INFO nova.service [-] Parent process has died >>>>>> unexpectedly, exiting >>>>>> 2012-12-10 17:51:31 INFO nova.service [-] Parent process has died >>>>>> unexpectedly, exiting >>>>>> 2012-12-10 17:51:31 INFO nova.wsgi [-] Stopping WSGI server. >>>>>> 2012-12-10 17:51:31 INFO nova.wsgi [-] Stopping WSGI server. >>>>>> >>>>>> [root@blade02 07-openstack-controller]# cat /etc/nova/nova.conf >>>>>> [DEFAULT] >>>>>> logdir = /var/log/nova >>>>>> state_path = /var/lib/nova >>>>>> lock_path = /var/lib/nova/tmp >>>>>> volumes_dir = /etc/nova/volumes >>>>>> dhcpbridge = /usr/bin/nova-dhcpbridge >>>>>> dhcpbridge_flagfile = /etc/nova/nova.conf >>>>>> force_dhcp_release = False >>>>>> injected_network_template = /usr/share/nova/interfaces.template >>>>>> libvirt_nonblocking = True >>>>>> libvirt_inject_partition = -1 >>>>>> network_manager = nova.network.manager.FlatDHCPManager >>>>>> iscsi_helper = tgtadm >>>>>> sql_connection = mysql://nova:x7deix7dei@controller/nova >>>>>> compute_driver = libvirt.LibvirtDriver >>>>>> firewall_driver = nova.virt.libvirt.firewall.IptablesFirewallDriver >>>>>> rpc_backend = nova.openstack.common.rpc.impl_qpid >>>>>> rootwrap_config = /etc/nova/rootwrap.conf >>>>>> verbose = True >>>>>> auth_strategy = keystone >>>>>> qpid_hostname = controller >>>>>> network_host = compute1 >>>>>> fixed_range = 10.0.0.0/24 >>>>>> flat_interface = eth1 >>>>>> flat_network_bridge = br100 >>>>>> public_interface = eth1 >>>>>> glance_host = controller >>>>>> vncserver_listen = 0.0.0.0 >>>>>> vncserver_proxyclient_address = controller >>>>>> novncproxy_base_url = http://37.123.104.3:6080/vnc_auto.html >>>>>> xvpvncproxy_base_url = http://37.123.104.3:6081/console >>>>>> metadata_host = 10.141.6.2 >>>>>> enabled_apis=ec2,osapi_compute,metadata >>>>>> >>>>>> #[keystone_authtoken] >>>>>> admin_tenant_name = %SERVICE_TENANT_NAME% >>>>>> admin_user = %SERVICE_USER% >>>>>> admin_password = %SERVICE_PASSWORD% >>>>>> auth_host = 127.0.0.1 >>>>>> auth_port = 35357 >>>>>> auth_protocol = http >>>>>> signing_dirname = /tmp/keystone-signing-nova >>>>>> >>>>>> There is no process using port 8774. >>>>>> >>>>>> [root@blade02 07-openstack-controller]# netstat -tunlp | grep 877 >>>>>> tcp 0 0 0.0.0.0:8775 0.0.0.0:* >>>>>> LISTEN 2157/python >>>>>> >>>>>> Maybe it is something similar to: >>>>>> >>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=877606#c3 >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Andrew >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Mailing list: https://launchpad.net/~openstack >>>>>> Post to : openstack@lists.launchpad.net >>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>> More help : https://help.launchpad.net/ListHelp >>>>> >>>> >>>> >>> >> >> > _______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp