I know it's "Available ", however that doesn't imply attachment. Cinder
uses iSCSI or NFS, to attach the volume to a running instance on a compute
node. If you're missing the required protocol packages, the attachment will
fail. You can have "Available " volumes, and lack tgtadm (or nfs-utils if
that's your protocol).

Secondly, Is your compute node able to resolve "controller"?

On Wednesday, January 14, 2015, Geo Varghese <gvargh...@aqorn.com> wrote:

> Hi Abel,
>
> Thanks for the reply.
>
> I have created volume and its in available state. Please check attached
> screenshot.
>
>
>
> On Thu, Jan 15, 2015 at 11:34 AM, Abel Lopez <alopg...@gmail.com
> <javascript:_e(%7B%7D,'cvml','alopg...@gmail.com');>> wrote:
>
>> Do your compute nodes have the required iSCSI packages installed?
>>
>>
>> On Wednesday, January 14, 2015, Geo Varghese <gvargh...@aqorn.com
>> <javascript:_e(%7B%7D,'cvml','gvargh...@aqorn.com');>> wrote:
>>
>>> Hi Jay,
>>>
>>> Thanks for the reply. Just pasting the details below
>>>
>>> keystone catalog
>>> ================================================
>>> Service: compute
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   Property  |
>>> Value                            |
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   adminURL  |
>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |      id     |
>>> 02028b1f4c0849c68eb79f5887516299              |
>>> | internalURL |
>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |  publicURL  |
>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |    region   |
>>> RegionOne                          |
>>>
>>> +-------------+------------------------------------------------------------+
>>> Service: network
>>> +-------------+----------------------------------+
>>> |   Property  |              Value               |
>>> +-------------+----------------------------------+
>>> |   adminURL  |      http://controller:9696      |
>>> |      id     | 32f687d4f7474769852d88932288b893 |
>>> | internalURL |      http://controller:9696      |
>>> |  publicURL  |      http://controller:9696      |
>>> |    region   |            RegionOne             |
>>> +-------------+----------------------------------+
>>> Service: volumev2
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   Property  |
>>> Value                            |
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   adminURL  |
>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |      id     |
>>> 5bca493cdde2439887d54fb805c4d2d4              |
>>> | internalURL |
>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |  publicURL  |
>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>>> |    region   |
>>> RegionOne                          |
>>>
>>> +-------------+------------------------------------------------------------+
>>> Service: image
>>> +-------------+----------------------------------+
>>> |   Property  |              Value               |
>>> +-------------+----------------------------------+
>>> |   adminURL  |      http://controller:9292      |
>>> |      id     | 2e2294b9151e4fb9b6efccf33c62181b |
>>> | internalURL |      http://controller:9292      |
>>> |  publicURL  |      http://controller:9292      |
>>> |    region   |            RegionOne             |
>>> +-------------+----------------------------------+
>>> Service: volume
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   Property  |
>>> Value                            |
>>>
>>> +-------------+------------------------------------------------------------+
>>> |   adminURL  |
>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>>> |      id     |
>>> 0e29cfaa785e4e148c57601b182a5e26              |
>>> | internalURL |
>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>>> |  publicURL  |
>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>>> |    region   |
>>> RegionOne                          |
>>>
>>> +-------------+------------------------------------------------------------+
>>> Service: ec2
>>> +-------------+---------------------------------------+
>>> |   Property  |                 Value                 |
>>> +-------------+---------------------------------------+
>>> |   adminURL  | http://controller:8773/services/Admin |
>>> |      id     |    8f4957d98cd04130b055b8b80b051833   |
>>> | internalURL | http://controller:8773/services/Cloud |
>>> |  publicURL  | http://controller:8773/services/Cloud |
>>> |    region   |               RegionOne               |
>>> +-------------+---------------------------------------+
>>> Service: identity
>>> +-------------+----------------------------------+
>>> |   Property  |              Value               |
>>> +-------------+----------------------------------+
>>> |   adminURL  |   http://controller:35357/v2.0   |
>>> |      id     | 146f7bbb0ad54740b95f8499f04b2ee2 |
>>> | internalURL |   http://controller:5000/v2.0    |
>>> |  publicURL  |   http://controller:5000/v2.0    |
>>> |    region   |            RegionOne             |
>>> +-------------+----------------------------------+
>>> ==============================================
>>>
>>> Nova.conf
>>>
>>> ================================================
>>> # This file autogenerated by Chef
>>> # Do not edit, changes will be overwritten
>>>
>>>
>>> [DEFAULT]
>>>
>>> # LOGS/STATE
>>> debug=False
>>> verbose=False
>>> auth_strategy=keystone
>>> dhcpbridge_flagfile=/etc/nova/nova.conf
>>> dhcpbridge=/usr/bin/nova-dhcpbridge
>>> log_dir=/var/log/nova
>>> state_path=/var/lib/nova
>>> instances_path=/var/lib/nova/instances
>>> instance_name_template=instance-%08x
>>> network_allocate_retries=0
>>> lock_path=/var/lib/nova/lock
>>>
>>> ssl_only=false
>>> cert=self.pem
>>> key=
>>>
>>> # Command prefix to use for running commands as root (default: sudo)
>>> rootwrap_config=/etc/nova/rootwrap.conf
>>>
>>> # Should unused base images be removed? (default: false)
>>> remove_unused_base_images=true
>>>
>>> # Unused unresized base images younger than this will not be removed
>>> (default: 86400)
>>> remove_unused_original_minimum_age_seconds=3600
>>>
>>> # Options defined in nova.openstack.common.rpc
>>> rpc_thread_pool_size=64
>>> rpc_conn_pool_size=30
>>> rpc_response_timeout=60
>>> rpc_backend=nova.openstack.common.rpc.impl_kombu
>>> amqp_durable_queues=false
>>> amqp_auto_delete=false
>>>
>>> ##### RABBITMQ #####
>>> rabbit_userid=guest
>>> rabbit_password=guest
>>> rabbit_virtual_host=/
>>> rabbit_hosts=rabbit1:5672,rabbit2:5672
>>> rabbit_retry_interval=1
>>> rabbit_retry_backoff=2
>>> rabbit_max_retries=0
>>> rabbit_durable_queues=false
>>> rabbit_ha_queues=True
>>>
>>>
>>>
>>> ##### SCHEDULER #####
>>> scheduler_manager=nova.scheduler.manager.SchedulerManager
>>> scheduler_driver=nova.scheduler.filter_scheduler.FilterScheduler
>>> scheduler_available_filters=nova.scheduler.filters.all_filters
>>> # which filter class names to use for filtering hosts when not specified
>>> in the request.
>>>
>>> scheduler_default_filters=AvailabilityZoneFilter,RamFilter,ComputeFilter,CoreFilter,SameHostFilter,DifferentHostFilter
>>> default_availability_zone=nova
>>> default_schedule_zone=nova
>>>
>>> ##### NETWORK #####
>>>
>>>
>>>
>>> # N.B. due to https://bugs.launchpad.net/nova/+bug/1206330
>>> # we override the endpoint scheme below, ignore the port
>>> # and essentially force http
>>> neutron_url=http://controller:9696
>>> neutron_api_insecure=false
>>> network_api_class=nova.network.neutronv2.api.API
>>> neutron_auth_strategy=keystone
>>> neutron_admin_tenant_name=service
>>> neutron_admin_username=neutron
>>> neutron_admin_password=openstack-network
>>> neutron_admin_auth_url=http://controller:5000/v2.0
>>> neutron_url_timeout=30
>>> neutron_region_name=
>>> neutron_ovs_bridge=br-int
>>> neutron_extension_sync_interval=600
>>> neutron_ca_certificates_file=
>>> linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver
>>> firewall_driver = nova.virt.firewall.NoopFirewallDriver
>>> security_group_api=neutron
>>> service_neutron_metadata_proxy=true
>>> neutron_metadata_proxy_shared_secret=secret123
>>> default_floating_pool=public
>>> dns_server=8.8.8.8
>>>
>>> use_ipv6=false
>>>
>>> ##### GLANCE #####
>>> image_service=nova.image.glance.GlanceImageService
>>> glance_api_servers=http://controller:9292
>>> glance_api_insecure=false
>>>
>>> ##### Cinder #####
>>> # Location of ca certificates file to use for cinder client requests
>>> cinder_ca_certificates_file=
>>>
>>> # Allow to perform insecure SSL requests to cinder
>>> cinder_api_insecure=false
>>>
>>> # Info to match when looking for cinder in the service catalog
>>> cinder_catalog_info=volumev2:cinderv2:publicURL
>>>
>>> ##### COMPUTE #####
>>> compute_driver=libvirt.LibvirtDriver
>>> preallocate_images=none
>>> use_cow_images=true
>>> vif_plugging_is_fatal=false
>>> vif_plugging_timeout=0
>>> compute_manager=nova.compute.manager.ComputeManager
>>> sql_connection=mysql://nova:nova@loadbalancer:3306/nova?charset=utf8
>>> connection_type=libvirt
>>>
>>> ##### NOTIFICATIONS #####
>>> # Driver or drivers to handle sending notifications (multi valued)
>>>
>>> # AMQP topic used for OpenStack notifications. (list value)
>>> # Deprecated group/name - [rpc_notifier2]/topics
>>> notification_topics=notifications
>>>
>>> # Generate periodic compute.instance.exists notifications
>>> instance_usage_audit=False
>>>
>>> # Time period to generate instance usages for.  Time period
>>> # must be hour, day, month or year (string value)
>>> instance_usage_audit_period=month
>>>
>>>
>>> # The IP address on which the OpenStack API will listen. (string value)
>>> osapi_compute_listen=0.0.0.0
>>> # The port on which the OpenStack API will listen. (integer value)
>>> osapi_compute_listen_port=8774
>>>
>>> # The IP address on which the metadata will listen. (string value)
>>> metadata_listen=0.0.0.0
>>> # The port on which the metadata will listen. (integer value)
>>> metadata_listen_port=8775
>>>
>>> ##### VNCPROXY #####
>>> novncproxy_base_url=http://controller:6080/vnc_auto.html
>>> xvpvncproxy_base_url=http://controller:6081/console
>>>
>>> # This is only required on the server running xvpvncproxy
>>> xvpvncproxy_host=0.0.0.0
>>> xvpvncproxy_port=6081
>>>
>>> # This is only required on the server running novncproxy
>>> novncproxy_host=0.0.0.0
>>> novncproxy_port=6080
>>>
>>> vncserver_listen=0.0.0.0
>>> vncserver_proxyclient_address=0.0.0.0
>>>
>>> vnc_keymap=en-us
>>>
>>> # store consoleauth tokens in memcached
>>>
>>> ##### MISC #####
>>> # force backing images to raw format
>>> force_raw_images=false
>>> allow_same_net_traffic=true
>>> osapi_max_limit=1000
>>> # If you terminate SSL with a load balancer, the HTTP_HOST environ
>>> # variable that generates the request_uri in webob.Request will lack
>>> # the HTTPS scheme. Setting this overrides the default and allows
>>> # URIs returned in the various links collections to contain the proper
>>> # HTTPS endpoint.
>>> osapi_compute_link_prefix = http://controller:8774/v2/%(tenant_id)s
>>> start_guests_on_host_boot=false
>>> resume_guests_state_on_host_boot=true
>>> allow_resize_to_same_host=false
>>> resize_confirm_window=0
>>> live_migration_retry_count=30
>>>
>>> ##### QUOTAS #####
>>> # (StrOpt) default driver to use for quota checks (default:
>>> nova.quota.DbQuotaDriver)
>>> quota_driver=nova.quota.DbQuotaDriver
>>> # number of security groups per project (default: 10)
>>> quota_security_groups=50
>>> # number of security rules per security group (default: 20)
>>> quota_security_group_rules=20
>>> # number of instance cores allowed per project (default: 20)
>>> quota_cores=20
>>> # number of fixed ips allowed per project (this should be at least the
>>> number of instances allowed) (default: -1)
>>> quota_fixed_ips=-1
>>> # number of floating ips allowed per project (default: 10)
>>> quota_floating_ips=10
>>> # number of bytes allowed per injected file (default: 10240)
>>> quota_injected_file_content_bytes=10240
>>> # number of bytes allowed per injected file path (default: 255)
>>> quota_injected_file_path_length=255
>>> # number of injected files allowed (default: 5)
>>> quota_injected_files=5
>>> # number of instances allowed per project (defailt: 10)
>>> quota_instances=10
>>> # number of key pairs per user (default: 100)
>>> quota_key_pairs=100
>>> # number of metadata items allowed per instance (default: 128)
>>> quota_metadata_items=128
>>> # megabytes of instance ram allowed per project (default: 51200)
>>> quota_ram=51200
>>>
>>> # virtual CPU to Physical CPU allocation ratio (default: 16.0)
>>> cpu_allocation_ratio=16.0
>>> # virtual ram to physical ram allocation ratio (default: 1.5)
>>> ram_allocation_ratio=1.5
>>>
>>> mkisofs_cmd=genisoimage
>>> injected_network_template=$pybasedir/nova/virt/interfaces.template
>>> flat_injected=false
>>>
>>> # The IP address on which the EC2 API will listen. (string value)
>>> ec2_listen=0.0.0.0
>>> # The port on which the EC2 API will listen. (integer value)
>>> ec2_listen_port=8773
>>>
>>>
>>> ##### WORKERS ######
>>>
>>> ##### KEYSTONE #####
>>> keystone_ec2_url=http://controller:5000/v2.0/ec2tokens
>>>
>>> # a list of APIs to enable by default (list value)
>>> enabled_apis=ec2,osapi_compute,metadata
>>>
>>> ##### WORKERS ######
>>>
>>> ##### MONITORS ######
>>> # Monitor classes available to the compute which may be
>>> # specified more than once. (multi valued)
>>> compute_available_monitors=nova.compute.monitors.all_monitors
>>>
>>> # A list of monitors that can be used for getting compute
>>> # metrics. (list value)
>>> compute_monitors=
>>>
>>> ##### VOLUMES #####
>>> # iscsi target user-land tool to use
>>> iscsi_helper=tgtadm
>>> volume_api_class=nova.volume.cinder.API
>>> # Region name of this node (string value)
>>> os_region_name=RegionOne
>>>
>>> # Override the default dnsmasq settings with this file (String value)
>>> dnsmasq_config_file=
>>>
>>> ##### THIRD PARTY ADDITIONS #####
>>>
>>>
>>> [ssl]
>>>
>>> # CA certificate file to use to verify connecting clients
>>> ca_file=
>>>
>>> # Certificate file to use when starting the server securely
>>> cert_file=
>>>
>>> # Private key file to use when starting the server securely
>>> key_file=
>>>
>>> [conductor]
>>>
>>> use_local=False
>>>
>>>
>>> [libvirt]
>>>
>>> #
>>> # Options defined in nova.virt.libvirt.driver
>>> #
>>>
>>> # Rescue ami image (string value)
>>> #rescue_image_id=<None>
>>>
>>> # Rescue aki image (string value)
>>> #rescue_kernel_id=<None>
>>>
>>> # Rescue ari image (string value)
>>> #rescue_ramdisk_id=<None>
>>>
>>> # Libvirt domain type (valid options are: kvm, lxc, qemu, uml,
>>> # xen) (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_type
>>> virt_type=kvm
>>>
>>> # Override the default libvirt URI (which is dependent on
>>> # virt_type) (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_uri
>>> #connection_uri=
>>>
>>> # Inject the admin password at boot time, without an agent.
>>> # (boolean value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_password
>>> inject_password=false
>>>
>>> # Inject the ssh public key at boot time (boolean value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_key
>>> inject_key=true
>>>
>>> # The partition to inject to : -2 => disable, -1 => inspect
>>> # (libguestfs only), 0 => not partitioned, >0 => partition
>>> # number (integer value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_partition
>>> inject_partition=-2
>>>
>>> # Sync virtual and real mouse cursors in Windows VMs (boolean
>>> # value)
>>> #use_usb_tablet=true
>>>
>>> # Migration target URI (any included "%s" is replaced with the
>>> # migration target hostname) (string value)
>>> live_migration_uri=qemu+tcp://%s/system
>>>
>>> # Migration flags to be set for live migration (string value)
>>> live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER
>>>
>>> # Migration flags to be set for block migration (string value)
>>> block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER,
>>> VIR_MIGRATE_NON_SHARED_INC
>>>
>>> # Maximum bandwidth to be used during migration, in Mbps
>>> # (integer value)
>>> live_migration_bandwidth=0
>>>
>>> # Snapshot image format (valid options are : raw, qcow2, vmdk,
>>> # vdi). Defaults to same as source image (string value)
>>> snapshot_image_format=qcow2
>>>
>>> # The libvirt VIF driver to configure the VIFs. (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_vif_driver
>>> vif_driver=nova.virt.libvirt.vif.LibvirtGenericVIFDriver
>>>
>>> # Libvirt handlers for remote volumes. (list value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_volume_drivers
>>>
>>> #volume_drivers=iscsi=nova.virt.libvirt.volume.LibvirtISCSIVolumeDriver,iser=nova.virt.libvirt.volume.LibvirtISERVolumeDriver,local=nova.virt.libvirt.volume.LibvirtVolumeDriver,fake=nova.virt.libvirt.volume.LibvirtFakeVolumeDriver,rbd=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,sheepdog=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,nfs=nova.virt.libvirt.volume.LibvirtNFSVolumeDriver,aoe=nova.virt.libvirt.volume.LibvirtAOEVolumeDriver,glusterfs=nova.virt.libvirt.volume.LibvirtGlusterfsVolumeDriver,fibre_channel=nova.virt.libvirt.volume.LibvirtFibreChannelVolumeDriver,scality=nova.virt.libvirt.volume.LibvirtScalityVolumeDriver
>>>
>>> # Override the default disk prefix for the devices attached to
>>> # a server, which is dependent on virt_type. (valid options
>>> # are: sd, xvd, uvd, vd) (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_disk_prefix
>>> #disk_prefix=<None>
>>>
>>> # Number of seconds to wait for instance to shut down after
>>> # soft reboot request is made. We fall back to hard reboot if
>>> # instance does not shutdown within this window. (integer
>>> # value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_wait_soft_reboot_seconds
>>> #wait_soft_reboot_seconds=120
>>>
>>> # Set to "host-model" to clone the host CPU feature flags; to
>>> # "host-passthrough" to use the host CPU model exactly; to
>>> # "custom" to use a named CPU model; to "none" to not set any
>>> # CPU model. If virt_type="kvm|qemu", it will default to
>>> # "host-model", otherwise it will default to "none" (string
>>> # value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_mode
>>>
>>> # Set to a named libvirt CPU model (see names listed in
>>> # /usr/share/libvirt/cpu_map.xml). Only has effect if
>>> # cpu_mode="custom" and virt_type="kvm|qemu" (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_model
>>> #cpu_model=<none>
>>>
>>> # Location where libvirt driver will store snapshots before
>>> # uploading them to image service (string value)
>>> # Deprecated group/name - [DEFAULT]/libvirt_snapshots_directory
>>> #snapshots_directory=$instances_path/snapshots
>>>
>>> # Location where the Xen hvmloader is kept (string value)
>>> #xen_hvmloader_path=/usr/lib/xen/boot/hvmloader
>>>
>>> # Specific cachemodes to use for different disk types e.g:
>>> # file=directsync,block=none (list value)
>>>
>>> # A path to a device that will be used as source of entropy on
>>> # the host. Permitted options are: /dev/random or /dev/hwrng
>>> # (string value)
>>>
>>> #
>>> # Options defined in nova.virt.libvirt.imagecache
>>> #
>>>
>>> # Unused resized base images younger than this will not be removed
>>> (default: 3600)
>>> remove_unused_resized_minimum_age_seconds=3600
>>>
>>> # Write a checksum for files in _base to disk (default: false)
>>> checksum_base_images=false
>>>
>>> #
>>> # Options defined in nova.virt.libvirt.vif
>>> #
>>>
>>> use_virtio_for_bridges=true
>>>
>>> #
>>> # Options defined in nova.virt.libvirt.imagebackend
>>> #
>>>
>>> # VM Images format. Acceptable values are: raw, qcow2, lvm, rbd,
>>> default. If default is specified,
>>> # then use_cow_images flag is used instead of this one.
>>> images_type=default
>>>
>>>
>>> [keystone_authtoken]
>>> auth_uri = http://controller:5000/v2.0
>>> auth_host = controller
>>> auth_port = 35357
>>> auth_protocol = http
>>> auth_version = v2.0
>>> admin_tenant_name = service
>>> admin_user = nova
>>> admin_password = openstack-compute
>>> signing_dir = /var/cache/nova/api
>>> hash_algorithms = md5
>>> insecure = false
>>> ========================================
>>>
>>>
>>> Please check it.
>>>
>>>
>>> On Wed, Jan 14, 2015 at 8:23 PM, Jay Pipes <jaypi...@gmail.com> wrote:
>>>
>>>> Could you pastebin the output of:
>>>>
>>>>  keystone catalog
>>>>
>>>> and also pastebin your nova.conf for the node running the Nova API
>>>> service?
>>>>
>>>> Thanks!
>>>> -jay
>>>>
>>>>
>>>> On 01/14/2015 02:25 AM, Geo Varghese wrote:
>>>>
>>>>> Hi Team,
>>>>>
>>>>> I need a help with cinder volume attachment with an instance.
>>>>>
>>>>> I have succesfully created cinder volume and it is in available state.
>>>>> Please check attached screenshot.
>>>>>
>>>>> Later I tried to attach volume to an instance but attachment failed.
>>>>> While checking logs in /var/log/nova/nova-api-os-compute.log
>>>>>
>>>>> ========
>>>>>
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     res =
>>>>> method(self, ctx, volume_id, *args, **kwargs)
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 206,
>>>>> in get
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     item =
>>>>> cinderclient(context).volumes.get(volume_id)
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 91, in
>>>>> cinderclient
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>>>> endpoint_type=endpoint_type)
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>>> "/usr/lib/python2.7/dist-packages/cinderclient/service_catalog.py",
>>>>> line
>>>>> 80, in url_for
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     raise
>>>>> cinderclient.exceptions.EndpointNotFound()
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack EndpointNotFound
>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>>>> =========================================
>>>>>
>>>>>
>>>>> I have already created endpints in v1 and v2.
>>>>> Please check endpints I have created for cinder below
>>>>>
>>>>> ============================================================
>>>>> =============
>>>>> root@controller:/home/geo# keystone endpoint-list | grep 8776
>>>>> | 5c7bcc79daa74532ac9ca19949e0d872 | regionOne |
>>>>> http://controller:8776/v1/%(tenant_id)s |
>>>>> http://controller:8776/v1/%(tenant_id)s |
>>>>> http://controller:8776/v1/%(tenant_id)s |
>>>>> 8ce0898aa7c84fec9b011823d34b55cb |
>>>>> | 5d71e0a1237c483990b84c36346602b4 | RegionOne |
>>>>> http://controller:8776/v2/%(tenant_id)s |
>>>>> http://controller:8776/v2/%(tenant_id)s |
>>>>> http://controller:8776/v2/%(tenant_id)s |
>>>>> 251eca5fdb6b4550a9f521c10fa9f2ca |
>>>>> ============================================================
>>>>> ===================
>>>>>
>>>>> Anyone please help me. Thanks for the support guys.
>>>>>
>>>>> --
>>>>> Regards,
>>>>> Geo Varghese
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> OpenStack-operators mailing list
>>>>> OpenStack-operators@lists.openstack.org
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/
>>>>> openstack-operators
>>>>>
>>>>>
>>>> _______________________________________________
>>>> OpenStack-operators mailing list
>>>> OpenStack-operators@lists.openstack.org
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>
>>>
>>>
>>>
>>> --
>>> --
>>> Regards,
>>> Geo Varghese
>>>
>>
>
>
> --
> --
> Regards,
> Geo Varghese
>
_______________________________________________
OpenStack-operators mailing list
OpenStack-operators@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators

Reply via email to