I know it's "Available ", however that doesn't imply attachment. Cinder uses iSCSI or NFS, to attach the volume to a running instance on a compute node. If you're missing the required protocol packages, the attachment will fail. You can have "Available " volumes, and lack tgtadm (or nfs-utils if that's your protocol).
Secondly, Is your compute node able to resolve "controller"? On Wednesday, January 14, 2015, Geo Varghese <gvargh...@aqorn.com> wrote: > Hi Abel, > > Thanks for the reply. > > I have created volume and its in available state. Please check attached > screenshot. > > > > On Thu, Jan 15, 2015 at 11:34 AM, Abel Lopez <alopg...@gmail.com > <javascript:_e(%7B%7D,'cvml','alopg...@gmail.com');>> wrote: > >> Do your compute nodes have the required iSCSI packages installed? >> >> >> On Wednesday, January 14, 2015, Geo Varghese <gvargh...@aqorn.com >> <javascript:_e(%7B%7D,'cvml','gvargh...@aqorn.com');>> wrote: >> >>> Hi Jay, >>> >>> Thanks for the reply. Just pasting the details below >>> >>> keystone catalog >>> ================================================ >>> Service: compute >>> >>> +-------------+------------------------------------------------------------+ >>> | Property | >>> Value | >>> >>> +-------------+------------------------------------------------------------+ >>> | adminURL | >>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | id | >>> 02028b1f4c0849c68eb79f5887516299 | >>> | internalURL | >>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | publicURL | >>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | region | >>> RegionOne | >>> >>> +-------------+------------------------------------------------------------+ >>> Service: network >>> +-------------+----------------------------------+ >>> | Property | Value | >>> +-------------+----------------------------------+ >>> | adminURL | http://controller:9696 | >>> | id | 32f687d4f7474769852d88932288b893 | >>> | internalURL | http://controller:9696 | >>> | publicURL | http://controller:9696 | >>> | region | RegionOne | >>> +-------------+----------------------------------+ >>> Service: volumev2 >>> >>> +-------------+------------------------------------------------------------+ >>> | Property | >>> Value | >>> >>> +-------------+------------------------------------------------------------+ >>> | adminURL | >>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | id | >>> 5bca493cdde2439887d54fb805c4d2d4 | >>> | internalURL | >>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | publicURL | >>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>> | region | >>> RegionOne | >>> >>> +-------------+------------------------------------------------------------+ >>> Service: image >>> +-------------+----------------------------------+ >>> | Property | Value | >>> +-------------+----------------------------------+ >>> | adminURL | http://controller:9292 | >>> | id | 2e2294b9151e4fb9b6efccf33c62181b | >>> | internalURL | http://controller:9292 | >>> | publicURL | http://controller:9292 | >>> | region | RegionOne | >>> +-------------+----------------------------------+ >>> Service: volume >>> >>> +-------------+------------------------------------------------------------+ >>> | Property | >>> Value | >>> >>> +-------------+------------------------------------------------------------+ >>> | adminURL | >>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>> | id | >>> 0e29cfaa785e4e148c57601b182a5e26 | >>> | internalURL | >>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>> | publicURL | >>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>> | region | >>> RegionOne | >>> >>> +-------------+------------------------------------------------------------+ >>> Service: ec2 >>> +-------------+---------------------------------------+ >>> | Property | Value | >>> +-------------+---------------------------------------+ >>> | adminURL | http://controller:8773/services/Admin | >>> | id | 8f4957d98cd04130b055b8b80b051833 | >>> | internalURL | http://controller:8773/services/Cloud | >>> | publicURL | http://controller:8773/services/Cloud | >>> | region | RegionOne | >>> +-------------+---------------------------------------+ >>> Service: identity >>> +-------------+----------------------------------+ >>> | Property | Value | >>> +-------------+----------------------------------+ >>> | adminURL | http://controller:35357/v2.0 | >>> | id | 146f7bbb0ad54740b95f8499f04b2ee2 | >>> | internalURL | http://controller:5000/v2.0 | >>> | publicURL | http://controller:5000/v2.0 | >>> | region | RegionOne | >>> +-------------+----------------------------------+ >>> ============================================== >>> >>> Nova.conf >>> >>> ================================================ >>> # This file autogenerated by Chef >>> # Do not edit, changes will be overwritten >>> >>> >>> [DEFAULT] >>> >>> # LOGS/STATE >>> debug=False >>> verbose=False >>> auth_strategy=keystone >>> dhcpbridge_flagfile=/etc/nova/nova.conf >>> dhcpbridge=/usr/bin/nova-dhcpbridge >>> log_dir=/var/log/nova >>> state_path=/var/lib/nova >>> instances_path=/var/lib/nova/instances >>> instance_name_template=instance-%08x >>> network_allocate_retries=0 >>> lock_path=/var/lib/nova/lock >>> >>> ssl_only=false >>> cert=self.pem >>> key= >>> >>> # Command prefix to use for running commands as root (default: sudo) >>> rootwrap_config=/etc/nova/rootwrap.conf >>> >>> # Should unused base images be removed? (default: false) >>> remove_unused_base_images=true >>> >>> # Unused unresized base images younger than this will not be removed >>> (default: 86400) >>> remove_unused_original_minimum_age_seconds=3600 >>> >>> # Options defined in nova.openstack.common.rpc >>> rpc_thread_pool_size=64 >>> rpc_conn_pool_size=30 >>> rpc_response_timeout=60 >>> rpc_backend=nova.openstack.common.rpc.impl_kombu >>> amqp_durable_queues=false >>> amqp_auto_delete=false >>> >>> ##### RABBITMQ ##### >>> rabbit_userid=guest >>> rabbit_password=guest >>> rabbit_virtual_host=/ >>> rabbit_hosts=rabbit1:5672,rabbit2:5672 >>> rabbit_retry_interval=1 >>> rabbit_retry_backoff=2 >>> rabbit_max_retries=0 >>> rabbit_durable_queues=false >>> rabbit_ha_queues=True >>> >>> >>> >>> ##### SCHEDULER ##### >>> scheduler_manager=nova.scheduler.manager.SchedulerManager >>> scheduler_driver=nova.scheduler.filter_scheduler.FilterScheduler >>> scheduler_available_filters=nova.scheduler.filters.all_filters >>> # which filter class names to use for filtering hosts when not specified >>> in the request. >>> >>> scheduler_default_filters=AvailabilityZoneFilter,RamFilter,ComputeFilter,CoreFilter,SameHostFilter,DifferentHostFilter >>> default_availability_zone=nova >>> default_schedule_zone=nova >>> >>> ##### NETWORK ##### >>> >>> >>> >>> # N.B. due to https://bugs.launchpad.net/nova/+bug/1206330 >>> # we override the endpoint scheme below, ignore the port >>> # and essentially force http >>> neutron_url=http://controller:9696 >>> neutron_api_insecure=false >>> network_api_class=nova.network.neutronv2.api.API >>> neutron_auth_strategy=keystone >>> neutron_admin_tenant_name=service >>> neutron_admin_username=neutron >>> neutron_admin_password=openstack-network >>> neutron_admin_auth_url=http://controller:5000/v2.0 >>> neutron_url_timeout=30 >>> neutron_region_name= >>> neutron_ovs_bridge=br-int >>> neutron_extension_sync_interval=600 >>> neutron_ca_certificates_file= >>> linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver >>> firewall_driver = nova.virt.firewall.NoopFirewallDriver >>> security_group_api=neutron >>> service_neutron_metadata_proxy=true >>> neutron_metadata_proxy_shared_secret=secret123 >>> default_floating_pool=public >>> dns_server=8.8.8.8 >>> >>> use_ipv6=false >>> >>> ##### GLANCE ##### >>> image_service=nova.image.glance.GlanceImageService >>> glance_api_servers=http://controller:9292 >>> glance_api_insecure=false >>> >>> ##### Cinder ##### >>> # Location of ca certificates file to use for cinder client requests >>> cinder_ca_certificates_file= >>> >>> # Allow to perform insecure SSL requests to cinder >>> cinder_api_insecure=false >>> >>> # Info to match when looking for cinder in the service catalog >>> cinder_catalog_info=volumev2:cinderv2:publicURL >>> >>> ##### COMPUTE ##### >>> compute_driver=libvirt.LibvirtDriver >>> preallocate_images=none >>> use_cow_images=true >>> vif_plugging_is_fatal=false >>> vif_plugging_timeout=0 >>> compute_manager=nova.compute.manager.ComputeManager >>> sql_connection=mysql://nova:nova@loadbalancer:3306/nova?charset=utf8 >>> connection_type=libvirt >>> >>> ##### NOTIFICATIONS ##### >>> # Driver or drivers to handle sending notifications (multi valued) >>> >>> # AMQP topic used for OpenStack notifications. (list value) >>> # Deprecated group/name - [rpc_notifier2]/topics >>> notification_topics=notifications >>> >>> # Generate periodic compute.instance.exists notifications >>> instance_usage_audit=False >>> >>> # Time period to generate instance usages for. Time period >>> # must be hour, day, month or year (string value) >>> instance_usage_audit_period=month >>> >>> >>> # The IP address on which the OpenStack API will listen. (string value) >>> osapi_compute_listen=0.0.0.0 >>> # The port on which the OpenStack API will listen. (integer value) >>> osapi_compute_listen_port=8774 >>> >>> # The IP address on which the metadata will listen. (string value) >>> metadata_listen=0.0.0.0 >>> # The port on which the metadata will listen. (integer value) >>> metadata_listen_port=8775 >>> >>> ##### VNCPROXY ##### >>> novncproxy_base_url=http://controller:6080/vnc_auto.html >>> xvpvncproxy_base_url=http://controller:6081/console >>> >>> # This is only required on the server running xvpvncproxy >>> xvpvncproxy_host=0.0.0.0 >>> xvpvncproxy_port=6081 >>> >>> # This is only required on the server running novncproxy >>> novncproxy_host=0.0.0.0 >>> novncproxy_port=6080 >>> >>> vncserver_listen=0.0.0.0 >>> vncserver_proxyclient_address=0.0.0.0 >>> >>> vnc_keymap=en-us >>> >>> # store consoleauth tokens in memcached >>> >>> ##### MISC ##### >>> # force backing images to raw format >>> force_raw_images=false >>> allow_same_net_traffic=true >>> osapi_max_limit=1000 >>> # If you terminate SSL with a load balancer, the HTTP_HOST environ >>> # variable that generates the request_uri in webob.Request will lack >>> # the HTTPS scheme. Setting this overrides the default and allows >>> # URIs returned in the various links collections to contain the proper >>> # HTTPS endpoint. >>> osapi_compute_link_prefix = http://controller:8774/v2/%(tenant_id)s >>> start_guests_on_host_boot=false >>> resume_guests_state_on_host_boot=true >>> allow_resize_to_same_host=false >>> resize_confirm_window=0 >>> live_migration_retry_count=30 >>> >>> ##### QUOTAS ##### >>> # (StrOpt) default driver to use for quota checks (default: >>> nova.quota.DbQuotaDriver) >>> quota_driver=nova.quota.DbQuotaDriver >>> # number of security groups per project (default: 10) >>> quota_security_groups=50 >>> # number of security rules per security group (default: 20) >>> quota_security_group_rules=20 >>> # number of instance cores allowed per project (default: 20) >>> quota_cores=20 >>> # number of fixed ips allowed per project (this should be at least the >>> number of instances allowed) (default: -1) >>> quota_fixed_ips=-1 >>> # number of floating ips allowed per project (default: 10) >>> quota_floating_ips=10 >>> # number of bytes allowed per injected file (default: 10240) >>> quota_injected_file_content_bytes=10240 >>> # number of bytes allowed per injected file path (default: 255) >>> quota_injected_file_path_length=255 >>> # number of injected files allowed (default: 5) >>> quota_injected_files=5 >>> # number of instances allowed per project (defailt: 10) >>> quota_instances=10 >>> # number of key pairs per user (default: 100) >>> quota_key_pairs=100 >>> # number of metadata items allowed per instance (default: 128) >>> quota_metadata_items=128 >>> # megabytes of instance ram allowed per project (default: 51200) >>> quota_ram=51200 >>> >>> # virtual CPU to Physical CPU allocation ratio (default: 16.0) >>> cpu_allocation_ratio=16.0 >>> # virtual ram to physical ram allocation ratio (default: 1.5) >>> ram_allocation_ratio=1.5 >>> >>> mkisofs_cmd=genisoimage >>> injected_network_template=$pybasedir/nova/virt/interfaces.template >>> flat_injected=false >>> >>> # The IP address on which the EC2 API will listen. (string value) >>> ec2_listen=0.0.0.0 >>> # The port on which the EC2 API will listen. (integer value) >>> ec2_listen_port=8773 >>> >>> >>> ##### WORKERS ###### >>> >>> ##### KEYSTONE ##### >>> keystone_ec2_url=http://controller:5000/v2.0/ec2tokens >>> >>> # a list of APIs to enable by default (list value) >>> enabled_apis=ec2,osapi_compute,metadata >>> >>> ##### WORKERS ###### >>> >>> ##### MONITORS ###### >>> # Monitor classes available to the compute which may be >>> # specified more than once. (multi valued) >>> compute_available_monitors=nova.compute.monitors.all_monitors >>> >>> # A list of monitors that can be used for getting compute >>> # metrics. (list value) >>> compute_monitors= >>> >>> ##### VOLUMES ##### >>> # iscsi target user-land tool to use >>> iscsi_helper=tgtadm >>> volume_api_class=nova.volume.cinder.API >>> # Region name of this node (string value) >>> os_region_name=RegionOne >>> >>> # Override the default dnsmasq settings with this file (String value) >>> dnsmasq_config_file= >>> >>> ##### THIRD PARTY ADDITIONS ##### >>> >>> >>> [ssl] >>> >>> # CA certificate file to use to verify connecting clients >>> ca_file= >>> >>> # Certificate file to use when starting the server securely >>> cert_file= >>> >>> # Private key file to use when starting the server securely >>> key_file= >>> >>> [conductor] >>> >>> use_local=False >>> >>> >>> [libvirt] >>> >>> # >>> # Options defined in nova.virt.libvirt.driver >>> # >>> >>> # Rescue ami image (string value) >>> #rescue_image_id=<None> >>> >>> # Rescue aki image (string value) >>> #rescue_kernel_id=<None> >>> >>> # Rescue ari image (string value) >>> #rescue_ramdisk_id=<None> >>> >>> # Libvirt domain type (valid options are: kvm, lxc, qemu, uml, >>> # xen) (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_type >>> virt_type=kvm >>> >>> # Override the default libvirt URI (which is dependent on >>> # virt_type) (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_uri >>> #connection_uri= >>> >>> # Inject the admin password at boot time, without an agent. >>> # (boolean value) >>> # Deprecated group/name - [DEFAULT]/libvirt_inject_password >>> inject_password=false >>> >>> # Inject the ssh public key at boot time (boolean value) >>> # Deprecated group/name - [DEFAULT]/libvirt_inject_key >>> inject_key=true >>> >>> # The partition to inject to : -2 => disable, -1 => inspect >>> # (libguestfs only), 0 => not partitioned, >0 => partition >>> # number (integer value) >>> # Deprecated group/name - [DEFAULT]/libvirt_inject_partition >>> inject_partition=-2 >>> >>> # Sync virtual and real mouse cursors in Windows VMs (boolean >>> # value) >>> #use_usb_tablet=true >>> >>> # Migration target URI (any included "%s" is replaced with the >>> # migration target hostname) (string value) >>> live_migration_uri=qemu+tcp://%s/system >>> >>> # Migration flags to be set for live migration (string value) >>> live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER >>> >>> # Migration flags to be set for block migration (string value) >>> block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER, >>> VIR_MIGRATE_NON_SHARED_INC >>> >>> # Maximum bandwidth to be used during migration, in Mbps >>> # (integer value) >>> live_migration_bandwidth=0 >>> >>> # Snapshot image format (valid options are : raw, qcow2, vmdk, >>> # vdi). Defaults to same as source image (string value) >>> snapshot_image_format=qcow2 >>> >>> # The libvirt VIF driver to configure the VIFs. (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_vif_driver >>> vif_driver=nova.virt.libvirt.vif.LibvirtGenericVIFDriver >>> >>> # Libvirt handlers for remote volumes. (list value) >>> # Deprecated group/name - [DEFAULT]/libvirt_volume_drivers >>> >>> #volume_drivers=iscsi=nova.virt.libvirt.volume.LibvirtISCSIVolumeDriver,iser=nova.virt.libvirt.volume.LibvirtISERVolumeDriver,local=nova.virt.libvirt.volume.LibvirtVolumeDriver,fake=nova.virt.libvirt.volume.LibvirtFakeVolumeDriver,rbd=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,sheepdog=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,nfs=nova.virt.libvirt.volume.LibvirtNFSVolumeDriver,aoe=nova.virt.libvirt.volume.LibvirtAOEVolumeDriver,glusterfs=nova.virt.libvirt.volume.LibvirtGlusterfsVolumeDriver,fibre_channel=nova.virt.libvirt.volume.LibvirtFibreChannelVolumeDriver,scality=nova.virt.libvirt.volume.LibvirtScalityVolumeDriver >>> >>> # Override the default disk prefix for the devices attached to >>> # a server, which is dependent on virt_type. (valid options >>> # are: sd, xvd, uvd, vd) (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_disk_prefix >>> #disk_prefix=<None> >>> >>> # Number of seconds to wait for instance to shut down after >>> # soft reboot request is made. We fall back to hard reboot if >>> # instance does not shutdown within this window. (integer >>> # value) >>> # Deprecated group/name - [DEFAULT]/libvirt_wait_soft_reboot_seconds >>> #wait_soft_reboot_seconds=120 >>> >>> # Set to "host-model" to clone the host CPU feature flags; to >>> # "host-passthrough" to use the host CPU model exactly; to >>> # "custom" to use a named CPU model; to "none" to not set any >>> # CPU model. If virt_type="kvm|qemu", it will default to >>> # "host-model", otherwise it will default to "none" (string >>> # value) >>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_mode >>> >>> # Set to a named libvirt CPU model (see names listed in >>> # /usr/share/libvirt/cpu_map.xml). Only has effect if >>> # cpu_mode="custom" and virt_type="kvm|qemu" (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_model >>> #cpu_model=<none> >>> >>> # Location where libvirt driver will store snapshots before >>> # uploading them to image service (string value) >>> # Deprecated group/name - [DEFAULT]/libvirt_snapshots_directory >>> #snapshots_directory=$instances_path/snapshots >>> >>> # Location where the Xen hvmloader is kept (string value) >>> #xen_hvmloader_path=/usr/lib/xen/boot/hvmloader >>> >>> # Specific cachemodes to use for different disk types e.g: >>> # file=directsync,block=none (list value) >>> >>> # A path to a device that will be used as source of entropy on >>> # the host. Permitted options are: /dev/random or /dev/hwrng >>> # (string value) >>> >>> # >>> # Options defined in nova.virt.libvirt.imagecache >>> # >>> >>> # Unused resized base images younger than this will not be removed >>> (default: 3600) >>> remove_unused_resized_minimum_age_seconds=3600 >>> >>> # Write a checksum for files in _base to disk (default: false) >>> checksum_base_images=false >>> >>> # >>> # Options defined in nova.virt.libvirt.vif >>> # >>> >>> use_virtio_for_bridges=true >>> >>> # >>> # Options defined in nova.virt.libvirt.imagebackend >>> # >>> >>> # VM Images format. Acceptable values are: raw, qcow2, lvm, rbd, >>> default. If default is specified, >>> # then use_cow_images flag is used instead of this one. >>> images_type=default >>> >>> >>> [keystone_authtoken] >>> auth_uri = http://controller:5000/v2.0 >>> auth_host = controller >>> auth_port = 35357 >>> auth_protocol = http >>> auth_version = v2.0 >>> admin_tenant_name = service >>> admin_user = nova >>> admin_password = openstack-compute >>> signing_dir = /var/cache/nova/api >>> hash_algorithms = md5 >>> insecure = false >>> ======================================== >>> >>> >>> Please check it. >>> >>> >>> On Wed, Jan 14, 2015 at 8:23 PM, Jay Pipes <jaypi...@gmail.com> wrote: >>> >>>> Could you pastebin the output of: >>>> >>>> keystone catalog >>>> >>>> and also pastebin your nova.conf for the node running the Nova API >>>> service? >>>> >>>> Thanks! >>>> -jay >>>> >>>> >>>> On 01/14/2015 02:25 AM, Geo Varghese wrote: >>>> >>>>> Hi Team, >>>>> >>>>> I need a help with cinder volume attachment with an instance. >>>>> >>>>> I have succesfully created cinder volume and it is in available state. >>>>> Please check attached screenshot. >>>>> >>>>> Later I tried to attach volume to an instance but attachment failed. >>>>> While checking logs in /var/log/nova/nova-api-os-compute.log >>>>> >>>>> ======== >>>>> >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack res = >>>>> method(self, ctx, volume_id, *args, **kwargs) >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 206, >>>>> in get >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack item = >>>>> cinderclient(context).volumes.get(volume_id) >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 91, in >>>>> cinderclient >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack >>>>> endpoint_type=endpoint_type) >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>> "/usr/lib/python2.7/dist-packages/cinderclient/service_catalog.py", >>>>> line >>>>> 80, in url_for >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack raise >>>>> cinderclient.exceptions.EndpointNotFound() >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack EndpointNotFound >>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack >>>>> ========================================= >>>>> >>>>> >>>>> I have already created endpints in v1 and v2. >>>>> Please check endpints I have created for cinder below >>>>> >>>>> ============================================================ >>>>> ============= >>>>> root@controller:/home/geo# keystone endpoint-list | grep 8776 >>>>> | 5c7bcc79daa74532ac9ca19949e0d872 | regionOne | >>>>> http://controller:8776/v1/%(tenant_id)s | >>>>> http://controller:8776/v1/%(tenant_id)s | >>>>> http://controller:8776/v1/%(tenant_id)s | >>>>> 8ce0898aa7c84fec9b011823d34b55cb | >>>>> | 5d71e0a1237c483990b84c36346602b4 | RegionOne | >>>>> http://controller:8776/v2/%(tenant_id)s | >>>>> http://controller:8776/v2/%(tenant_id)s | >>>>> http://controller:8776/v2/%(tenant_id)s | >>>>> 251eca5fdb6b4550a9f521c10fa9f2ca | >>>>> ============================================================ >>>>> =================== >>>>> >>>>> Anyone please help me. Thanks for the support guys. >>>>> >>>>> -- >>>>> Regards, >>>>> Geo Varghese >>>>> >>>>> >>>>> _______________________________________________ >>>>> OpenStack-operators mailing list >>>>> OpenStack-operators@lists.openstack.org >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/ >>>>> openstack-operators >>>>> >>>>> >>>> _______________________________________________ >>>> OpenStack-operators mailing list >>>> OpenStack-operators@lists.openstack.org >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>>> >>> >>> >>> >>> -- >>> -- >>> Regards, >>> Geo Varghese >>> >> > > > -- > -- > Regards, > Geo Varghese >
_______________________________________________ OpenStack-operators mailing list OpenStack-operators@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators