Hi Jay/Abel, Thanks for your help.
Just fixed issue by changing following line in nova.conf cinder_catalog_info=volumev2:cinderv2:publicURL to cinder_catalog_info=volume:cinder:publicURL Now attachment successfuly done. Do guys know how this fixed the issue? On Thu, Jan 15, 2015 at 12:01 PM, Geo Varghese <[email protected]> wrote: > Hi Abel, > > Oh okay. yes sure compute nod ecan access the controller. I have added it > in /etc/hosts > > The current error is, it couldn't find an endpoint. is this related with > anything aboe you mentioned? > > On Thu, Jan 15, 2015 at 11:56 AM, Abel Lopez <[email protected]> wrote: > >> I know it's "Available ", however that doesn't imply attachment. Cinder >> uses iSCSI or NFS, to attach the volume to a running instance on a compute >> node. If you're missing the required protocol packages, the attachment will >> fail. You can have "Available " volumes, and lack tgtadm (or nfs-utils if >> that's your protocol). >> >> Secondly, Is your compute node able to resolve "controller"? >> >> >> On Wednesday, January 14, 2015, Geo Varghese <[email protected]> wrote: >> >>> Hi Abel, >>> >>> Thanks for the reply. >>> >>> I have created volume and its in available state. Please check attached >>> screenshot. >>> >>> >>> >>> On Thu, Jan 15, 2015 at 11:34 AM, Abel Lopez <[email protected]> wrote: >>> >>>> Do your compute nodes have the required iSCSI packages installed? >>>> >>>> >>>> On Wednesday, January 14, 2015, Geo Varghese <[email protected]> >>>> wrote: >>>> >>>>> Hi Jay, >>>>> >>>>> Thanks for the reply. Just pasting the details below >>>>> >>>>> keystone catalog >>>>> ================================================ >>>>> Service: compute >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | Property | >>>>> Value | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | adminURL | >>>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | id | >>>>> 02028b1f4c0849c68eb79f5887516299 | >>>>> | internalURL | >>>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | publicURL | >>>>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | region | >>>>> RegionOne | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> Service: network >>>>> +-------------+----------------------------------+ >>>>> | Property | Value | >>>>> +-------------+----------------------------------+ >>>>> | adminURL | http://controller:9696 | >>>>> | id | 32f687d4f7474769852d88932288b893 | >>>>> | internalURL | http://controller:9696 | >>>>> | publicURL | http://controller:9696 | >>>>> | region | RegionOne | >>>>> +-------------+----------------------------------+ >>>>> Service: volumev2 >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | Property | >>>>> Value | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | adminURL | >>>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | id | >>>>> 5bca493cdde2439887d54fb805c4d2d4 | >>>>> | internalURL | >>>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | publicURL | >>>>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 | >>>>> | region | >>>>> RegionOne | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> Service: image >>>>> +-------------+----------------------------------+ >>>>> | Property | Value | >>>>> +-------------+----------------------------------+ >>>>> | adminURL | http://controller:9292 | >>>>> | id | 2e2294b9151e4fb9b6efccf33c62181b | >>>>> | internalURL | http://controller:9292 | >>>>> | publicURL | http://controller:9292 | >>>>> | region | RegionOne | >>>>> +-------------+----------------------------------+ >>>>> Service: volume >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | Property | >>>>> Value | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> | adminURL | >>>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>>>> | id | >>>>> 0e29cfaa785e4e148c57601b182a5e26 | >>>>> | internalURL | >>>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>>>> | publicURL | >>>>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 | >>>>> | region | >>>>> RegionOne | >>>>> >>>>> +-------------+------------------------------------------------------------+ >>>>> Service: ec2 >>>>> +-------------+---------------------------------------+ >>>>> | Property | Value | >>>>> +-------------+---------------------------------------+ >>>>> | adminURL | http://controller:8773/services/Admin | >>>>> | id | 8f4957d98cd04130b055b8b80b051833 | >>>>> | internalURL | http://controller:8773/services/Cloud | >>>>> | publicURL | http://controller:8773/services/Cloud | >>>>> | region | RegionOne | >>>>> +-------------+---------------------------------------+ >>>>> Service: identity >>>>> +-------------+----------------------------------+ >>>>> | Property | Value | >>>>> +-------------+----------------------------------+ >>>>> | adminURL | http://controller:35357/v2.0 | >>>>> | id | 146f7bbb0ad54740b95f8499f04b2ee2 | >>>>> | internalURL | http://controller:5000/v2.0 | >>>>> | publicURL | http://controller:5000/v2.0 | >>>>> | region | RegionOne | >>>>> +-------------+----------------------------------+ >>>>> ============================================== >>>>> >>>>> Nova.conf >>>>> >>>>> ================================================ >>>>> # This file autogenerated by Chef >>>>> # Do not edit, changes will be overwritten >>>>> >>>>> >>>>> [DEFAULT] >>>>> >>>>> # LOGS/STATE >>>>> debug=False >>>>> verbose=False >>>>> auth_strategy=keystone >>>>> dhcpbridge_flagfile=/etc/nova/nova.conf >>>>> dhcpbridge=/usr/bin/nova-dhcpbridge >>>>> log_dir=/var/log/nova >>>>> state_path=/var/lib/nova >>>>> instances_path=/var/lib/nova/instances >>>>> instance_name_template=instance-%08x >>>>> network_allocate_retries=0 >>>>> lock_path=/var/lib/nova/lock >>>>> >>>>> ssl_only=false >>>>> cert=self.pem >>>>> key= >>>>> >>>>> # Command prefix to use for running commands as root (default: sudo) >>>>> rootwrap_config=/etc/nova/rootwrap.conf >>>>> >>>>> # Should unused base images be removed? (default: false) >>>>> remove_unused_base_images=true >>>>> >>>>> # Unused unresized base images younger than this will not be removed >>>>> (default: 86400) >>>>> remove_unused_original_minimum_age_seconds=3600 >>>>> >>>>> # Options defined in nova.openstack.common.rpc >>>>> rpc_thread_pool_size=64 >>>>> rpc_conn_pool_size=30 >>>>> rpc_response_timeout=60 >>>>> rpc_backend=nova.openstack.common.rpc.impl_kombu >>>>> amqp_durable_queues=false >>>>> amqp_auto_delete=false >>>>> >>>>> ##### RABBITMQ ##### >>>>> rabbit_userid=guest >>>>> rabbit_password=guest >>>>> rabbit_virtual_host=/ >>>>> rabbit_hosts=rabbit1:5672,rabbit2:5672 >>>>> rabbit_retry_interval=1 >>>>> rabbit_retry_backoff=2 >>>>> rabbit_max_retries=0 >>>>> rabbit_durable_queues=false >>>>> rabbit_ha_queues=True >>>>> >>>>> >>>>> >>>>> ##### SCHEDULER ##### >>>>> scheduler_manager=nova.scheduler.manager.SchedulerManager >>>>> scheduler_driver=nova.scheduler.filter_scheduler.FilterScheduler >>>>> scheduler_available_filters=nova.scheduler.filters.all_filters >>>>> # which filter class names to use for filtering hosts when not >>>>> specified in the request. >>>>> >>>>> scheduler_default_filters=AvailabilityZoneFilter,RamFilter,ComputeFilter,CoreFilter,SameHostFilter,DifferentHostFilter >>>>> default_availability_zone=nova >>>>> default_schedule_zone=nova >>>>> >>>>> ##### NETWORK ##### >>>>> >>>>> >>>>> >>>>> # N.B. due to https://bugs.launchpad.net/nova/+bug/1206330 >>>>> # we override the endpoint scheme below, ignore the port >>>>> # and essentially force http >>>>> neutron_url=http://controller:9696 >>>>> neutron_api_insecure=false >>>>> network_api_class=nova.network.neutronv2.api.API >>>>> neutron_auth_strategy=keystone >>>>> neutron_admin_tenant_name=service >>>>> neutron_admin_username=neutron >>>>> neutron_admin_password=openstack-network >>>>> neutron_admin_auth_url=http://controller:5000/v2.0 >>>>> neutron_url_timeout=30 >>>>> neutron_region_name= >>>>> neutron_ovs_bridge=br-int >>>>> neutron_extension_sync_interval=600 >>>>> neutron_ca_certificates_file= >>>>> >>>>> linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver >>>>> firewall_driver = nova.virt.firewall.NoopFirewallDriver >>>>> security_group_api=neutron >>>>> service_neutron_metadata_proxy=true >>>>> neutron_metadata_proxy_shared_secret=secret123 >>>>> default_floating_pool=public >>>>> dns_server=8.8.8.8 >>>>> >>>>> use_ipv6=false >>>>> >>>>> ##### GLANCE ##### >>>>> image_service=nova.image.glance.GlanceImageService >>>>> glance_api_servers=http://controller:9292 >>>>> glance_api_insecure=false >>>>> >>>>> ##### Cinder ##### >>>>> # Location of ca certificates file to use for cinder client requests >>>>> cinder_ca_certificates_file= >>>>> >>>>> # Allow to perform insecure SSL requests to cinder >>>>> cinder_api_insecure=false >>>>> >>>>> # Info to match when looking for cinder in the service catalog >>>>> cinder_catalog_info=volumev2:cinderv2:publicURL >>>>> >>>>> ##### COMPUTE ##### >>>>> compute_driver=libvirt.LibvirtDriver >>>>> preallocate_images=none >>>>> use_cow_images=true >>>>> vif_plugging_is_fatal=false >>>>> vif_plugging_timeout=0 >>>>> compute_manager=nova.compute.manager.ComputeManager >>>>> sql_connection=mysql://nova:nova@loadbalancer:3306/nova?charset=utf8 >>>>> connection_type=libvirt >>>>> >>>>> ##### NOTIFICATIONS ##### >>>>> # Driver or drivers to handle sending notifications (multi valued) >>>>> >>>>> # AMQP topic used for OpenStack notifications. (list value) >>>>> # Deprecated group/name - [rpc_notifier2]/topics >>>>> notification_topics=notifications >>>>> >>>>> # Generate periodic compute.instance.exists notifications >>>>> instance_usage_audit=False >>>>> >>>>> # Time period to generate instance usages for. Time period >>>>> # must be hour, day, month or year (string value) >>>>> instance_usage_audit_period=month >>>>> >>>>> >>>>> # The IP address on which the OpenStack API will listen. (string value) >>>>> osapi_compute_listen=0.0.0.0 >>>>> # The port on which the OpenStack API will listen. (integer value) >>>>> osapi_compute_listen_port=8774 >>>>> >>>>> # The IP address on which the metadata will listen. (string value) >>>>> metadata_listen=0.0.0.0 >>>>> # The port on which the metadata will listen. (integer value) >>>>> metadata_listen_port=8775 >>>>> >>>>> ##### VNCPROXY ##### >>>>> novncproxy_base_url=http://controller:6080/vnc_auto.html >>>>> xvpvncproxy_base_url=http://controller:6081/console >>>>> >>>>> # This is only required on the server running xvpvncproxy >>>>> xvpvncproxy_host=0.0.0.0 >>>>> xvpvncproxy_port=6081 >>>>> >>>>> # This is only required on the server running novncproxy >>>>> novncproxy_host=0.0.0.0 >>>>> novncproxy_port=6080 >>>>> >>>>> vncserver_listen=0.0.0.0 >>>>> vncserver_proxyclient_address=0.0.0.0 >>>>> >>>>> vnc_keymap=en-us >>>>> >>>>> # store consoleauth tokens in memcached >>>>> >>>>> ##### MISC ##### >>>>> # force backing images to raw format >>>>> force_raw_images=false >>>>> allow_same_net_traffic=true >>>>> osapi_max_limit=1000 >>>>> # If you terminate SSL with a load balancer, the HTTP_HOST environ >>>>> # variable that generates the request_uri in webob.Request will lack >>>>> # the HTTPS scheme. Setting this overrides the default and allows >>>>> # URIs returned in the various links collections to contain the proper >>>>> # HTTPS endpoint. >>>>> osapi_compute_link_prefix = http://controller:8774/v2/%(tenant_id)s >>>>> start_guests_on_host_boot=false >>>>> resume_guests_state_on_host_boot=true >>>>> allow_resize_to_same_host=false >>>>> resize_confirm_window=0 >>>>> live_migration_retry_count=30 >>>>> >>>>> ##### QUOTAS ##### >>>>> # (StrOpt) default driver to use for quota checks (default: >>>>> nova.quota.DbQuotaDriver) >>>>> quota_driver=nova.quota.DbQuotaDriver >>>>> # number of security groups per project (default: 10) >>>>> quota_security_groups=50 >>>>> # number of security rules per security group (default: 20) >>>>> quota_security_group_rules=20 >>>>> # number of instance cores allowed per project (default: 20) >>>>> quota_cores=20 >>>>> # number of fixed ips allowed per project (this should be at least the >>>>> number of instances allowed) (default: -1) >>>>> quota_fixed_ips=-1 >>>>> # number of floating ips allowed per project (default: 10) >>>>> quota_floating_ips=10 >>>>> # number of bytes allowed per injected file (default: 10240) >>>>> quota_injected_file_content_bytes=10240 >>>>> # number of bytes allowed per injected file path (default: 255) >>>>> quota_injected_file_path_length=255 >>>>> # number of injected files allowed (default: 5) >>>>> quota_injected_files=5 >>>>> # number of instances allowed per project (defailt: 10) >>>>> quota_instances=10 >>>>> # number of key pairs per user (default: 100) >>>>> quota_key_pairs=100 >>>>> # number of metadata items allowed per instance (default: 128) >>>>> quota_metadata_items=128 >>>>> # megabytes of instance ram allowed per project (default: 51200) >>>>> quota_ram=51200 >>>>> >>>>> # virtual CPU to Physical CPU allocation ratio (default: 16.0) >>>>> cpu_allocation_ratio=16.0 >>>>> # virtual ram to physical ram allocation ratio (default: 1.5) >>>>> ram_allocation_ratio=1.5 >>>>> >>>>> mkisofs_cmd=genisoimage >>>>> injected_network_template=$pybasedir/nova/virt/interfaces.template >>>>> flat_injected=false >>>>> >>>>> # The IP address on which the EC2 API will listen. (string value) >>>>> ec2_listen=0.0.0.0 >>>>> # The port on which the EC2 API will listen. (integer value) >>>>> ec2_listen_port=8773 >>>>> >>>>> >>>>> ##### WORKERS ###### >>>>> >>>>> ##### KEYSTONE ##### >>>>> keystone_ec2_url=http://controller:5000/v2.0/ec2tokens >>>>> >>>>> # a list of APIs to enable by default (list value) >>>>> enabled_apis=ec2,osapi_compute,metadata >>>>> >>>>> ##### WORKERS ###### >>>>> >>>>> ##### MONITORS ###### >>>>> # Monitor classes available to the compute which may be >>>>> # specified more than once. (multi valued) >>>>> compute_available_monitors=nova.compute.monitors.all_monitors >>>>> >>>>> # A list of monitors that can be used for getting compute >>>>> # metrics. (list value) >>>>> compute_monitors= >>>>> >>>>> ##### VOLUMES ##### >>>>> # iscsi target user-land tool to use >>>>> iscsi_helper=tgtadm >>>>> volume_api_class=nova.volume.cinder.API >>>>> # Region name of this node (string value) >>>>> os_region_name=RegionOne >>>>> >>>>> # Override the default dnsmasq settings with this file (String value) >>>>> dnsmasq_config_file= >>>>> >>>>> ##### THIRD PARTY ADDITIONS ##### >>>>> >>>>> >>>>> [ssl] >>>>> >>>>> # CA certificate file to use to verify connecting clients >>>>> ca_file= >>>>> >>>>> # Certificate file to use when starting the server securely >>>>> cert_file= >>>>> >>>>> # Private key file to use when starting the server securely >>>>> key_file= >>>>> >>>>> [conductor] >>>>> >>>>> use_local=False >>>>> >>>>> >>>>> [libvirt] >>>>> >>>>> # >>>>> # Options defined in nova.virt.libvirt.driver >>>>> # >>>>> >>>>> # Rescue ami image (string value) >>>>> #rescue_image_id=<None> >>>>> >>>>> # Rescue aki image (string value) >>>>> #rescue_kernel_id=<None> >>>>> >>>>> # Rescue ari image (string value) >>>>> #rescue_ramdisk_id=<None> >>>>> >>>>> # Libvirt domain type (valid options are: kvm, lxc, qemu, uml, >>>>> # xen) (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_type >>>>> virt_type=kvm >>>>> >>>>> # Override the default libvirt URI (which is dependent on >>>>> # virt_type) (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_uri >>>>> #connection_uri= >>>>> >>>>> # Inject the admin password at boot time, without an agent. >>>>> # (boolean value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_password >>>>> inject_password=false >>>>> >>>>> # Inject the ssh public key at boot time (boolean value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_key >>>>> inject_key=true >>>>> >>>>> # The partition to inject to : -2 => disable, -1 => inspect >>>>> # (libguestfs only), 0 => not partitioned, >0 => partition >>>>> # number (integer value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_inject_partition >>>>> inject_partition=-2 >>>>> >>>>> # Sync virtual and real mouse cursors in Windows VMs (boolean >>>>> # value) >>>>> #use_usb_tablet=true >>>>> >>>>> # Migration target URI (any included "%s" is replaced with the >>>>> # migration target hostname) (string value) >>>>> live_migration_uri=qemu+tcp://%s/system >>>>> >>>>> # Migration flags to be set for live migration (string value) >>>>> live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER >>>>> >>>>> # Migration flags to be set for block migration (string value) >>>>> block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, >>>>> VIR_MIGRATE_PEER2PEER, VIR_MIGRATE_NON_SHARED_INC >>>>> >>>>> # Maximum bandwidth to be used during migration, in Mbps >>>>> # (integer value) >>>>> live_migration_bandwidth=0 >>>>> >>>>> # Snapshot image format (valid options are : raw, qcow2, vmdk, >>>>> # vdi). Defaults to same as source image (string value) >>>>> snapshot_image_format=qcow2 >>>>> >>>>> # The libvirt VIF driver to configure the VIFs. (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_vif_driver >>>>> vif_driver=nova.virt.libvirt.vif.LibvirtGenericVIFDriver >>>>> >>>>> # Libvirt handlers for remote volumes. (list value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_volume_drivers >>>>> >>>>> #volume_drivers=iscsi=nova.virt.libvirt.volume.LibvirtISCSIVolumeDriver,iser=nova.virt.libvirt.volume.LibvirtISERVolumeDriver,local=nova.virt.libvirt.volume.LibvirtVolumeDriver,fake=nova.virt.libvirt.volume.LibvirtFakeVolumeDriver,rbd=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,sheepdog=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,nfs=nova.virt.libvirt.volume.LibvirtNFSVolumeDriver,aoe=nova.virt.libvirt.volume.LibvirtAOEVolumeDriver,glusterfs=nova.virt.libvirt.volume.LibvirtGlusterfsVolumeDriver,fibre_channel=nova.virt.libvirt.volume.LibvirtFibreChannelVolumeDriver,scality=nova.virt.libvirt.volume.LibvirtScalityVolumeDriver >>>>> >>>>> # Override the default disk prefix for the devices attached to >>>>> # a server, which is dependent on virt_type. (valid options >>>>> # are: sd, xvd, uvd, vd) (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_disk_prefix >>>>> #disk_prefix=<None> >>>>> >>>>> # Number of seconds to wait for instance to shut down after >>>>> # soft reboot request is made. We fall back to hard reboot if >>>>> # instance does not shutdown within this window. (integer >>>>> # value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_wait_soft_reboot_seconds >>>>> #wait_soft_reboot_seconds=120 >>>>> >>>>> # Set to "host-model" to clone the host CPU feature flags; to >>>>> # "host-passthrough" to use the host CPU model exactly; to >>>>> # "custom" to use a named CPU model; to "none" to not set any >>>>> # CPU model. If virt_type="kvm|qemu", it will default to >>>>> # "host-model", otherwise it will default to "none" (string >>>>> # value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_mode >>>>> >>>>> # Set to a named libvirt CPU model (see names listed in >>>>> # /usr/share/libvirt/cpu_map.xml). Only has effect if >>>>> # cpu_mode="custom" and virt_type="kvm|qemu" (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_model >>>>> #cpu_model=<none> >>>>> >>>>> # Location where libvirt driver will store snapshots before >>>>> # uploading them to image service (string value) >>>>> # Deprecated group/name - [DEFAULT]/libvirt_snapshots_directory >>>>> #snapshots_directory=$instances_path/snapshots >>>>> >>>>> # Location where the Xen hvmloader is kept (string value) >>>>> #xen_hvmloader_path=/usr/lib/xen/boot/hvmloader >>>>> >>>>> # Specific cachemodes to use for different disk types e.g: >>>>> # file=directsync,block=none (list value) >>>>> >>>>> # A path to a device that will be used as source of entropy on >>>>> # the host. Permitted options are: /dev/random or /dev/hwrng >>>>> # (string value) >>>>> >>>>> # >>>>> # Options defined in nova.virt.libvirt.imagecache >>>>> # >>>>> >>>>> # Unused resized base images younger than this will not be removed >>>>> (default: 3600) >>>>> remove_unused_resized_minimum_age_seconds=3600 >>>>> >>>>> # Write a checksum for files in _base to disk (default: false) >>>>> checksum_base_images=false >>>>> >>>>> # >>>>> # Options defined in nova.virt.libvirt.vif >>>>> # >>>>> >>>>> use_virtio_for_bridges=true >>>>> >>>>> # >>>>> # Options defined in nova.virt.libvirt.imagebackend >>>>> # >>>>> >>>>> # VM Images format. Acceptable values are: raw, qcow2, lvm, rbd, >>>>> default. If default is specified, >>>>> # then use_cow_images flag is used instead of this one. >>>>> images_type=default >>>>> >>>>> >>>>> [keystone_authtoken] >>>>> auth_uri = http://controller:5000/v2.0 >>>>> auth_host = controller >>>>> auth_port = 35357 >>>>> auth_protocol = http >>>>> auth_version = v2.0 >>>>> admin_tenant_name = service >>>>> admin_user = nova >>>>> admin_password = openstack-compute >>>>> signing_dir = /var/cache/nova/api >>>>> hash_algorithms = md5 >>>>> insecure = false >>>>> ======================================== >>>>> >>>>> >>>>> Please check it. >>>>> >>>>> >>>>> On Wed, Jan 14, 2015 at 8:23 PM, Jay Pipes <[email protected]> wrote: >>>>> >>>>>> Could you pastebin the output of: >>>>>> >>>>>> keystone catalog >>>>>> >>>>>> and also pastebin your nova.conf for the node running the Nova API >>>>>> service? >>>>>> >>>>>> Thanks! >>>>>> -jay >>>>>> >>>>>> >>>>>> On 01/14/2015 02:25 AM, Geo Varghese wrote: >>>>>> >>>>>>> Hi Team, >>>>>>> >>>>>>> I need a help with cinder volume attachment with an instance. >>>>>>> >>>>>>> I have succesfully created cinder volume and it is in available >>>>>>> state. >>>>>>> Please check attached screenshot. >>>>>>> >>>>>>> Later I tried to attach volume to an instance but attachment failed. >>>>>>> While checking logs in /var/log/nova/nova-api-os-compute.log >>>>>>> >>>>>>> ======== >>>>>>> >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack res = >>>>>>> method(self, ctx, volume_id, *args, **kwargs) >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 206, >>>>>>> in get >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack item = >>>>>>> cinderclient(context).volumes.get(volume_id) >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 91, >>>>>>> in >>>>>>> cinderclient >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack >>>>>>> endpoint_type=endpoint_type) >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack File >>>>>>> "/usr/lib/python2.7/dist-packages/cinderclient/service_catalog.py", >>>>>>> line >>>>>>> 80, in url_for >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack raise >>>>>>> cinderclient.exceptions.EndpointNotFound() >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack >>>>>>> EndpointNotFound >>>>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack >>>>>>> ========================================= >>>>>>> >>>>>>> >>>>>>> I have already created endpints in v1 and v2. >>>>>>> Please check endpints I have created for cinder below >>>>>>> >>>>>>> ============================================================ >>>>>>> ============= >>>>>>> root@controller:/home/geo# keystone endpoint-list | grep 8776 >>>>>>> | 5c7bcc79daa74532ac9ca19949e0d872 | regionOne | >>>>>>> http://controller:8776/v1/%(tenant_id)s | >>>>>>> http://controller:8776/v1/%(tenant_id)s | >>>>>>> http://controller:8776/v1/%(tenant_id)s | >>>>>>> 8ce0898aa7c84fec9b011823d34b55cb | >>>>>>> | 5d71e0a1237c483990b84c36346602b4 | RegionOne | >>>>>>> http://controller:8776/v2/%(tenant_id)s | >>>>>>> http://controller:8776/v2/%(tenant_id)s | >>>>>>> http://controller:8776/v2/%(tenant_id)s | >>>>>>> 251eca5fdb6b4550a9f521c10fa9f2ca | >>>>>>> ============================================================ >>>>>>> =================== >>>>>>> >>>>>>> Anyone please help me. Thanks for the support guys. >>>>>>> >>>>>>> -- >>>>>>> Regards, >>>>>>> Geo Varghese >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> OpenStack-operators mailing list >>>>>>> [email protected] >>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/ >>>>>>> openstack-operators >>>>>>> >>>>>>> >>>>>> _______________________________________________ >>>>>> OpenStack-operators mailing list >>>>>> [email protected] >>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/ >>>>>> openstack-operators >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> -- >>>>> Regards, >>>>> Geo Varghese >>>>> >>>> >>> >>> >>> -- >>> -- >>> Regards, >>> Geo Varghese >>> >> > > > -- > -- > Regards, > Geo Varghese > -- -- Regards, Geo Varghese
_______________________________________________ OpenStack-operators mailing list [email protected] http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
