Ops... forgot the link: http://pastebin.com/LereJgyw The requested infor is in the pastebin. Regards, -- Fernando Fuentes ffuen...@txweather.org http://www.txweather.org On Mon, Jul 18, 2016, at 03:16 PM, Fernando Fuentes wrote: > Nir, > > After some playing around with pvscan I was able to get all of the > need it information. > > Please see: > > > -- > Fernando Fuentes > ffuen...@txweather.org > http://www.txweather.org > > > > On Mon, Jul 18, 2016, at 02:30 PM, Nir Soffer wrote: >> On Mon, Jul 18, 2016 at 6:48 PM, Fernando Fuentes >> <ffuen...@darktcp.net> wrote: >> > Nir, >> > >> > As requested: >> > >> > [root@gamma ~]# lsblk >> > NAME >> > MAJ:MIN RM >> > SIZE RO TYPE MOUNTPOINT >> > sda >> > 8:0 0 >> > 557G 0 disk >> > ├─sda1 >> > 8:1 0 >> > 500M 0 part /boot >> > └─sda2 >> > 8:2 0 >> > 556.5G 0 part >> > ├─vg_gamma-lv_root (dm-0) >> > 253:0 0 >> > 50G 0 lvm / >> > ├─vg_gamma-lv_swap (dm-1) >> > 253:1 0 >> > 4G 0 lvm [SWAP] >> > └─vg_gamma-lv_home (dm-2) >> > 253:2 0 >> > 502.4G 0 lvm /home >> > sr0 >> > 11:0 1 >> > 1024M 0 rom >> > sdb >> > 8:16 0 >> > 2T 0 disk >> > └─36589cfc000000881b9b93c2623780840 (dm-4) >> > 253:4 0 >> > 2T 0 mpath >> > sdc >> > 8:32 0 >> > 2T 0 disk >> > └─36589cfc00000050564002c7e51978316 (dm-3) >> > 253:3 0 >> > 2T 0 mpath >> > ├─3ccb7b67--8067--4315--9656--d68ba10975ba-metadata (dm-5) >> > 253:5 0 >> > 512M 0 lvm >> > ├─3ccb7b67--8067--4315--9656--d68ba10975ba-outbox (dm-6) >> > 253:6 0 >> > 128M 0 lvm >> > ├─3ccb7b67--8067--4315--9656--d68ba10975ba-leases (dm-7) >> > 253:7 0 >> > 2G 0 lvm >> > ├─3ccb7b67--8067--4315--9656--d68ba10975ba-ids (dm-8) >> > 253:8 0 >> > 128M 0 lvm >> > ├─3ccb7b67--8067--4315--9656--d68ba10975ba-inbox (dm-9) >> > 253:9 0 >> > 128M 0 lvm >> > └─3ccb7b67--8067--4315--9656--d68ba10975ba-master (dm-10) >> > 253:10 0 >> > 1G 0 lvm >> > sdd >> > 8:48 0 >> > 4T 0 disk >> > └─36589cfc00000059ccab70662b71c47ef (dm-11) >> > 253:11 0 >> > 4T 0 mpath >> > ├─4861322b--352f--41c6--890a--5cbf1c2c1f01-metadata (dm-12) >> > 253:12 0 >> > 512M 0 lvm >> > ├─4861322b--352f--41c6--890a--5cbf1c2c1f01-ids (dm-13) >> > 253:13 0 >> > 128M 0 lvm >> > ├─4861322b--352f--41c6--890a--5cbf1c2c1f01-leases (dm-14) >> > 253:14 0 >> > 2G 0 lvm >> > ├─4861322b--352f--41c6--890a--5cbf1c2c1f01-outbox (dm-15) >> > 253:15 0 >> > 128M 0 lvm >> > ├─4861322b--352f--41c6--890a--5cbf1c2c1f01-inbox (dm-16) >> > 253:16 0 >> > 128M 0 lvm >> > └─4861322b--352f--41c6--890a--5cbf1c2c1f01-master (dm-17) >> > 253:17 0 >> > 1G 0 lvm >> >> So you have 2 storage domains: >> >> - 3ccb7b67-8067-4315-9656-d68ba10975ba >> - 4861322b-352f-41c6-890a-5cbf1c2c1f01 >> >> But most likely both of them are not active now. >> >> Can you share the output of: >> >> iscsiadm -m session >> >> On a system connected to iscsi storage you will see something like: >> >> # iscsiadm -m session >> tcp: [5] 10.35.0.99:3260,1 iqn.2003-01.org.dumbo.target1 (non-flash) >> >> The special lvs (ids, leases, ...) should be active, and you should >> see also >> regular disks lvs (used for snapshot for vm disks). >> >> Here is an example from machine connected to active iscsi domain: >> >> # lvs >> LV VG >> Attr LSize >> 27c4c795-bca4-4d7b-9b40-cda9098790f5 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 1.00g >> 35be1f52-5b28-4c90-957a-710dbbb8f13f 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 8.00g >> 36d9b41b-4b01-4fc2-8e93-ccf79af0f766 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 128.00m >> 4fda3b44-27a5-4ce4-b8c3-66744aa9937b 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 2.12g >> c2e78f72-d499-44f0-91f5-9930a599dc87 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 128.00m >> d49919b4-30fc-440f-9b21-3367ddfdf396 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 4.00g >> f3b10280-43ed-4772-b122-18c92e098171 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 1.00g >> f409cc48-8248-4239-a4ea-66b0b1084416 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi------- 1.00g >> ids 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-ao---- 128.00m >> inbox 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-a----- 128.00m >> leases 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-a----- 2.00g >> master 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-a----- 1.00g >> metadata 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-a----- 512.00m >> outbox 5f35b5c0-17d7-4475-9125- >> e97f1cdb06f9 -wi-a----- 128.00m >> >> >> >> > [root@gamma ~]# >> > >> > Regards, >> > >> > -- >> > Fernando Fuentes >> > ffuen...@txweather.org >> > http://www.txweather.org >> > >> > On Mon, Jul 18, 2016, at 07:43 AM, Nir Soffer wrote: >> >> Can you share output of lsblk on this host? >> >> >> >> On Mon, Jul 18, 2016 at 3:52 AM, Fernando Fuentes >> >> <ffuen...@darktcp.net> >> >> wrote: >> >> > Nir, >> >> > >> >> > That's odd. gamma is my iscsi host, its in up state and it has >> >> > active >> >> > VM's. >> >> > What am I missing? >> >> > >> >> > Regards, >> >> > >> >> > -- >> >> > Fernando Fuentes >> >> > ffuen...@txweather.org >> >> > http://www.txweather.org >> >> > >> >> > On Sun, Jul 17, 2016, at 07:24 PM, Nir Soffer wrote: >> >> >> On Sun, Jul 17, 2016 at 1:24 AM, Fernando Fuentes >> >> >> <ffuen...@darktcp.net> >> >> >> wrote: >> >> >> > Nir, >> >> >> > >> >> >> > Ok I got the uuid but I am getting the same results as >> >> >> > before. >> >> >> > Nothing comes up. >> >> >> > >> >> >> > [root@gamma ~]# pvscan --cache >> >> >> > [root@gamma ~]# lvs -o vg_name,lv_name,tags | grep >> >> >> > 3b7d9349-9eb1-42f8-9e04-7bbb97c02b98 >> >> >> > [root@gamma ~]# >> >> >> > >> >> >> > without the grep all I get is: >> >> >> > >> >> >> > [root@gamma ~]# lvs -o vg_name,lv_name,tags >> >> >> > VG LV LV Tags >> >> >> > vg_gamma lv_home >> >> >> > vg_gamma lv_root >> >> >> > vg_gamma lv_swap >> >> >> >> >> >> You are not connected to the iscsi storage domain. >> >> >> >> >> >> Please try this from a host in up state in engine. >> >> >> >> >> >> Nir >> >> >> >> >> >> > >> >> >> > On the other hand an fdisk shows a bunch of disks and here is >> >> >> > one >> >> >> > example: >> >> >> > >> >> >> > Disk /dev/mapper/36589cfc00000050564002c7e51978316: 2199.0 >> >> >> > GB, >> >> >> > 2199023255552 bytes >> >> >> > 255 heads, 63 sectors/track, 267349 cylinders >> >> >> > Units = cylinders of 16065 * 512 = 8225280 bytes >> >> >> > Sector size (logical/physical): 512 bytes / 32768 bytes >> >> >> > I/O size (minimum/optimal): 32768 bytes / 1048576 bytes >> >> >> > Disk identifier: 0x00000000 >> >> >> > >> >> >> > >> >> >> > Disk /dev/mapper/36589cfc000000881b9b93c2623780840: 2199.0 >> >> >> > GB, >> >> >> > 2199023255552 bytes >> >> >> > 255 heads, 63 sectors/track, 267349 cylinders >> >> >> > Units = cylinders of 16065 * 512 = 8225280 bytes >> >> >> > Sector size (logical/physical): 512 bytes / 32768 bytes >> >> >> > I/O size (minimum/optimal): 32768 bytes / 1048576 bytes >> >> >> > Disk identifier: 0x00000000 >> >> >> > >> >> >> > >> >> >> > Disk /dev/mapper/3ccb7b67--8067--4315--9656--d68ba10975ba- >> >> >> > metadata: 536 >> >> >> > MB, 536870912 bytes >> >> >> > 255 heads, 63 sectors/track, 65 cylinders >> >> >> > Units = cylinders of 16065 * 512 = 8225280 bytes >> >> >> > Sector size (logical/physical): 512 bytes / 32768 bytes >> >> >> > I/O size (minimum/optimal): 32768 bytes / 1048576 bytes >> >> >> > Disk identifier: 0x00000000 >> >> >> > >> >> >> > Disk /dev/mapper/4861322b--352f--41c6--890a--5cbf1c2c1f01- >> >> >> > master: 1073 >> >> >> > MB, 1073741824 bytes >> >> >> > 255 heads, 63 sectors/track, 130 cylinders >> >> >> > Units = cylinders of 16065 * 512 = 8225280 bytes >> >> >> > Sector size (logical/physical): 512 bytes / 32768 bytes >> >> >> > I/O size (minimum/optimal): 32768 bytes / 1048576 bytes >> >> >> > Disk identifier: 0x00000000 >> >> >> > >> >> >> > Regards, >> >> >> > >> >> >> > -- >> >> >> > Fernando Fuentes >> >> >> > ffuen...@txweather.org >> >> >> > http://www.txweather.org >> >> >> > >> >> >> > On Sat, Jul 16, 2016, at 04:25 PM, Fernando Fuentes wrote: >> >> >> >> Nir, >> >> >> >> >> >> >> >> Ok ill look for it here in a few. >> >> >> >> Thanks for your reply and help! >> >> >> >> >> >> >> >> -- >> >> >> >> Fernando Fuentes >> >> >> >> ffuen...@txweather.org >> >> >> >> http://www.txweather.org >> >> >> >> >> >> >> >> On Sat, Jul 16, 2016, at 04:16 PM, Nir Soffer wrote: >> >> >> >> > On Fri, Jul 15, 2016 at 3:50 PM, Fernando Fuentes >> >> >> >> > <ffuen...@darktcp.net> >> >> >> >> > wrote: >> >> >> >> > > Nir, >> >> >> >> > > >> >> >> >> > > I try to follow your steps but I cant seem to find the >> >> >> >> > > ID of the >> >> >> >> > > template. >> >> >> >> > >> >> >> >> > The image-uuid of the template is displayed in the Disks >> >> >> >> > tab in engine. >> >> >> >> > >> >> >> >> > To find the volume-uuid on block storage, you can do: >> >> >> >> > >> >> >> >> > pvscan --cache >> >> >> >> > lvs -o vg_name,lv_name,tags | grep image-uuid >> >> >> >> > >> >> >> >> > > >> >> >> >> > > Regards, >> >> >> >> > > >> >> >> >> > > -- >> >> >> >> > > Fernando Fuentes >> >> >> >> > > ffuen...@txweather.org >> >> >> >> > > http://www.txweather.org >> >> >> >> > > >> >> >> >> > > On Sun, Jul 10, 2016, at 02:15 PM, Nir Soffer wrote: >> >> >> >> > >> On Thu, Jul 7, 2016 at 7:46 PM, Melissa Mesler >> >> >> >> > >> <meli...@justmelly.com> >> >> >> >> > >> wrote: >> >> >> >> > >> > All, I did a test for Fernando in our ovirt >> >> >> >> > >> > environment. I created a vm >> >> >> >> > >> > called win7melly in the nfs domain. I then migrated >> >> >> >> > >> > it to the iscsi >> >> >> >> > >> > domain. It booted without any issue. So it has to be >> >> >> >> > >> > something with the >> >> >> >> > >> > templates. I have attached the vdsm log for the host >> >> >> >> > >> > the vm resides on. >> >> >> >> > >> >> >> >> >> > >> The log show a working vm, so it does not help much. >> >> >> >> > >> >> >> >> >> > >> I think that the template you copied from the nfs >> >> >> >> > >> domain to the block >> >> >> >> > >> domain is >> >> >> >> > >> corrupted, or the volume metadata are incorrect. >> >> >> >> > >> >> >> >> >> > >> If I understand this correctly, this started when >> >> >> >> > >> Fernando could not copy >> >> >> >> > >> the vm >> >> >> >> > >> disk to the block storage, and I guess the issue was >> >> >> >> > >> that the template >> >> >> >> > >> was missing >> >> >> >> > >> on that storage domain. I assume that he copied the >> >> >> >> > >> template to the >> >> >> >> > >> block storage >> >> >> >> > >> domain by opening the templates tab, selecting the >> >> >> >> > >> template, and choosing >> >> >> >> > >> copy >> >> >> >> > >> from the menu. >> >> >> >> > >> >> >> >> >> > >> Lets compare the template on both nfs and block storage >> >> >> >> > >> domain. >> >> >> >> > >> >> >> >> >> > >> 1. Find the template on the nfs storage domain, using >> >> >> >> > >> the image uuid in >> >> >> >> > >> engine. >> >> >> >> > >> >> >> >> >> > >> It should be at >> >> >> >> > >> >> >> >> >> > >> /rhev/data-center/mnt/server:_path/domain- >> >> >> >> > >> uuid/images/image-uuid/volume-uuid >> >> >> >> > >> >> >> >> >> > >> 2. Please share the output of: >> >> >> >> > >> >> >> >> >> > >> cat /path/to/volume.meta >> >> >> >> > >> qemu-img info /path/to/volume >> >> >> >> > >> qemu-img check /path/to/volume >> >> >> >> > >> >> >> >> >> > >> 4. Find the template on the block storage domain >> >> >> >> > >> >> >> >> >> > >> You should have an lv using the same volume uuid and >> >> >> >> > >> the image-uuid >> >> >> >> > >> should be in the lv tag. >> >> >> >> > >> >> >> >> >> > >> Find it using: >> >> >> >> > >> >> >> >> >> > >> lvs -o vg_name,lv_name,tags | grep volume-uuid >> >> >> >> > >> >> >> >> >> > >> 5. Activate the lv >> >> >> >> > >> >> >> >> >> > >> lvchange -ay vg_name/lv_name >> >> >> >> > >> >> >> >> >> > >> 6. Share the output of >> >> >> >> > >> >> >> >> >> > >> qemu-img info /dev/vg_name/lv_name >> >> >> >> > >> qemu-img check /dev/vg_name/lv_name >> >> >> >> > >> >> >> >> >> > >> 7. Deactivate the lv >> >> >> >> > >> >> >> >> >> > >> lvchange -an vg_name/lv_name >> >> >> >> > >> >> >> >> >> > >> 8. Find the lv metadata >> >> >> >> > >> >> >> >> >> > >> The metadata is stored in /dev/vg_name/metadata. To >> >> >> >> > >> find the correct >> >> >> >> > >> block, >> >> >> >> > >> find the tag named MD_N in the lv tags you found in >> >> >> >> > >> step 4 >> >> >> >> > >> >> >> >> >> > >> The block we need is located at offset N from start of >> >> >> >> > >> volume. >> >> >> >> > >> >> >> >> >> > >> 9. Share the output of: >> >> >> >> > >> >> >> >> >> > >> dd if=/dev/vg_name/metadata bs=512 skip=N count=1 >> >> >> >> > >> iflag=direct >> >> >> >> > >> >> >> >> >> > >> The output of this command should show the image-uuid. >> >> >> >> > >> >> >> >> >> > >> Nir >> >> >> >> > >> >> >> >> >> > >> > >> >> >> >> > >> > - MeLLy >> >> >> >> > >> > >> >> >> >> > >> > On Mon, Jul 4, 2016, at 11:52 PM, Fernando Fuentes >> >> >> >> > >> > wrote: >> >> >> >> > >> >> Nir, >> >> >> >> > >> >> >> >> >> >> > >> >> That's exactly how I did it Nir. >> >> >> >> > >> >> I will test tomorrow with a new Windows VM and >> >> >> >> > >> >> report back. >> >> >> >> > >> >> >> >> >> >> > >> >> Regards, >> >> >> >> > >> >> >> >> >> >> > >> >> -- >> >> >> >> > >> >> Fernando Fuentes >> >> >> >> > >> >> ffuen...@txweather.org >> >> >> >> > >> >> http://www.txweather.org >> >> >> >> > >> >> >> >> >> >> > >> >> On Mon, Jul 4, 2016, at 10:48 AM, Nir Soffer wrote: >> >> >> >> > >> >> > On Mon, Jul 4, 2016 at 6:43 PM, Francesco Romani >> >> >> >> > >> >> > <from...@redhat.com> >> >> >> >> > >> >> > wrote: >> >> >> >> > >> >> > > ----- Original Message ----- >> >> >> >> > >> >> > >> From: "Nir Soffer" <nsof...@redhat.com> >> >> >> >> > >> >> > >> To: "Fernando Fuentes" <ffuen...@darktcp.net> >> >> >> >> > >> >> > >> Cc: "Francesco Romani" <from...@redhat.com>, >> >> >> >> > >> >> > >> "users" <users@ovirt.org> >> >> >> >> > >> >> > >> Sent: Saturday, July 2, 2016 11:18:01 AM >> >> >> >> > >> >> > >> Subject: Re: [ovirt-users] disk not bootable >> >> >> >> > >> >> > >> >> >> >> >> > >> >> > >> On Sat, Jul 2, 2016 at 1:33 AM, Fernando >> >> >> >> > >> >> > >> Fuentes <ffuen...@darktcp.net> >> >> >> >> > >> >> > >> wrote: >> >> >> >> > >> >> > >> > Nir, >> >> >> >> > >> >> > >> > >> >> >> >> > >> >> > >> > Ok I ran another test and this one I moved >> >> >> >> > >> >> > >> > from NFS domain to iSCSI and >> >> >> >> > >> >> > >> > stop working than I moved it back and still >> >> >> >> > >> >> > >> > unable to run... Windows VM >> >> >> >> > >> >> > >> > is saying "no available boot disk" >> >> >> >> > >> >> > >> > VM: Win7-Test >> >> >> >> > >> >> > >> > Host: Zeta >> >> >> >> > >> >> > >> > Info as requested: >> >> >> >> > >> >> > >> > http://pastebin.com/1fSi3auz >> >> >> >> > >> >> > >> >> >> >> >> > >> >> > >> We need a working xml to compare to. >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > [snip expected changes] >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > >> >> >> >> > >> >> > >> <entry name="manufacturer">oVirt</entry> >> >> >> >> > >> >> > >> <entry name="product">oVirt Node</entry> >> >> >> >> > >> >> > >> <entry name="version">6- >> >> >> >> > >> >> > >> 5.el6.centos.11.2</entry> >> >> >> >> > >> >> > >> - <entry name="serial">C938F077-55E2-3E50-A694- >> >> >> >> > >> >> > >> 9FCB7661FD89</entry> >> >> >> >> > >> >> > >> + <entry name="serial">735C7A01-1F16-3CF0-AF8C- >> >> >> >> > >> >> > >> A99823E95AC0</entry> >> >> >> >> > >> >> > >> >> >> >> >> > >> >> > >> Not expected - maybe this is confusing windows? >> >> >> >> > >> >> > >> >> >> >> >> > >> >> > >> Francesco, why vm serial has changed after >> >> >> >> > >> >> > >> moving disks from one storage >> >> >> >> > >> >> > >> domain >> >> >> >> > >> >> > >> to another? >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > We put in serial either >> >> >> >> > >> >> > > 1. the UUID Engine send to us >> >> >> >> > >> >> > > 2. the host UUID as returned by our getHostUUID >> >> >> >> > >> >> > > utility function >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > the latter is unlikely to change, even after >> >> >> >> > >> >> > > this disk move. >> >> >> >> > >> >> > >> >> >> >> > >> >> > Fernando, can you describe exactly how you moved >> >> >> >> > >> >> > the disk? >> >> >> >> > >> >> > >> >> >> >> > >> >> > I assume that you selected the vm in the virtual >> >> >> >> > >> >> > machines tab, then >> >> >> >> > >> >> > selected >> >> >> >> > >> >> > disks from the sub tab, then selected move, and >> >> >> >> > >> >> > selected the target >> >> >> >> > >> >> > storage domain. >> >> >> >> > >> >> > >> >> >> >> > >> >> > Also, can you reproduce this with a new vm? >> >> >> >> > >> >> > (create vm with disk nfs, >> >> >> >> > >> >> > stop vm, >> >> >> >> > >> >> > move disk to iscsi, start vm). >> >> >> >> > >> >> > >> >> >> >> > >> >> > > So the first suspect in line is Engine >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > Arik, do you know if Engine is indeed supposed >> >> >> >> > >> >> > > to change the UUID in this flow? >> >> >> >> > >> >> > > That seems very surprising. >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > Thanks and bests, >> >> >> >> > >> >> > > >> >> >> >> > >> >> > > -- >> >> >> >> > >> >> > > Francesco Romani >> >> >> >> > >> >> > > RedHat Engineering Virtualization R & D >> >> >> >> > >> >> > > Phone: 8261328 >> >> >> >> > >> >> > > IRC: fromani >> >> >> >> > >> >> _______________________________________________ >> >> >> >> > >> >> Users mailing list >> >> >> >> > >> >> Users@ovirt.org >> >> >> >> > >> >> http://lists.ovirt.org/mailman/listinfo/users >> >> >> >> > >> > >> >> >> >> > >> > _______________________________________________ >> >> >> >> > >> > Users mailing list >> >> >> >> > >> > Users@ovirt.org >> >> >> >> > >> > http://lists.ovirt.org/mailman/listinfo/users >> >> >> >> > >> > >> >> >> >> _______________________________________________ >> >> >> >> Users mailing list >> >> >> >> Users@ovirt.org >> >> >> >> http://lists.ovirt.org/mailman/listinfo/users >> >> >> > _______________________________________________ >> >> >> > Users mailing list >> >> >> > Users@ovirt.org >> >> >> > http://lists.ovirt.org/mailman/listinfo/users >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users