Hi, Sergey:

During the time period when I had problem cloning template,  there are only a 
few unique entries in vmkernel.log, and they were repeated hundreds/thousands 
of times by all the cpu cores:

2019-06-02T16:47:00.633Z cpu9:8491061)FSS: 6751: Failed to open file 
'hpilo-d0ccb15'; Requested flags 0x5, world: 8491061 [ams-ahs], (Existing flags 
0x5, world: 8491029 [ams-main]): Busy
2019-06-02T16:47:49.320Z cpu1:66415)nhpsa: hpsa_vmkScsiCmdDone:6384: Sense 
data: error code: 0x70, key: 0x5, info:00 00 00 00 , cmdInfo:00 00 00 00 , 
CmdSN: 0xd5c, worldId: 0x818e8e, Cmd: 0x85, ASC: 0x20, ASCQ: 0x0
2019-06-02T16:47:49.320Z cpu1:66415)ScsiDeviceIO: 2948: Cmd(0x43954115be40) 
0x85, CmdSN 0xd5c from world 8490638 to dev 
"naa.600508b1001c6d77d7dd6a0cc0953df1" failed H:0x0 D:0x2 P:0x0 Valid sense 
data: 0x5 0x20 0x0.

The device " naa.600508b1001c6d77d7dd6a0cc0953df1" is the local disk on this 
host.

Yiping


On 6/5/19, 11:15 AM, "Sergey Levitskiy" <serg...@hotmail.com> wrote:

    This must be specific to that environment.  For a full clone mode ACS 
simply calls cloneVMTask of vSphere API so basically until cloning of that 
template succeeds when attmepted in vSphere client  it would keep failing in 
ACS. Can you post vmkernel.log from your ESX host esx-0001-a-001?
    
    
    On 6/5/19, 8:47 AM, "Yiping Zhang" <yipzh...@adobe.com.INVALID> wrote:
    
        Well,  I can always reproduce it in this particular vSphere set up,  
but in a different ACS+vSphere environment,  I don't see this problem.
        
        Yiping
        
        On 6/5/19, 1:00 AM, "Andrija Panic" <andrija.pa...@gmail.com> wrote:
        
            Yiping,
            
            if you are sure you can reproduce the issue, it would be good to 
raise a
            GitHub issue and provide as much detail as possible.
            
            Andrija
            
            On Wed, 5 Jun 2019 at 05:29, Yiping Zhang 
<yipzh...@adobe.com.invalid>
            wrote:
            
            > Hi, Sergey:
            >
            > Thanks for the tip. After setting vmware.create.full.clone=false, 
 I was
            > able to create and start system VM instances.    However,  I feel 
that the
            > underlying problem still exists, and I am just working around it 
instead of
            > fixing it,  because in my lab CloudStack instance with the same 
version of
            > ACS and vSphere,  I still have vmware.create.full.clone=true and 
all is
            > working as expected.
            >
            > I did some reading on VMware docs regarding full clone vs. linked 
clone.
            > It seems that the best practice is to use full clone for 
production,
            > especially if there are high rates of changes to the disks.  So
            > eventually,  I need to understand and fix the root cause for this 
issue.
            > At least for now,  I am over this hurdle and I can move on.
            >
            > Thanks again,
            >
            > Yiping
            >
            > On 6/4/19, 11:13 AM, "Sergey Levitskiy" <serg...@hotmail.com> 
wrote:
            >
            >     Everything looks good and consistent including all references 
in VMDK
            > and its snapshot. I would try these 2 routes:
            >     1. Figure out what vSphere error actually means from vmkernel 
log of
            > ESX when ACS tries to clone the template. If the same error 
happens while
            > doing it outside of ACS then a support case with VMware can be an 
option
            >     2. Try using link clones. This can be done by this global 
setting and
            > restarting management server
            >     vmware.create.full.clone                    false
            >
            >
            >     On 6/4/19, 9:57 AM, "Yiping Zhang" 
<yipzh...@adobe.com.INVALID> wrote:
            >
            >         Hi, Sergey:
            >
            >         Thanks for the help. By now, I have dropped and recreated 
DB,
            > re-deployed this zone multiple times, blow away primary and 
secondary
            > storage (including all contents on them) , or just delete 
template itself
            > from primary storage, multiple times.  Every time I ended up with 
the same
            > error at the same place.
            >
            >         The full management server log,  from the point I seeded 
the
            > systemvmtemplate for vmware, to deploying a new advanced zone and 
enable
            > the zone to let CS to create system VM's and finally disable the 
zone to
            > stop infinite loop of trying to recreate failed system VM's,  are 
posted
            > at pastebin:
            >
            >
            > 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2Fc05wiQ3R&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C44530fc614da4d42aeb208d6e9e1bf07%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636953553116209842&amp;sdata=oQLChzVf11KaM4bsFV9rraYkT%2F96AIhfR3SNQkpOBhs%3D&amp;reserved=0
            >
            >         Here are the content of relevant files for the template 
on primary
            > storage:
            >
            >         1) /vmfsvolumes:
            >
            >         ls -l /vmfs/volumes/
            >         total 2052
            >         drwxr-xr-x    1 root     root             8 Jan  1  1970
            > 414f6a73-87cd6dac-9585-133ddd409762
            >         lrwxr-xr-x    1 root     root            17 Jun  4 16:37
            > 42054b8459633172be231d72a52d59d4 -> afc5e946-03bfe3c2          
<== this is
            > the NFS datastore for primary storage
            >         drwxr-xr-x    1 root     root             8 Jan  1  1970
            > 5cd4b46b-fa4fcff0-d2a1-00215a9b31c0
            >         drwxr-xr-t    1 root     root          1400 Jun  3 22:50
            > 5cd4b471-c2318b91-8fb2-00215a9b31c0
            >         drwxr-xr-x    1 root     root             8 Jan  1  1970
            > 5cd4b471-da49a95b-bdb6-00215a9b31c0
            >         drwxr-xr-x    4 root     root          4096 Jun  3 23:38
            > afc5e946-03bfe3c2
            >         drwxr-xr-x    1 root     root             8 Jan  1  1970
            > b70c377c-54a9d28a-6a7b-3f462a475f73
            >
            >         2) content in template dir on primary storage:
            >
            >         ls -l
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/
            >         total 1154596
            >         -rw-------    1 root     root          8192 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a-000001-delta.vmdk
            >         -rw-------    1 root     root           366 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
            >         -rw-r--r--    1 root     root           268 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog
            >         -rw-------    1 root     root          9711 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn
            >         -rw-------    1 root     root     2097152000 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a-flat.vmdk
            >         -rw-------    1 root     root           518 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a.vmdk
            >         -rw-r--r--    1 root     root           471 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a.vmsd
            >         -rwxr-xr-x    1 root     root          1402 Jun  3 23:38
            > 533b6fcf3fa6301aadcc2b168f3f999a.vmtx
            >
            >         3) *.vmdk file content:
            >
            >         cat
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmdk
            >         # Disk DescriptorFile
            >         version=1
            >         encoding="UTF-8"
            >         CID=ecb01275
            >         parentCID=ffffffff
            >         isNativeSnapshot="no"
            >         createType="vmfs"
            >
            >         # Extent description
            >         RW 4096000 VMFS 
"533b6fcf3fa6301aadcc2b168f3f999a-flat.vmdk"
            >
            >         # The Disk Data Base
            >         #DDB
            >
            >         ddb.adapterType = "lsilogic"
            >         ddb.geometry.cylinders = "4063"
            >         ddb.geometry.heads = "16"
            >         ddb.geometry.sectors = "63"
            >         ddb.longContentID = "1c60ba48999abde959998f05ecb01275"
            >         ddb.thinProvisioned = "1"
            >         ddb.uuid = "60 00 C2 9b 52 6d 98 c4-1f 44 51 ce 1e 70 a9 
70"
            >         ddb.virtualHWVersion = "13"
            >
            >         4) *-0001.vmdk content:
            >
            >         cat
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
            >
            >         # Disk DescriptorFile
            >         version=1
            >         encoding="UTF-8"
            >         CID=ecb01275
            >         parentCID=ecb01275
            >         isNativeSnapshot="no"
            >         createType="vmfsSparse"
            >         parentFileNameHint="533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
            >         # Extent description
            >         RW 4096000 VMFSSPARSE
            > "533b6fcf3fa6301aadcc2b168f3f999a-000001-delta.vmdk"
            >
            >         # The Disk Data Base
            >         #DDB
            >
            >         ddb.longContentID = "1c60ba48999abde959998f05ecb01275"
            >
            >
            >         5) *.vmtx content:
            >
            >         cat
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmtx
            >
            >         .encoding = "UTF-8"
            >         config.version = "8"
            >         virtualHW.version = "8"
            >         nvram = "533b6fcf3fa6301aadcc2b168f3f999a.nvram"
            >         pciBridge0.present = "TRUE"
            >         svga.present = "TRUE"
            >         pciBridge4.present = "TRUE"
            >         pciBridge4.virtualDev = "pcieRootPort"
            >         pciBridge4.functions = "8"
            >         pciBridge5.present = "TRUE"
            >         pciBridge5.virtualDev = "pcieRootPort"
            >         pciBridge5.functions = "8"
            >         pciBridge6.present = "TRUE"
            >         pciBridge6.virtualDev = "pcieRootPort"
            >         pciBridge6.functions = "8"
            >         pciBridge7.present = "TRUE"
            >         pciBridge7.virtualDev = "pcieRootPort"
            >         pciBridge7.functions = "8"
            >         vmci0.present = "TRUE"
            >         hpet0.present = "TRUE"
            >         floppy0.present = "FALSE"
            >         memSize = "256"
            >         scsi0.virtualDev = "lsilogic"
            >         scsi0.present = "TRUE"
            >         ide0:0.startConnected = "FALSE"
            >         ide0:0.deviceType = "atapi-cdrom"
            >         ide0:0.fileName = "CD/DVD drive 0"
            >         ide0:0.present = "TRUE"
            >         scsi0:0.deviceType = "scsi-hardDisk"
            >         scsi0:0.fileName = 
"533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk"
            >         scsi0:0.present = "TRUE"
            >         displayName = "533b6fcf3fa6301aadcc2b168f3f999a"
            >         annotation = "systemvmtemplate-4.11.2.0-vmware"
            >         guestOS = "otherlinux-64"
            >         toolScripts.afterPowerOn = "TRUE"
            >         toolScripts.afterResume = "TRUE"
            >         toolScripts.beforeSuspend = "TRUE"
            >         toolScripts.beforePowerOff = "TRUE"
            >         uuid.bios = "42 02 f1 40 33 e8 de e5-1a c5 93 2a c9 12 47 
61"
            >         vc.uuid = "50 02 5b d9 e9 c9 77 86-28 3e 84 00 22 2b eb 
d3"
            >         firmware = "bios"
            >         migrate.hostLog = 
"533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog"
            >
            >
            >         6) *.vmsd file content:
            >
            >         cat
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmsd
            >         .encoding = "UTF-8"
            >         snapshot.lastUID = "1"
            >         snapshot.current = "1"
            >         snapshot0.uid = "1"
            >         snapshot0.filename =
            > "533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn"
            >         snapshot0.displayName = "cloud.template.base"
            >         snapshot0.description = "Base snapshot"
            >         snapshot0.createTimeHigh = "363123"
            >         snapshot0.createTimeLow = "-679076964"
            >         snapshot0.numDisks = "1"
            >         snapshot0.disk0.fileName = 
"533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
            >         snapshot0.disk0.node = "scsi0:0"
            >         snapshot.numSnapshots = "1"
            >
            >         7) *-Snapshot1.vmsn content:
            >
            >         cat
            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn
            >
            >         ҾSnapshot\?%?cfgFilet%t%.encoding = "UTF-8"
            >         config.version = "8"
            >         virtualHW.version = "8"
            >         nvram = "533b6fcf3fa6301aadcc2b168f3f999a.nvram"
            >         pciBridge0.present = "TRUE"
            >         svga.present = "TRUE"
            >         pciBridge4.present = "TRUE"
            >         pciBridge4.virtualDev = "pcieRootPort"
            >         pciBridge4.functions = "8"
            >         pciBridge5.present = "TRUE"
            >         pciBridge5.virtualDev = "pcieRootPort"
            >         pciBridge5.functions = "8"
            >         pciBridge6.present = "TRUE"
            >         pciBridge6.virtualDev = "pcieRootPort"
            >         pciBridge6.functions = "8"
            >         pciBridge7.present = "TRUE"
            >         pciBridge7.virtualDev = "pcieRootPort"
            >         pciBridge7.functions = "8"
            >         vmci0.present = "TRUE"
            >         hpet0.present = "TRUE"
            >         floppy0.present = "FALSE"
            >         memSize = "256"
            >         scsi0.virtualDev = "lsilogic"
            >         scsi0.present = "TRUE"
            >         ide0:0.startConnected = "FALSE"
            >         ide0:0.deviceType = "atapi-cdrom"
            >         ide0:0.fileName = "CD/DVD drive 0"
            >         ide0:0.present = "TRUE"
            >         scsi0:0.deviceType = "scsi-hardDisk"
            >         scsi0:0.fileName = "533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
            >         scsi0:0.present = "TRUE"
            >         displayName = "533b6fcf3fa6301aadcc2b168f3f999a"
            >         annotation = "systemvmtemplate-4.11.2.0-vmware"
            >         guestOS = "otherlinux-64"
            >         toolScripts.afterPowerOn = "TRUE"
            >         toolScripts.afterResume = "TRUE"
            >         toolScripts.beforeSuspend = "TRUE"
            >         toolScripts.beforePowerOff = "TRUE"
            >         uuid.bios = "42 02 f1 40 33 e8 de e5-1a c5 93 2a c9 12 47 
61"
            >         vc.uuid = "50 02 5b d9 e9 c9 77 86-28 3e 84 00 22 2b eb 
d3"
            >         firmware = "bios"
            >         migrate.hostLog = 
"533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog"
            >
            >
            >         ------------
            >
            >         That's all the data on the template VMDK.
            >
            >         Much appreciate your time!
            >
            >         Yiping
            >
            >
            >
            >         On 6/4/19, 9:29 AM, "Sergey Levitskiy" 
<serg...@hotmail.com>
            > wrote:
            >
            >             Have you tried deleting template from PS and let ACS 
to recopy
            > it again? If the issue is reproducible we can try to look what is 
wrong
            > with VMDK. Please post content of 
533b6fcf3fa6301aadcc2b168f3f999a.vmdk ,
            > 533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk and
            > 533b6fcf3fa6301aadcc2b168f3f999a.vmx (their equitant after ACS 
finishes
            > copying template). Also from one of your ESX hosts output of this
            >             ls -al /vmfs/volumes
            >             ls -al 
/vmfs/volumes/*/533b6fcf3fa6301aadcc2b168f3f999a (their
            > equitant after ACS finishes copying template)
            >
            >              Can you also post management server log starting 
from the
            > point you unregister and delete template from the vCenter.
            >
            >             On 6/4/19, 8:37 AM, "Yiping Zhang" 
<yipzh...@adobe.com.INVALID>
            > wrote:
            >
            >                 I have manually imported the OVA to vCenter and
            > successfully cloned a VM instance with it, on the same NFS 
datastore.
            >
            >
            >                 On 6/4/19, 8:25 AM, "Sergey Levitskiy" <
            > serg...@hotmail.com> wrote:
            >
            >                     I would suspect the template is corrupted on 
the
            > secondary storage. You can try disabling/enabling link clone 
feature and
            > see if it works the other way.
            >                     vmware.create.full.clone                    
false
            >
            >                     Also systemVM template might have been 
generated on a
            > newer version of vSphere and not compatible with ESXi 6.5. What 
you can do
            > to validate this is to manually deploy OVA that is in Secondary 
storage and
            > try to spin up VM from it directly in vCenter.
            >
            >
            >
            >                     On 6/3/19, 5:41 PM, "Yiping Zhang"
            > <yipzh...@adobe.com.INVALID> wrote:
            >
            >                         Hi, list:
            >
            >                         I am struggling with deploying a new 
advanced zone
            > using ACS 4.11.2.0 + vSphere 6.5 + NetApp volumes for primary and 
secondary
            > storage devices. The initial setup of CS management server, 
seeding of
            > systemVM template, and advanced zone deployment all went smoothly.
            >
            >                         Once I enabled the zone in web UI and the 
systemVM
            > template gets copied/staged on to primary storage device. But 
subsequent VM
            > creations from this template would fail with errors:
            >
            >
            >                         2019-06-03 18:38:15,764 INFO  
[c.c.h.v.m.HostMO]
            > (DirectAgent-7:ctx-d01169cb esx-0001-a-001.example.org, 
job-3/job-29,
            > cmd: CopyCommand) VM 533b6fcf3fa6301aadcc2b168f3f999a not found 
in host
            > cache
            >
            >                         2019-06-03 18:38:17,017 INFO
            > [c.c.h.v.r.VmwareResource] (DirectAgent-4:ctx-08b54fbd
            > esx-0001-a-001.example.org, job-3/job-29, cmd: CopyCommand)
            > VmwareStorageProcessor and VmwareStorageSubsystemCommandHandler
            > successfully reconfigured
            >
            >                         2019-06-03 18:38:17,128 INFO
            > [c.c.s.r.VmwareStorageProcessor] (DirectAgent-4:ctx-08b54fbd
            > esx-0001-a-001.example.org, job-3/job-29, cmd: CopyCommand) 
creating full
            > clone from template
            >
            >                         2019-06-03 18:38:17,657 INFO
            > [c.c.h.v.u.VmwareHelper] (DirectAgent-4:ctx-08b54fbd
            > esx-0001-a-001.example.org, job-3/job-29, cmd: CopyCommand)
            > [ignored]failed toi get message for exception: Error caused by 
file
            > 
/vmfs/volumes/afc5e946-03bfe3c2/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
            >
            >                         2019-06-03 18:38:17,658 ERROR
            > [c.c.s.r.VmwareStorageProcessor] (DirectAgent-4:ctx-08b54fbd
            > esx-0001-a-001.example.org, job-3/job-29, cmd: CopyCommand) clone 
volume
            > from base image failed due to Exception: 
java.lang.RuntimeException
            >
            >                         Message: Error caused by file
            > 
/vmfs/volumes/afc5e946-03bfe3c2/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
            >
            >
            >
            >                         If I try to create “new VM from template”
            > (533b6fcf3fa6301aadcc2b168f3f999a) on vCenter UI manually,  I 
will receive
            > exactly the same error message. The name of the VMDK file in the 
error
            > message is a snapshot of the base disk image, but it is not part 
of the
            > original template OVA on the secondary storage.  So, in the 
process of
            > copying the template from secondary to primary storage, a 
snapshot got
            > created and the disk became corrupted/unusable.
            >
            >                         Much later in the log file,  there is 
another
            > error message “failed to fetch any free public IP address” (for 
ssvm, I
            > think).  I don’t know if these two errors are related or if one 
is the root
            > cause for the other error.
            >
            >                         The full management server log is 
uploaded as
            > 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2Fc05wiQ3R&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C44530fc614da4d42aeb208d6e9e1bf07%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636953553116209842&amp;sdata=oQLChzVf11KaM4bsFV9rraYkT%2F96AIhfR3SNQkpOBhs%3D&amp;reserved=0
            >
            >                         Any help or insight on what went wrong 
here are
            > much appreciated.
            >
            >                         Thanks
            >
            >                         Yiping
            >
            >
            >
            >
            >
            >
            >
            >
            >
            >
            >
            >
            >
            
            -- 
            
            Andrija Panić
            
        
        
    
    

Reply via email to