The nfs volume definitely allows root mount and have RW permissions, as we 
already see the volume mounted and template staged on primary storage. The 
volume is mounted as NFS3 datastore in vSphere.

Volume snapshot is enabled,  I can ask to have snapshot disabled to see if it 
makes any differentces.   I need to find out more about NFS version and qtree 
mode from our storage admin.  

One thing I noticed is that when cloudstack templates are staged on to primary 
storage, a snapshot was created, which does not exist In the original OVA or on 
secondary storage.  I suppose this is the expected behavior?

Yiping

On 6/6/19, 6:59 AM, "Sergey Levitskiy" <serg...@hotmail.com> wrote:

    This option is 'vol options name_of_volume nosnapdir on' however if I 
recall it right is supposed to work even with .snapshot directory visible
    Can you find out all vol options on your netapp volume? I would be most 
concerned about:
    - NFS version - NFS v4 should be disabled
    - security qtree mode to be set to UNIX
    - allow root mount
    
    I am also wondering if ACS is able to create ROOT-XX folder so you might 
want to watch the content of the DS when ACS tries the operations.
     
    
    On 6/5/19, 11:43 PM, "Paul Angus" <paul.an...@shapeblue.com> wrote:
    
        Hi Yiping,
        
        do you have snapshots enabled on the NetApp filer?  (it used to be seen 
as a  ".snapshot"  subdirectory in each directory)
        
        If so try disabling snapshots - there used to be a bug where the 
.snapshot directory would confuse CloudStack.
        
        paul.an...@shapeblue.com 
        
https://nam04.safelinks.protection.outlook.com/?url=www.shapeblue.com&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C557bf647ff13413c66b708d6ea87220d%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636954263448727822&amp;sdata=NhoxwF0x4%2F8yn%2B8ck%2BCI8RUKEEDGnI73QfDDQeSmZUc%3D&amp;reserved=0
        Amadeus House, Floral Street, London  WC2E 9DPUK
        @shapeblue
          
         
        
        
        -----Original Message-----
        From: Yiping Zhang <yipzh...@adobe.com.INVALID> 
        Sent: 05 June 2019 23:38
        To: users@cloudstack.apache.org
        Subject: Re: Can't start systemVM in a new advanced zone deployment
        
        Hi, Sergey:
        
        I found more logs in vpxa.log ( the esxi hosts are using UTC time zone, 
 so I was looking at wrong time periods earlier).  I have uploaded more logs 
into pastebin.
        
        From these log entries,  it appears that when copying template to VM,  
it tried to open destination VMDK file and got error file not found.  
        
        In case that the CloudStack attempted to create a systemVM,  the 
destination VMDK file path it is looking for is 
"<datastore>/<disk-name>/<disk-name>.vmdk",  see uploaded log at 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2FaFysZkTy&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C557bf647ff13413c66b708d6ea87220d%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636954263448727822&amp;sdata=YyB9VdghCgiBuUmDZ8gIc0jPlM8miPzemX2UEAZ3sFA%3D&amp;reserved=0
        
        In case when I manually created new VM from a (different) template in 
vCenter UI,   the destination VMDK file path it is looking for is 
"<datastore>/<VM-NAME>/<VM-NAME>.vmdk", see uploaded log at 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2FyHcsD8xB&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C557bf647ff13413c66b708d6ea87220d%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636954263448732817&amp;sdata=N%2BZHteGo3LDU0pvhBtzv7wcocAv35gRE9b9yKVQa6%2FQ%3D&amp;reserved=0
        
        So, I am confused as to how the path for destination VMDK was 
determined and by CloudStack or VMware, how did I end up with this?
        
        Yiping
        
        
        On 6/5/19, 12:32 PM, "Sergey Levitskiy" <serg...@hotmail.com> wrote:
        
            Some operations log get transferred to vCenter log vpxd.log. It is 
not straightforward to trace I but Vmware will be able to help should you open 
case with them. 
            
            
            On 6/5/19, 11:39 AM, "Yiping Zhang" <yipzh...@adobe.com.INVALID> 
wrote:
            
                Hi, Sergey:
                
                During the time period when I had problem cloning template,  
there are only a few unique entries in vmkernel.log, and they were repeated 
hundreds/thousands of times by all the cpu cores:
                
                2019-06-02T16:47:00.633Z cpu9:8491061)FSS: 6751: Failed to open 
file 'hpilo-d0ccb15'; Requested flags 0x5, world: 8491061 [ams-ahs], (Existing 
flags 0x5, world: 8491029 [ams-main]): Busy
                2019-06-02T16:47:49.320Z cpu1:66415)nhpsa: 
hpsa_vmkScsiCmdDone:6384: Sense data: error code: 0x70, key: 0x5, info:00 00 00 
00 , cmdInfo:00 00 00 00 , CmdSN: 0xd5c, worldId: 0x818e8e, Cmd: 0x85, ASC: 
0x20, ASCQ: 0x0
                2019-06-02T16:47:49.320Z cpu1:66415)ScsiDeviceIO: 2948: 
Cmd(0x43954115be40) 0x85, CmdSN 0xd5c from world 8490638 to dev 
"naa.600508b1001c6d77d7dd6a0cc0953df1" failed H:0x0 D:0x2 P:0x0 Valid sense 
data: 0x5 0x20 0x0.
                
                The device " naa.600508b1001c6d77d7dd6a0cc0953df1" is the local 
disk on this host.
                
                Yiping
                
                
                On 6/5/19, 11:15 AM, "Sergey Levitskiy" <serg...@hotmail.com> 
wrote:
                
                    This must be specific to that environment.  For a full 
clone mode ACS simply calls cloneVMTask of vSphere API so basically until 
cloning of that template succeeds when attmepted in vSphere client  it would 
keep failing in ACS. Can you post vmkernel.log from your ESX host 
esx-0001-a-001?
                    
                    
                    On 6/5/19, 8:47 AM, "Yiping Zhang" 
<yipzh...@adobe.com.INVALID> wrote:
                    
                        Well,  I can always reproduce it in this particular 
vSphere set up,  but in a different ACS+vSphere environment,  I don't see this 
problem.
                        
                        Yiping
                        
                        On 6/5/19, 1:00 AM, "Andrija Panic" 
<andrija.pa...@gmail.com> wrote:
                        
                            Yiping,
                            
                            if you are sure you can reproduce the issue, it 
would be good to raise a
                            GitHub issue and provide as much detail as possible.
                            
                            Andrija
                            
                            On Wed, 5 Jun 2019 at 05:29, Yiping Zhang 
<yipzh...@adobe.com.invalid>
                            wrote:
                            
                            > Hi, Sergey:
                            >
                            > Thanks for the tip. After setting 
vmware.create.full.clone=false,  I was
                            > able to create and start system VM instances.    
However,  I feel that the
                            > underlying problem still exists, and I am just 
working around it instead of
                            > fixing it,  because in my lab CloudStack instance 
with the same version of
                            > ACS and vSphere,  I still have 
vmware.create.full.clone=true and all is
                            > working as expected.
                            >
                            > I did some reading on VMware docs regarding full 
clone vs. linked clone.
                            > It seems that the best practice is to use full 
clone for production,
                            > especially if there are high rates of changes to 
the disks.  So
                            > eventually,  I need to understand and fix the 
root cause for this issue.
                            > At least for now,  I am over this hurdle and I 
can move on.
                            >
                            > Thanks again,
                            >
                            > Yiping
                            >
                            > On 6/4/19, 11:13 AM, "Sergey Levitskiy" 
<serg...@hotmail.com> wrote:
                            >
                            >     Everything looks good and consistent 
including all references in VMDK
                            > and its snapshot. I would try these 2 routes:
                            >     1. Figure out what vSphere error actually 
means from vmkernel log of
                            > ESX when ACS tries to clone the template. If the 
same error happens while
                            > doing it outside of ACS then a support case with 
VMware can be an option
                            >     2. Try using link clones. This can be done by 
this global setting and
                            > restarting management server
                            >     vmware.create.full.clone                    
false
                            >
                            >
                            >     On 6/4/19, 9:57 AM, "Yiping Zhang" 
<yipzh...@adobe.com.INVALID> wrote:
                            >
                            >         Hi, Sergey:
                            >
                            >         Thanks for the help. By now, I have 
dropped and recreated DB,
                            > re-deployed this zone multiple times, blow away 
primary and secondary
                            > storage (including all contents on them) , or 
just delete template itself
                            > from primary storage, multiple times.  Every time 
I ended up with the same
                            > error at the same place.
                            >
                            >         The full management server log,  from the 
point I seeded the
                            > systemvmtemplate for vmware, to deploying a new 
advanced zone and enable
                            > the zone to let CS to create system VM's and 
finally disable the zone to
                            > stop infinite loop of trying to recreate failed 
system VM's,  are posted
                            > at pastebin:
                            >
                            >
                            > 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2Fc05wiQ3R&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C557bf647ff13413c66b708d6ea87220d%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636954263448732817&amp;sdata=NY2nAe8v%2BP7ANlpjD4xcmOSc7IoBpFizoX3eCuclUHo%3D&amp;reserved=0
                            >
                            >         Here are the content of relevant files 
for the template on primary
                            > storage:
                            >
                            >         1) /vmfsvolumes:
                            >
                            >         ls -l /vmfs/volumes/
                            >         total 2052
                            >         drwxr-xr-x    1 root     root             
8 Jan  1  1970
                            > 414f6a73-87cd6dac-9585-133ddd409762
                            >         lrwxr-xr-x    1 root     root            
17 Jun  4 16:37
                            > 42054b8459633172be231d72a52d59d4 -> 
afc5e946-03bfe3c2          <== this is
                            > the NFS datastore for primary storage
                            >         drwxr-xr-x    1 root     root             
8 Jan  1  1970
                            > 5cd4b46b-fa4fcff0-d2a1-00215a9b31c0
                            >         drwxr-xr-t    1 root     root          
1400 Jun  3 22:50
                            > 5cd4b471-c2318b91-8fb2-00215a9b31c0
                            >         drwxr-xr-x    1 root     root             
8 Jan  1  1970
                            > 5cd4b471-da49a95b-bdb6-00215a9b31c0
                            >         drwxr-xr-x    4 root     root          
4096 Jun  3 23:38
                            > afc5e946-03bfe3c2
                            >         drwxr-xr-x    1 root     root             
8 Jan  1  1970
                            > b70c377c-54a9d28a-6a7b-3f462a475f73
                            >
                            >         2) content in template dir on primary 
storage:
                            >
                            >         ls -l
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/
                            >         total 1154596
                            >         -rw-------    1 root     root          
8192 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a-000001-delta.vmdk
                            >         -rw-------    1 root     root           
366 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
                            >         -rw-r--r--    1 root     root           
268 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog
                            >         -rw-------    1 root     root          
9711 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn
                            >         -rw-------    1 root     root     
2097152000 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a-flat.vmdk
                            >         -rw-------    1 root     root           
518 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a.vmdk
                            >         -rw-r--r--    1 root     root           
471 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a.vmsd
                            >         -rwxr-xr-x    1 root     root          
1402 Jun  3 23:38
                            > 533b6fcf3fa6301aadcc2b168f3f999a.vmtx
                            >
                            >         3) *.vmdk file content:
                            >
                            >         cat
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmdk
                            >         # Disk DescriptorFile
                            >         version=1
                            >         encoding="UTF-8"
                            >         CID=ecb01275
                            >         parentCID=ffffffff
                            >         isNativeSnapshot="no"
                            >         createType="vmfs"
                            >
                            >         # Extent description
                            >         RW 4096000 VMFS 
"533b6fcf3fa6301aadcc2b168f3f999a-flat.vmdk"
                            >
                            >         # The Disk Data Base
                            >         #DDB
                            >
                            >         ddb.adapterType = "lsilogic"
                            >         ddb.geometry.cylinders = "4063"
                            >         ddb.geometry.heads = "16"
                            >         ddb.geometry.sectors = "63"
                            >         ddb.longContentID = 
"1c60ba48999abde959998f05ecb01275"
                            >         ddb.thinProvisioned = "1"
                            >         ddb.uuid = "60 00 C2 9b 52 6d 98 c4-1f 44 
51 ce 1e 70 a9 70"
                            >         ddb.virtualHWVersion = "13"
                            >
                            >         4) *-0001.vmdk content:
                            >
                            >         cat
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
                            >
                            >         # Disk DescriptorFile
                            >         version=1
                            >         encoding="UTF-8"
                            >         CID=ecb01275
                            >         parentCID=ecb01275
                            >         isNativeSnapshot="no"
                            >         createType="vmfsSparse"
                            >         
parentFileNameHint="533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
                            >         # Extent description
                            >         RW 4096000 VMFSSPARSE
                            > 
"533b6fcf3fa6301aadcc2b168f3f999a-000001-delta.vmdk"
                            >
                            >         # The Disk Data Base
                            >         #DDB
                            >
                            >         ddb.longContentID = 
"1c60ba48999abde959998f05ecb01275"
                            >
                            >
                            >         5) *.vmtx content:
                            >
                            >         cat
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmtx
                            >
                            >         .encoding = "UTF-8"
                            >         config.version = "8"
                            >         virtualHW.version = "8"
                            >         nvram = 
"533b6fcf3fa6301aadcc2b168f3f999a.nvram"
                            >         pciBridge0.present = "TRUE"
                            >         svga.present = "TRUE"
                            >         pciBridge4.present = "TRUE"
                            >         pciBridge4.virtualDev = "pcieRootPort"
                            >         pciBridge4.functions = "8"
                            >         pciBridge5.present = "TRUE"
                            >         pciBridge5.virtualDev = "pcieRootPort"
                            >         pciBridge5.functions = "8"
                            >         pciBridge6.present = "TRUE"
                            >         pciBridge6.virtualDev = "pcieRootPort"
                            >         pciBridge6.functions = "8"
                            >         pciBridge7.present = "TRUE"
                            >         pciBridge7.virtualDev = "pcieRootPort"
                            >         pciBridge7.functions = "8"
                            >         vmci0.present = "TRUE"
                            >         hpet0.present = "TRUE"
                            >         floppy0.present = "FALSE"
                            >         memSize = "256"
                            >         scsi0.virtualDev = "lsilogic"
                            >         scsi0.present = "TRUE"
                            >         ide0:0.startConnected = "FALSE"
                            >         ide0:0.deviceType = "atapi-cdrom"
                            >         ide0:0.fileName = "CD/DVD drive 0"
                            >         ide0:0.present = "TRUE"
                            >         scsi0:0.deviceType = "scsi-hardDisk"
                            >         scsi0:0.fileName = 
"533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk"
                            >         scsi0:0.present = "TRUE"
                            >         displayName = 
"533b6fcf3fa6301aadcc2b168f3f999a"
                            >         annotation = 
"systemvmtemplate-4.11.2.0-vmware"
                            >         guestOS = "otherlinux-64"
                            >         toolScripts.afterPowerOn = "TRUE"
                            >         toolScripts.afterResume = "TRUE"
                            >         toolScripts.beforeSuspend = "TRUE"
                            >         toolScripts.beforePowerOff = "TRUE"
                            >         uuid.bios = "42 02 f1 40 33 e8 de e5-1a 
c5 93 2a c9 12 47 61"
                            >         vc.uuid = "50 02 5b d9 e9 c9 77 86-28 3e 
84 00 22 2b eb d3"
                            >         firmware = "bios"
                            >         migrate.hostLog = 
"533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog"
                            >
                            >
                            >         6) *.vmsd file content:
                            >
                            >         cat
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a.vmsd
                            >         .encoding = "UTF-8"
                            >         snapshot.lastUID = "1"
                            >         snapshot.current = "1"
                            >         snapshot0.uid = "1"
                            >         snapshot0.filename =
                            > "533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn"
                            >         snapshot0.displayName = 
"cloud.template.base"
                            >         snapshot0.description = "Base snapshot"
                            >         snapshot0.createTimeHigh = "363123"
                            >         snapshot0.createTimeLow = "-679076964"
                            >         snapshot0.numDisks = "1"
                            >         snapshot0.disk0.fileName = 
"533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
                            >         snapshot0.disk0.node = "scsi0:0"
                            >         snapshot.numSnapshots = "1"
                            >
                            >         7) *-Snapshot1.vmsn content:
                            >
                            >         cat
                            > 
/vmfs/volumes/42054b8459633172be231d72a52d59d4/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-Snapshot1.vmsn
                            >
                            >         ҾSnapshot\?%?cfgFilet%t%.encoding = 
"UTF-8"
                            >         config.version = "8"
                            >         virtualHW.version = "8"
                            >         nvram = 
"533b6fcf3fa6301aadcc2b168f3f999a.nvram"
                            >         pciBridge0.present = "TRUE"
                            >         svga.present = "TRUE"
                            >         pciBridge4.present = "TRUE"
                            >         pciBridge4.virtualDev = "pcieRootPort"
                            >         pciBridge4.functions = "8"
                            >         pciBridge5.present = "TRUE"
                            >         pciBridge5.virtualDev = "pcieRootPort"
                            >         pciBridge5.functions = "8"
                            >         pciBridge6.present = "TRUE"
                            >         pciBridge6.virtualDev = "pcieRootPort"
                            >         pciBridge6.functions = "8"
                            >         pciBridge7.present = "TRUE"
                            >         pciBridge7.virtualDev = "pcieRootPort"
                            >         pciBridge7.functions = "8"
                            >         vmci0.present = "TRUE"
                            >         hpet0.present = "TRUE"
                            >         floppy0.present = "FALSE"
                            >         memSize = "256"
                            >         scsi0.virtualDev = "lsilogic"
                            >         scsi0.present = "TRUE"
                            >         ide0:0.startConnected = "FALSE"
                            >         ide0:0.deviceType = "atapi-cdrom"
                            >         ide0:0.fileName = "CD/DVD drive 0"
                            >         ide0:0.present = "TRUE"
                            >         scsi0:0.deviceType = "scsi-hardDisk"
                            >         scsi0:0.fileName = 
"533b6fcf3fa6301aadcc2b168f3f999a.vmdk"
                            >         scsi0:0.present = "TRUE"
                            >         displayName = 
"533b6fcf3fa6301aadcc2b168f3f999a"
                            >         annotation = 
"systemvmtemplate-4.11.2.0-vmware"
                            >         guestOS = "otherlinux-64"
                            >         toolScripts.afterPowerOn = "TRUE"
                            >         toolScripts.afterResume = "TRUE"
                            >         toolScripts.beforeSuspend = "TRUE"
                            >         toolScripts.beforePowerOff = "TRUE"
                            >         uuid.bios = "42 02 f1 40 33 e8 de e5-1a 
c5 93 2a c9 12 47 61"
                            >         vc.uuid = "50 02 5b d9 e9 c9 77 86-28 3e 
84 00 22 2b eb d3"
                            >         firmware = "bios"
                            >         migrate.hostLog = 
"533b6fcf3fa6301aadcc2b168f3f999a-7d5d73de.hlog"
                            >
                            >
                            >         ------------
                            >
                            >         That's all the data on the template VMDK.
                            >
                            >         Much appreciate your time!
                            >
                            >         Yiping
                            >
                            >
                            >
                            >         On 6/4/19, 9:29 AM, "Sergey Levitskiy" 
<serg...@hotmail.com>
                            > wrote:
                            >
                            >             Have you tried deleting template from 
PS and let ACS to recopy
                            > it again? If the issue is reproducible we can try 
to look what is wrong
                            > with VMDK. Please post content of 
533b6fcf3fa6301aadcc2b168f3f999a.vmdk ,
                            > 533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk and
                            > 533b6fcf3fa6301aadcc2b168f3f999a.vmx (their 
equitant after ACS finishes
                            > copying template). Also from one of your ESX 
hosts output of this
                            >             ls -al /vmfs/volumes
                            >             ls -al 
/vmfs/volumes/*/533b6fcf3fa6301aadcc2b168f3f999a (their
                            > equitant after ACS finishes copying template)
                            >
                            >              Can you also post management server 
log starting from the
                            > point you unregister and delete template from the 
vCenter.
                            >
                            >             On 6/4/19, 8:37 AM, "Yiping Zhang" 
<yipzh...@adobe.com.INVALID>
                            > wrote:
                            >
                            >                 I have manually imported the OVA 
to vCenter and
                            > successfully cloned a VM instance with it, on the 
same NFS datastore.
                            >
                            >
                            >                 On 6/4/19, 8:25 AM, "Sergey 
Levitskiy" <
                            > serg...@hotmail.com> wrote:
                            >
                            >                     I would suspect the template 
is corrupted on the
                            > secondary storage. You can try disabling/enabling 
link clone feature and
                            > see if it works the other way.
                            >                     vmware.create.full.clone      
              false
                            >
                            >                     Also systemVM template might 
have been generated on a
                            > newer version of vSphere and not compatible with 
ESXi 6.5. What you can do
                            > to validate this is to manually deploy OVA that 
is in Secondary storage and
                            > try to spin up VM from it directly in vCenter.
                            >
                            >
                            >
                            >                     On 6/3/19, 5:41 PM, "Yiping 
Zhang"
                            > <yipzh...@adobe.com.INVALID> wrote:
                            >
                            >                         Hi, list:
                            >
                            >                         I am struggling with 
deploying a new advanced zone
                            > using ACS 4.11.2.0 + vSphere 6.5 + NetApp volumes 
for primary and secondary
                            > storage devices. The initial setup of CS 
management server, seeding of
                            > systemVM template, and advanced zone deployment 
all went smoothly.
                            >
                            >                         Once I enabled the zone 
in web UI and the systemVM
                            > template gets copied/staged on to primary storage 
device. But subsequent VM
                            > creations from this template would fail with 
errors:
                            >
                            >
                            >                         2019-06-03 18:38:15,764 
INFO  [c.c.h.v.m.HostMO]
                            > (DirectAgent-7:ctx-d01169cb 
esx-0001-a-001.example.org, job-3/job-29,
                            > cmd: CopyCommand) VM 
533b6fcf3fa6301aadcc2b168f3f999a not found in host
                            > cache
                            >
                            >                         2019-06-03 18:38:17,017 
INFO
                            > [c.c.h.v.r.VmwareResource] 
(DirectAgent-4:ctx-08b54fbd
                            > esx-0001-a-001.example.org, job-3/job-29, cmd: 
CopyCommand)
                            > VmwareStorageProcessor and 
VmwareStorageSubsystemCommandHandler
                            > successfully reconfigured
                            >
                            >                         2019-06-03 18:38:17,128 
INFO
                            > [c.c.s.r.VmwareStorageProcessor] 
(DirectAgent-4:ctx-08b54fbd
                            > esx-0001-a-001.example.org, job-3/job-29, cmd: 
CopyCommand) creating full
                            > clone from template
                            >
                            >                         2019-06-03 18:38:17,657 
INFO
                            > [c.c.h.v.u.VmwareHelper] 
(DirectAgent-4:ctx-08b54fbd
                            > esx-0001-a-001.example.org, job-3/job-29, cmd: 
CopyCommand)
                            > [ignored]failed toi get message for exception: 
Error caused by file
                            > 
/vmfs/volumes/afc5e946-03bfe3c2/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
                            >
                            >                         2019-06-03 18:38:17,658 
ERROR
                            > [c.c.s.r.VmwareStorageProcessor] 
(DirectAgent-4:ctx-08b54fbd
                            > esx-0001-a-001.example.org, job-3/job-29, cmd: 
CopyCommand) clone volume
                            > from base image failed due to Exception: 
java.lang.RuntimeException
                            >
                            >                         Message: Error caused by 
file
                            > 
/vmfs/volumes/afc5e946-03bfe3c2/533b6fcf3fa6301aadcc2b168f3f999a/533b6fcf3fa6301aadcc2b168f3f999a-000001.vmdk
                            >
                            >
                            >
                            >                         If I try to create “new 
VM from template”
                            > (533b6fcf3fa6301aadcc2b168f3f999a) on vCenter UI 
manually,  I will receive
                            > exactly the same error message. The name of the 
VMDK file in the error
                            > message is a snapshot of the base disk image, but 
it is not part of the
                            > original template OVA on the secondary storage.  
So, in the process of
                            > copying the template from secondary to primary 
storage, a snapshot got
                            > created and the disk became corrupted/unusable.
                            >
                            >                         Much later in the log 
file,  there is another
                            > error message “failed to fetch any free public IP 
address” (for ssvm, I
                            > think).  I don’t know if these two errors are 
related or if one is the root
                            > cause for the other error.
                            >
                            >                         The full management 
server log is uploaded as
                            > 
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpastebin.com%2Fc05wiQ3R&amp;data=02%7C01%7Cyipzhang%40adobe.com%7C557bf647ff13413c66b708d6ea87220d%7Cfa7b1b5a7b34438794aed2c178decee1%7C0%7C0%7C636954263448732817&amp;sdata=NY2nAe8v%2BP7ANlpjD4xcmOSc7IoBpFizoX3eCuclUHo%3D&amp;reserved=0
                            >
                            >                         Any help or insight on 
what went wrong here are
                            > much appreciated.
                            >
                            >                         Thanks
                            >
                            >                         Yiping
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            >
                            
                            -- 
                            
                            Andrija Panić
                            
                        
                        
                    
                    
                
                
            
            
        
        
    
    

Reply via email to