Hi Leo,

Leo Liu wrote, On 16/06/2009 09:39:
> ENV:
> SPARC-Enterprise-T5220 boxes
> Sunt5220b is the source node with source domain ldom2, ldom2 used a flat file 
> as its vdisk(as a full disk) which locates on a share disk;
> Sunt5220a is the target node, the target domain still used a flat file as its 
> vdisk which also locates on a share disk;
> 

Are both domains backed by the same file though ? It doesn't look
like it below. The same disk backend needs to be added as a vdsdev
on both the source & target systems - it's not sufficient to have
the just the same name on the target for a different backend.
The disk contents are not migrated across.

- Liam


> 1.    Create ldom2 on sunt5220b, ldom2 used a flat file /ldom2/bootdisk as 
> its vdisk, which locates on a vxfs file-system /dev/vx/dsk/ldom2/ldomvol2;
> 
> 2.    On target node sunt5220a, mount a vxfs file-system on /ldom2 and create 
> a flatfile on it;
> [root at sunt5220a config]#>mount -F vxfs /dev/vx/dsk/ldomdg3/ldomvol3 /ldom2/
> [root at sunt5220a config]#>mkfile 17G /ldom2/bootdisk
> 
> 3.    Add /ldom2/bootdisk  in VDS of primary domain, which is the same 
> configuration as source node sunt5220b;
> [root at sunt5220b config]#>ldm list-bindings
> VDS
>     NAME             VOLUME         OPTIONS          MPGROUP        DEVICE
>     primary-vds0     ldom2_boot                                     
> /ldom2/bootdisk
> [root at sunt5220a config]#>ldm list-bindings
> VDS
> NAME             VOLUME         OPTIONS          MPGROUP        DEVICE
> primary-vds0     ldom2_boot                                     
> /ldom2/bootdisk
> 
> 4.    Migrate ldom2 from source node sunt5220b to target node sunt5220a, the 
> migration was successful without any error.
> [root at sunt5220b config]#>ldm migrate-domain ldom2 sunt5220a
> Target Password:
> 
> 8.    On the target node, we can see that ldom2 is here with ?n? normal flag
> [root at sunt5220a /]#>ldm list
> NAME             STATE      FLAGS   CONS    VCPU  MEMORY   UTIL  UPTIME
> primary          active     -n-cv-  SP      48    14G      0.1%  3h 49m
> ldom2            active     -n----  5000    8     1G       0.1%  21m
> 
> Everything looks fine here EXCEPT:
> 1.    We failed to login ldom2,  login hang here
> [root at sunt5220a /]#>telnet localhost 5000
> Trying 127.0.0.1...
> Connected to localhost.
> Escape character is '^]'.
> 
> Connecting to console "ldom2" in group "ldom2" ....
> Press ~? for control options ..
> 
> 2.    stop ldom on target node, then restart it, ldom failed to start
> [root at sunt5220a /]#>telnet localhost 5000
> Trying 127.0.0.1...
> Connected to localhost.
> Escape character is '^]'.
> 
> Connecting to console "ldom2" in group "ldom2" ....
> Press ~? for control options ..
> 
> {0} ok boot
> Bad magic number in disk label
> ERROR: /virtual-devices at 100/channel-devices at 200/disk at 0: Can't open 
> disk label package
> 
> ERROR: boot-read fail
> Boot device: ldom2_boot  File and args:
> Bad magic number in disk label
> ERROR: /virtual-devices at 100/channel-devices at 200/disk at 0: Can't open 
> disk label package
> 
> ERROR: boot-read fail
> 
> Can't open boot device
> 
> 
> Could anybody please make some comments on it, thanks so much.



Reply via email to