Hi Gilberto, and Thank you for getting back to me.

Just to be 100% clear: the Proxmox (with Hyper-Converged Ceph) Cluster is working AOK, except for the face I can't migrate *any* of the VMs (live or shutdown).

Yes, I can SSH into each node from every other node using:

 * the hostname of the "management" NIC
 * the hostname of the "migration traffic" NIC
 * the ip address of the "managment" NIC
 * the IP address of the "migration traffic" NIC

The VM's HDD is on the rbd storage (see below).

As requested:

/etc/pve/storage.cfg

```

dir: local
       path /var/lib/vz
       content vztmpl,iso,backup

lvmthin: local-lvm
       thinpool data
       vgname pve
       content images,rootdir

rbd: rbd
       content images,rootdir
       krbd 0
       pool rbd

cephfs: cephfs
       path /data/cephfs
       content backup,vztmpl,iso
       fs-name cephfs
```

/etc/pve/datacenter.cfg

```

console: html5
crs: ha-rebalance-on-start=1
ha: shutdown_policy=migrate
keyboard: en-us
migration: secure,network=192.168.200.0/24
next-id: lower=1000
```

/etc/pve/corosync.conf

```

logging {
 debug: off
 to_syslog: yes
}

nodelist {
 node {
   name: pven1
   nodeid: 1
   quorum_votes: 1
   ring0_addr: 192.168.100.101
   ring1_addr: 192.168.200.101
 }
 node {
   name: pven2
   nodeid: 2
   quorum_votes: 1
   ring0_addr: 192.168.100.102
   ring1_addr: 192.168.200.102
}
 node {
   name: pven3
   nodeid: 3
   quorum_votes: 1
   ring0_addr: 192.168.100.103
   ring1_addr: 192.168.200.103
}
}

quorum {
 provider: corosync_votequorum
}

totem {
 cluster_name: cluster1
 config_version: 4
 interface {
   knet_link_priority: 10
   linknumber: 0
 }
 interface {
   knet_link_priority: 20
   linknumber: 1
 }
 ip_version: ipv4-6
 link_mode: passive
 secauth: on
 version: 2
}

```

/etc/hosts

```

127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1         localhost localhost.localdomain localhost6 localhost6.localdomain6 ip6-localhost ip6-loopback
192.168.100.101  pven1.mydomain.local pven1
192.168.100.102 pven2.mydomain.local pven2
192.168.100.103 pven3.mydomain.local pven3
192.168.200.101 pvent1.mydomain.local pvent1
192.168.200.102 pvent2.mydomain.local pvent2
192.168.200.103 pvent3.mydomain.local pvent3
```

/etc/network/interfaces (for pven1; pven2 & pven3 are the same, except for the IP Address (see above))

```

auto lo
iface lo inet loopback

iface eno1 inet manual

iface eno2 inet manual

auto enp2s0
iface enp2s0 inet static
 address 192.168.200.20/24
 mtu 9000

auto bond0
iface bond0 inet manual
 bond-slaves eno1 eno2
 bond-mode 802.3ad
 bond-xmit-hash-policy layer2+3
 bond-miimon 100
 bond-downdelay 200
 bond-updelay 200

auto vmbr0
iface vmbr0 inet static
 bridge-ports bond0
 bridge-stp off
 bridge-fd 0
 bridge-vlan-aware yes
 bridge-vids 100,200

auto vmbr0.200
iface vmbr0.200 inet static
 address 192.168.100.101/24
 gateway 192.168.100.1
```

Note: iface enp2s0 (on all 3 Nodes) sits on an isolated VLAN which (obviously) has no gateway ie the only hosts on that VLAN are pven1, pven2, & pven3, and all are "pingable" from each other.

Thanks for the help

Dulux-Oz


On 2/9/24 08:02, Gilberto Ferreira wrote:
Hi
We need more details.
Send us the following:

cat /etc/pve/storage
cat /etc/pvedatacenter
cat /etc/pve/corosync.conf
cat /etc/hosts
cat /etc/network/interfaces

Can you ssh between the nodes?


---


Gilberto Nunes Ferreira
(47) 99676-7530 - Whatsapp / Telegram






Em dom., 1 de set. de 2024 às 05:11, duluxoz <dulu...@gmail.com> escreveu:

    Hi All,

    I need help with figuring out why I can't migrate a VM from one
    Proxmox
    Node to another (in the same cluster, of course).

    These are the details provided by the Proxmox Task Log:

    ```

    task started by HA resource agent
    2024-09-01 18:02:30 use dedicated network address for sending
    migration
    traffic (192.168.200.103)
    2024-09-01 18:02:30 starting migration of VM 100 to node 'pven3'
    (192.168.200.103)
    2024-09-01 18:02:30 starting VM 100 on remote node 'pven3'
    2024-09-01 18:02:30 [pven3]
    2024-09-01 18:02:32 start remote tunnel
    2024-09-01 18:02:33 ssh tunnel ver 1
    2024-09-01 18:02:33 starting online/live migration on
    unix:/run/qemu-server/100.migrate
    2024-09-01 18:02:33 set migration capabilities
    2024-09-01 18:02:33 migration downtime limit: 100 ms
    2024-09-01 18:02:33 migration cachesize: 256.0 MiB
    2024-09-01 18:02:33 set migration parameters
    2024-09-01 18:02:33 start migrate command to
    unix:/run/qemu-server/100.migrate
    channel 2: open failed: connect failed: open failed
    2024-09-01 18:02:34 migration status error: failed - Unable to
    write to
    socket: Broken pipe
    2024-09-01 18:02:34 ERROR: online migrate failure - aborting
    2024-09-01 18:02:34 aborting phase 2 - cleanup resources
    2024-09-01 18:02:34 migrate_cancel
    2024-09-01 18:02:36 ERROR: migration finished with problems (duration
    00:00:07)
    TASK ERROR: migration problems
    ```

    If someone could point me in the correct direction to resolve this
    issue
    I'd be very grateful - thanks

    Cheer

    Dulux-Oz


    _______________________________________________
    pve-user mailing list
    pve-user@lists.proxmox.com
    https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user

_______________________________________________
pve-user mailing list
pve-user@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user

Reply via email to