Yes... I wait for that... It's work fine about 2/3 day ago... Since yesterday, I get this odd behavior...
Right now, I try create other VM, and get this in PVE01: pve01:/var/log# tail -f /var/log/syslog | grep drbd Sep 11 11:21:44 pve01 kernel: [ 2838.206262] drbd .drbdctrl: Preparing cluster-wide state change 3187852087 (0->-1 3/1) Sep 11 11:21:44 pve01 kernel: [ 2838.206265] drbd .drbdctrl: Committing cluster-wide state change 3187852087 (0ms) Sep 11 11:21:44 pve01 kernel: [ 2838.206274] drbd .drbdctrl: role( Secondary -> Primary ) Sep 11 11:21:44 pve01 kernel: [ 2838.268037] drbd .drbdctrl: role( Primary -> Secondary ) Sep 11 11:21:44 pve01 kernel: [ 2838.276870] drbd .drbdctrl: Preparing cluster-wide state change 3555025472 (0->-1 3/1) Sep 11 11:21:44 pve01 kernel: [ 2838.276872] drbd .drbdctrl: Committing cluster-wide state change 3555025472 (0ms) Sep 11 11:21:44 pve01 kernel: [ 2838.276880] drbd .drbdctrl: role( Secondary -> Primary ) Sep 11 11:21:44 pve01 kernel: [ 2838.302317] drbd .drbdctrl: role( Primary -> Secondary ) Sep 11 11:21:44 pve01 kernel: [ 2838.310209] drbd .drbdctrl: Preparing cluster-wide state change 2077508299 (0->-1 3/1) Sep 11 11:21:44 pve01 kernel: [ 2838.310211] drbd .drbdctrl: Committing cluster-wide state change 2077508299 (0ms) Sep 11 11:21:44 pve01 kernel: [ 2838.310219] drbd .drbdctrl: role( Secondary -> Primary ) Sep 11 11:21:44 pve01 kernel: [ 2838.335623] drbd .drbdctrl: role( Primary -> Secondary ) Sep 11 11:21:44 pve01 kernel: [ 2838.376006] drbd .drbdctrl: Preparing cluster-wide state change 4247485156 (0->-1 3/1) Sep 11 11:21:44 pve01 kernel: [ 2838.376008] drbd .drbdctrl: Committing cluster-wide state change 4247485156 (0ms) Sep 11 11:21:44 pve01 kernel: [ 2838.376017] drbd .drbdctrl: role( Secondary -> Primary ) Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: Failed to find logical volume "drbdpool/vm-103-disk-1_00" Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: Rounding up size to full physical extent 32.01 GiB Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: Logical volume "vm-103-disk-1_00" created. Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: drbdmeta: unrecognized option '--max-peers=7' Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: NOT initializing bitmap Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: initializing activity log Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: Writing meta data... Sep 11 11:21:45 pve01 org.drbd.drbdmanaged[1309]: New drbd meta data block successfully created. Sep 11 11:21:45 pve01 kernel: [ 2839.476032] drbd vm-103-disk-1: Starting worker thread (from drbdsetup [3813]) Sep 11 11:21:45 pve01 kernel: [ 2839.516041] drbd vm-103-disk-1/0 drbd10: disk( Diskless -> Attaching ) Sep 11 11:21:45 pve01 kernel: [ 2839.516052] drbd vm-103-disk-1/0 drbd10: Maximum number of peer devices = 7 Sep 11 11:21:45 pve01 kernel: [ 2839.516184] drbd vm-103-disk-1: Method to ensure write ordering: flush Sep 11 11:21:45 pve01 kernel: [ 2839.516187] drbd vm-103-disk-1/0 drbd10: drbd_bm_resize called with capacity == 67108864 Sep 11 11:21:45 pve01 kernel: [ 2839.522395] drbd vm-103-disk-1/0 drbd10: resync bitmap: bits=8388608 words=917504 pages=1792 Sep 11 11:21:45 pve01 kernel: [ 2839.530295] drbd vm-103-disk-1/0 drbd10: recounting of set bits took additional 8ms Sep 11 11:21:45 pve01 kernel: [ 2839.530305] drbd vm-103-disk-1/0 drbd10: disk( Attaching -> Inconsistent ) Sep 11 11:21:45 pve01 kernel: [ 2839.530309] drbd vm-103-disk-1/0 drbd10: attached to current UUID: 0000000000000004 Sep 11 11:21:45 pve01 kernel: [ 2839.537821] drbd vm-103-disk-1 pve02: Starting sender thread (from drbdsetup [3826]) Sep 11 11:21:45 pve01 kernel: [ 2839.539000] drbd vm-103-disk-1 pve02: conn( StandAlone -> Unconnected ) Sep 11 11:21:45 pve01 kernel: [ 2839.541112] drbd vm-103-disk-1: Preparing cluster-wide state change 839975239 (1->-1 7683/4609) Sep 11 11:21:45 pve01 kernel: [ 2839.541114] drbd vm-103-disk-1: Committing cluster-wide state change 839975239 (0ms) Sep 11 11:21:45 pve01 kernel: [ 2839.541123] drbd vm-103-disk-1: role( Secondary -> Primary ) Sep 11 11:21:45 pve01 kernel: [ 2839.541124] drbd vm-103-disk-1/0 drbd10: disk( Inconsistent -> UpToDate ) Sep 11 11:21:45 pve01 kernel: [ 2839.545440] drbd vm-103-disk-1 pve02: Starting receiver thread (from drbd_w_vm-103-d [3815]) Sep 11 11:21:45 pve01 kernel: [ 2839.545509] drbd vm-103-disk-1/0 drbd10: size = 32 GB (33554432 KB) Sep 11 11:21:45 pve01 kernel: [ 2839.553772] drbd vm-103-disk-1: Forced to consider local data as UpToDate! Sep 11 11:21:45 pve01 kernel: [ 2839.553778] drbd vm-103-disk-1/0 drbd10: new current UUID: F485EDDF7BDE1589 weak: FFFFFFFFFFFFFFFD Sep 11 11:21:45 pve01 kernel: [ 2839.562119] drbd vm-103-disk-1 pve02: conn( Unconnected -> Connecting ) Sep 11 11:21:45 pve01 kernel: [ 2839.563982] drbd vm-103-disk-1: role( Primary -> Secondary ) Sep 11 11:21:46 pve01 kernel: [ 2839.726144] drbd .drbdctrl: role( Primary -> Secondary ) But nothing in syslog from PVE02.... 2015-09-11 11:17 GMT-03:00 Dietmar Maurer <diet...@proxmox.com>: > > As can you see, the disc 101 exist in PVE01 but not in PVE02. > > How can I force drbdmanage to sync or whatever to show disc 101 in both > > servers?? > > Is that reproducible? Did you wait until initial sync was finished? > Any hints in /var/log/syslog? > > -- Gilberto Ferreira +55 (47) 9676-7530 Skype: gilberto.nunes36
_______________________________________________ pve-devel mailing list pve-devel@pve.proxmox.com http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel