Hello Rainer,
I've got pretty the same issue (drbd 8.3.0):
I've got a huge drbd in pri/pri-role. Then ontop LVM with xfs in the LV's.
Now the XM-VM boots into the LV, no prob.
If i now want to migrate, i get the same error as you, but in my case, the drbd
logs something like "split-brain detected -> disconnected" in dmesg, perhaps
you've got the same?
drbd0: Split-Brain detected, dropping connection!
drbd0: self 20D88E3F20F7E8C9:11227E17F1A34EBD:F14436F7DEC14D2E:D51BA840A9E19E2D
drbd0: peer 5664952031DE8E53:11227E17F1A34EBD:F14436F7DEC14D2E:D51BA840A9E19E2D
drbd0: helper command: /sbin/drbdadm split-brain minor-0
drbd0: helper command: /sbin/drbdadm split-brain minor-0 exit code 0 (0x0)
drbd0: conn( WFReportParams -> Disconnecting )
drbd0: error receiving ReportState, l: 4!
drbd0: asender terminated
drbd0: Terminating asender thread
drbd0: Connection closed
drbd0: conn( Disconnecting -> StandAlone )
And on the other node
drbd0: meta connection shut down by peer.
drbd0: peer( Primary -> Unknown ) conn( Connected -> NetworkFailure ) pdsk(
UpToDate -> DUnknown )
drbd0: asender terminated
drbd0: Terminating asender thread
drbd0: sock was shut down by peer
drbd0: short read expecting header on sock: r=0
drbd0: Creating new current UUID
drbd0: Connection closed
drbd0: conn( NetworkFailure -> Unconnected )
drbd0: receiver terminated
It seems, that while migrating there is some overlapping write access to the
drbd, what let's drbd decide to have s split-brain.
Even if i
after-sb-0pri discard-zero-changes;
after-sb-1pri violently-as0p;
after-sb-2pri violently-as0p;
I get this log-entries....
So that's the point where a DLM-aware filesystem comes into play.
The question is now, how we have to "mix" drbd, lvm and ocfs2/gfs for a working
playground... ;)
I previously had implemented the drbd -> ocfs2 -> sparse-files-way, that one
failed with drbd-split-brain, too.
Perhaps drbd -> lvm -> ocfs2 and mounting that volume in the vm could do the
trick?
Cheers, Florian
> -----Ursprüngliche Nachricht-----
> Von: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
> [mailto:xen-users-bounces@xxxxxxxxxxxxxxxxxxx] Im Auftrag von
> Rainer Sokoll
> Gesendet: Dienstag, 3. März 2009 15:36
> An: Nick Couchman
> Cc: xen-users@xxxxxxxxxxxxxxxxxxx
> Betreff: Re: [Xen-users] xm migrate headache
>
> On Tue, Mar 03, 2009 at 07:27:18AM -0700, Nick Couchman wrote:
>
> > So, in your Xen configs, are you using phy: to access the
> DRBD devices
> > directly?
>
> Yes, you are correct. I've used a file based storage backend
> before, but it was too slow.
>
> drbd.conf:
> ----8<----
> resource stunnel {
> on jitxen01 {
> address 192.168.0.1:7794;
> device /dev/drbd4;
> disk /dev/XEN/stunnel;
> meta-disk internal;
> }
> on jitxen02 {
> address 192.168.0.2:7794;
> device /dev/drbd4;
> disk /dev/XEN/stunnel;
> meta-disk internal;
> }
> net {
> allow-two-primaries;
> after-sb-0pri discard-zero-changes;
> after-sb-1pri discard-secondary;
> }
> }
> ----8<----
>
> machine config in xen:
>
> ----8<----
> name="stunnel"
> [...]
> disk=[ 'phy:/dev/drbd4,xvda,w', ]
> ----8<----
>
> > If so, then you should be okay without a cluster-aware filesystem
>
> Now I feel much better :-)
>
> > - sorry about that, I was under the impression that you
> were mounting
> > the DRBD device on the dom0s and then storing disk files on it.
>
> This sounds like a not-so-clever idea to me :-)
>
> Rainer
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users
>
**********************************************************************************************
IMPORTANT: The contents of this email and any attachments are confidential.
They are intended for the
named recipient(s) only.
If you have received this email in error, please notify the system manager or
the sender immediately and do
not disclose the contents to anyone or make copies thereof.
*** eSafe scanned this email for viruses, vandals, and malicious content. ***
**********************************************************************************************
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
|