WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-users

Re: [Xen-users] Re: [Fedora-xen] Guest VM dropping connection.

To: "Mark Nielsen" <mnielsen@xxxxxxxxxx>
Subject: Re: [Xen-users] Re: [Fedora-xen] Guest VM dropping connection.
From: "Paul Reiber" <reiber@xxxxxxxxx>
Date: Tue, 15 Jan 2008 13:14:21 -0800
Cc: Ubaidul Khan <ukhanlists@xxxxxxxxxxx>, xen-users@xxxxxxxxxxxxxxxxxxx, fedora-xen@xxxxxxxxxx
Delivery-date: Tue, 15 Jan 2008 13:15:17 -0800
Dkim-signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:message-id:date:from:reply-to:to:subject:cc:in-reply-to:mime-version:content-type:content-transfer-encoding:content-disposition:references; bh=PNdiSH6VPcRZ0ZElhMzZhvGR12rbAZKDPZvAAWjMNyE=; b=QUWXYMGVTVKo6JCWbyT25cUdP0jv82gzgWsEZyPdRBRI212C2E6mFqzBXZ/Mok9nNEtXGo01wbIxrh/0yhShLQHZwetI1518VCWX3cIAbDEP4vD4+CeGqIsrZKsnZX6uVh11xGPJInr10kus1OY1Uxzrg+8Lj+7EacdlwtcEBz8=
Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=message-id:date:from:reply-to:to:subject:cc:in-reply-to:mime-version:content-type:content-transfer-encoding:content-disposition:references; b=ahtSQ3b9+uNghwOFuJ83YBbdd9dHAJI6Gw5s1G9wdJ80yaP6MsGnJ9PZkiG9NwRaRIv0GGr6chGJg0stdUsTIddT9JOv6W0JRyGla+QjBhIKsJnlzybutkWD5IIU3PAmRrIwQLr1SAajfDSOGIyAZuKwu7ZggnQnjsd5qP872c0=
Envelope-to: www-data@xxxxxxxxxxxxxxxxxx
In-reply-to: <478CFB41.7020501@xxxxxxxxxx>
List-help: <mailto:xen-users-request@lists.xensource.com?subject=help>
List-id: Xen user discussion <xen-users.lists.xensource.com>
List-post: <mailto:xen-users@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=unsubscribe>
References: <BAY119-F299C8C4AE3D185AF538A91C7EF0@xxxxxxx> <BAY119-W7DA205642E05359EE8F82C7470@xxxxxxx> <478CFB41.7020501@xxxxxxxxxx>
Reply-to: reiber@xxxxxxxxx
Sender: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
Ubaidul, you might wanna try this:

SSH into the VM from elsewhere.  Make sure the window SSH is running
in will be able to be left running for a _long_ while, (at least, in
so far as it's still there after the VM's lost its network connection)
... then leave it running a simple makeshift monitor.

Here's an easy makeshift error monitor: go into /var/log, and run a
bunch of backgrounded 'tail -f' commands on various logfiles on the
system, and leave that window up and running somewhere.  Then when the
problem happens again, you can look over and see what errors have been
reported recently.

Not all errors are logged in /var/log/messages... this is a common
misconception, though, so no worries.

Please let me know if this helps!

Regards,
-pbr



On Jan 15, 2008 10:28 AM, Mark Nielsen <mnielsen@xxxxxxxxxx> wrote:
> are you seeing any messages on the host about the network bouncing?
>
> Ubaidul Khan wrote:
> > I am still experiencing the same problems but I am seeing some weird 
> > messages in /var/log/messages:
> >
> > Jan 15 10:48:32 tsubo-dr kernel: device eth0 left promiscuous mode
> > Jan 15 10:49:50 tsubo-dr kernel: Removing netfilter NETLINK layer.
> > Jan 15 10:49:50 tsubo-dr kernel: ip_tables: (C) 2000-2006 Netfilter Core 
> > Team
> > Jan 15 10:49:50 tsubo-dr kernel: Netfilter messages via NETLINK v0.30.
> > Jan 15 10:49:50 tsubo-dr kernel: ip_conntrack version 2.4 (8192 buckets, 
> > 65536 max) - 304 bytes per conntrack
> > Jan 15 10:49:56 tsubo-dr kernel: device eth0 entered promiscuous mode
> > Jan 15 10:50:25 tsubo-dr kernel: device eth0 left promiscuous mode
> > Jan 15 10:54:08 tsubo-dr kernel: device eth0 entered promiscuous mode
> > Jan 15 10:54:37 tsubo-dr kernel: device eth0 left promiscuous mode
> >
> > The loss of connection is happening during the windows when the interface 
> > enters/leaves promiscuious mode.
> > ----------------------------------------
> >
> >> From: ukhanlists@xxxxxxxxxxx
> >> To: xen-users@xxxxxxxxxxxxxxxxxxx; fedora-xen@xxxxxxxxxx
> >> Date: Mon, 30 Jul 2007 22:04:45 -0400
> >> CC:
> >> Subject: [Fedora-xen] Guest VM dropping connection.
> >>
> >> I am running RHEL5 Guest and Host.  The guest OS is dropping network
> >> connections occasionally(I noticed my ssh session suddenly terminated).  
> >> The
> >> guest is configured with private network interface and a public network
> >> interface.  Following are the bridge/network settings for the host:
> >>
> >> bridge name     bridge id               STP enabled     interfaces
> >> xenbr0          8000.feffffffffff              no                 vif2.0
> >>
> >> vif1.0
> >>
> >> peth0
> >>
> >> vif0.0
> >>
> >> xenbr1          8000.fefffffffffe             no                 vif2.1
> >>
> >> peth1
> >>
> >> vif0.3
> >>
> >> xenbr1 is dedicated to provide the private network link for the guest os.
> >> Following is the guest OS configuration:
> >>
> >> # Automatically generated xen config file
> >> name = "saleen_webvm0"
> >> memory = "1000"
> >> disk = [
> >> 'tap:aio:/vms/img/redhat/linux/es/5.0/x86_64/saleen_webvm0.img,xvda,w','phy:/dev/sdc,xvdb,w'
> >> ]
> >> vif = [ 'mac=00:16:3e:2e:6f:e1, bridge=xenbr0','mac=00:16:3e:2f:9e:b1,
> >> bridge=xenbr1',]
> >> # vif = [ 'mac=00:16:3e:2e:6f:e1, bridge=xenbr0',]
> >> vfb = ["type=vnc,vncunused=1"]
> >> uuid = "122374dc-26a3-cb0d-99f8-3393e2481777"
> >> bootloader="/usr/bin/pygrub"
> >> vcpus=2
> >> on_reboot   = 'restart'
> >> on_crash    = 'restart'
> >>
> >> Following is the output of ifconfig:
> >>
> >> eth0      Link encap:Ethernet  HWaddr 00:14:4F:40:D8:70
> >>           inet addr:xxx.xxx.xxx.xxx  Bcast:xxx.xxx.xxx.255
> >> Mask:255.255.255.0
> >>           inet6 addr: fe80::214:4fff:fe40:d870/64 Scope:Link
> >>           UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>           RX packets:79313 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:1167 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:0
> >>           RX bytes:6619719 (6.3 MiB)  TX bytes:115448 (112.7 KiB)
> >>
> >> eth1      Link encap:Ethernet  HWaddr 00:14:4F:40:D8:71
> >>           inet addr:xxx.xxx.xxx.xxx  Bcast:xxx.xxx.xxx.255
> >> Mask:255.255.255.0
> >>           inet6 addr: fe80::214:4fff:fe40:d871/64 Scope:Link
> >>           UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>           RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:0
> >>           RX bytes:0 (0.0 b)  TX bytes:720 (720.0 b)
> >>
> >> peth0     Link encap:Ethernet  HWaddr FE:FF:FF:FF:FF:FF
> >>           inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>           UP BROADCAST RUNNING NOARP  MTU:1500  Metric:1
> >>           RX packets:105404 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:13455 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:1000
> >>           RX bytes:9122426 (8.6 MiB)  TX bytes:1997833 (1.9 MiB)
> >>           Base address:0x8c00 Memory:fa5e0000-fa600000
> >>
> >> peth1     Link encap:Ethernet  HWaddr FE:FF:FF:FF:FF:FE
> >>           inet6 addr: fe80::fcff:ffff:feff:fffe/64 Scope:Link
> >>           UP BROADCAST RUNNING NOARP  MTU:1500  Metric:1
> >>           RX packets:9239102 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:5337342 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:1000
> >>           RX bytes:12420212276 (11.5 GiB)  TX bytes:523849820 (499.5 MiB)
> >>           Base address:0x8800 Memory:fa5c0000-fa5e0000
> >>
> >> vif1.0    Link encap:Ethernet  HWaddr FE:FF:FF:FF:FF:FF
> >>           inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>           UP BROADCAST RUNNING NOARP  MTU:1500  Metric:1
> >>           RX packets:2174 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:80032 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:32
> >>           RX bytes:446031 (435.5 KiB)  TX bytes:6941997 (6.6 MiB)
> >>
> >> vif2.0    Link encap:Ethernet  HWaddr FE:FF:FF:FF:FF:FF
> >>           inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>           UP BROADCAST RUNNING NOARP  MTU:1500  Metric:1
> >>           RX packets:10084 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:89284 errors:0 dropped:0 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:32
> >>           RX bytes:1186096 (1.1 MiB)  TX bytes:7380350 (7.0 MiB)
> >>
> >> vif2.1    Link encap:Ethernet  HWaddr FE:FF:FF:FF:FF:FF
> >>           inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>           UP BROADCAST RUNNING NOARP  MTU:1500  Metric:1
> >>           RX packets:5337343 errors:0 dropped:0 overruns:0 frame:0
> >>           TX packets:9239082 errors:0 dropped:20 overruns:0 carrier:0
> >>           collisions:0 txqueuelen:32
> >>           RX bytes:403712612 (385.0 MiB)  TX bytes:12383227754 (11.5 GiB)
> >>
> >> Any help is greately appreciated.
> >>
> >> _________________________________________________________________
> >> http://liveearth.msn.com
> >>
> >> --
> >> Fedora-xen mailing list
> >> Fedora-xen@xxxxxxxxxx
> >> https://www.redhat.com/mailman/listinfo/fedora-xen
> >>
> >
> > _________________________________________________________________
> > Share life as it happens with the new Windows Live.
> > http://www.windowslive.com/share.html?ocid=TXT_TAGHM_Wave2_sharelife_012008
> >
> > --
> > Fedora-xen mailing list
> > Fedora-xen@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/fedora-xen
> >
>
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users
>

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

<Prev in Thread] Current Thread [Next in Thread>