WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] BUG at xen4.1/kernel 2.6.32.35 at a CentOS 5.5 when star

To: Teck Choon Giam <giamteckchoon@xxxxxxxxx>
Subject: Re: [Xen-devel] BUG at xen4.1/kernel 2.6.32.35 at a CentOS 5.5 when starting a VM
From: Gerd Jakobovitsch <gerd@xxxxxxxxxxx>
Date: Fri, 08 Apr 2011 12:06:36 -0300
Cc: xen-devel@xxxxxxxxxxxxxxxxxxx
Delivery-date: Fri, 08 Apr 2011 08:17:25 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <AANLkTimF8QkUE4G7x=p93jsVA6iCEL595=o1Z8WEfCjc@xxxxxxxxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <4D933ADB.8060106@xxxxxxxxxxx> <AANLkTimF8QkUE4G7x=p93jsVA6iCEL595=o1Z8WEfCjc@xxxxxxxxxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
User-agent: Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.14) Gecko/20110223 Lightning/1.0b2 Thunderbird/3.1.8
On 03/30/2011 11:44 PM, Teck Choon Giam wrote:
On Wed, Mar 30, 2011 at 10:14 PM, Gerd Jakobovitsch <gerd@xxxxxxxxxxx> wrote:
Hello all,

I used to run xen4.0 kernel 2.6.32.24 over CentOS 5.5, with a relative success, but the bug at mmu.c appeared once at a while. Therefore, I'm looking for a more stable option.
I compiled and ran the newly released xen 4.1, with kernel PVOPS 2.6.32.35 over CentOS 5.5. When trying to start a VM, the following bugs appeared at dmesg. After that, xl and xm commands do not longer respond:

[  145.749573]   alloc irq_desc for 2209 on node -1
[  145.749581]   alloc kstat_irqs on node -1
[  145.883515] block tda: sector-size: 512 capacity: 262144
[  145.889952] general protection fault: 0000 [#1] SMP
[  145.890109] last sysfs file: /sys/block/tda/removable
[  145.890164] CPU 7
[  145.890252] Modules linked in: bridge stp nfs fscache nfs_acl auth_rpcgss arptable_filter arp_tables xt_esp ipt_ah xt_physdev xt_multiport lockd sunrpc bonding dm_multipath bnx2 megaraid_sas
[  145.891125] Pid: 5179, comm: tapdisk2 Not tainted 2.6.32.35 #1 PowerEdge M610
[  145.891184] RIP: e030:[<ffffffff81281e79>]  [<ffffffff81281e79>] blktap_device_end_request+0x4e/0x63
[  145.891296] RSP: e02b:ffff880064061cd8  EFLAGS: 00010046
[  145.891351] RAX: 6b6b6b6b6b6b6b6b RBX: ffff88007d264690 RCX: 0000000000000028
[  145.891410] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000000000
[  145.891469] RBP: ffff880064061cf8 R08: 0000000064061c98 R09: ffff88007da42948
[  145.891528] R10: ffffea0000000008 R11: 0000000001f60080 R12: ffff88007da427f8
[  145.891587] R13: ffff88007c75f398 R14: 0000000000000000 R15: ffff88007c75f3a8
[  145.891651] FS:  00007ff33d9a4730(0000) GS:ffff8800189e5000(0000) knlGS:0000000000000000
[  145.891714] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
[  145.891771] CR2: 0000000002594cc8 CR3: 000000007be61000 CR4: 0000000000002660
[  145.891830] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[  145.891890] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[  145.892171] Process tapdisk2 (pid: 5179, threadinfo ffff880064060000, task ffff88007c272d60)
[  145.892669] Stack:
[  145.892934]  ffff88007c272d60 0000000000000000 0000000000000000 0000000000000000
[  145.893334] <0> ffff880064061e88 ffffffff812815ae ffff880064061e58 ffffffff811d234f
[  145.894035] <0> ffff88007e9bbfc0 ffff88007c75f398 00000001ffffffff 0000000000000000
[  145.895015] Call Trace:
[  145.895286]  [<ffffffff812815ae>] blktap_ring_ioctl+0x183/0x2d8
[  145.895566]  [<ffffffff811d234f>] ? inode_has_perm+0x77/0x89
[  145.895844]  [<ffffffff811d234f>] ? inode_has_perm+0x77/0x89
[  145.896124]  [<ffffffff81219e24>] ? _raw_spin_lock+0x77/0x12f
[  145.896403]  [<ffffffff81219d28>] ? _raw_spin_unlock+0xab/0xb2
[  145.896682]  [<ffffffff81529311>] ? _spin_unlock+0x9/0xb
[  145.896958]  [<ffffffff81219e24>] ? _raw_spin_lock+0x77/0x12f
[  145.897234]  [<ffffffff811d2415>] ? file_has_perm+0xb4/0xc6
[  145.897513]  [<ffffffff810fe868>] vfs_ioctl+0x5e/0x77
[  145.897786]  [<ffffffff810fed7d>] do_vfs_ioctl+0x484/0x4d5
[  145.898060]  [<ffffffff810fee25>] sys_ioctl+0x57/0x7a
[  145.898338]  [<ffffffff81013d02>] system_call_fastpath+0x16/0x1b
[  145.898614] Code: e8 5f f4 ff ff 49 8b 44 24 40 48 8b b8 80 03 00 00 e8 64 75 2a 00 41 8b 54 24 60 44 89 f6 4c 89 e7 e8 b5 89 f7 ff 49 8b 44 24 40 <48> 8b b8 80 03 00 00 e8 23 74 2a 00 5b 41 5c 41 5d 41 5e c9 c3
[  145.902008] RIP  [<ffffffff81281e79>] blktap_device_end_request+0x4e/0x63
[  145.902321]  RSP <ffff880064061cd8>
[  145.902585] ---[ end trace 2800cfa5aa85ca0a ]---
[  262.100689] BUG: spinlock lockup on CPU#4, vol_id/5181, ffff88007c75f520
[  262.100965] Pid: 5181, comm: vol_id Tainted: G      D    2.6.32.35 #1
[  262.101232] Call Trace:
[  262.101497]  [<ffffffff81219eae>] _raw_spin_lock+0x101/0x12f
[  262.101762]  [<ffffffff815293e6>] _spin_lock_irq+0x1e/0x20
[  262.102028]  [<ffffffff811fcc14>] __make_request+0x5e/0x402
[  262.102294]  [<ffffffff8101019f>] ? xen_restore_fl_direct_end+0x0/0x1
[  262.102563]  [<ffffffff811fa171>] generic_make_request+0x258/0x2f4
[  262.102832]  [<ffffffff811156d8>] ? bio_init+0x18/0x32
[  262.103099]  [<ffffffff811fbec8>] submit_bio+0xd0/0xd9
[  262.103366]  [<ffffffff81111574>] submit_bh+0xf7/0x11a
[  262.103631]  [<ffffffff8111448f>] block_read_full_page+0x246/0x264
[  262.103898]  [<ffffffff81117c13>] ? blkdev_get_block+0x0/0x4d
[  262.104165]  [<ffffffff815292c6>] ? _spin_unlock_irq+0x1e/0x20
[  262.104433]  [<ffffffff810ba73d>] ? add_to_page_cache_locked+0xa0/0xca
[  262.104702]  [<ffffffff81116ef9>] blkdev_readpage+0x13/0x15
[  262.104972]  [<ffffffff810c1d36>] __do_page_cache_readahead+0x144/0x177
[  262.105240]  [<ffffffff810c1f8f>] ondemand_readahead+0x126/0x18e
[  262.105507]  [<ffffffff810c20d7>] page_cache_sync_readahead+0x38/0x3a
[  262.105778]  [<ffffffff810bb833>] generic_file_aio_read+0x24c/0x5c1
[  262.106045]  [<ffffffff810f1808>] do_sync_read+0xe2/0x126
[  262.106315]  [<ffffffff81068a02>] ? autoremove_wake_function+0x0/0x38
[  262.106584]  [<ffffffff811d701c>] ? selinux_file_permission+0x5c/0x10e
[  262.106854]  [<ffffffff811ce9c4>] ? security_file_permission+0x11/0x13
[  262.107120]  [<ffffffff810f1f7b>] vfs_read+0xab/0x167
[  262.107385]  [<ffffffff810f2374>] sys_read+0x47/0x70
[  262.107652]  [<ffffffff81013d02>] system_call_fastpath+0x16/0x1b
[  262.107918] sending NMI to all CPUs:
[  262.108189] BUG: unable to handle kernel paging request at ffffffffff5fb310
[  262.108526] IP: [<ffffffff8102c7d1>] flat_send_IPI_mask+0x6a/0xc0
[  262.108832] PGD 1003067 PUD 1004067 PMD 18b7067 PTE 0
[  262.109235] Oops: 0002 [#2] SMP
[  262.109565] last sysfs file: /sys/class/blktap2/blktap1/dev
[  262.109830] CPU 4
[  262.110121] Modules linked in: bridge stp nfs fscache nfs_acl auth_rpcgss arptable_filter arp_tables xt_esp ipt_ah xt_physdev xt_multiport lockd sunrpc bonding dm_multipath bnx2 megaraid_sas
[  262.111520] Pid: 5181, comm: vol_id Tainted: G      D    2.6.32.35 #1 PowerEdge M610
[  262.112008] RIP: e030:[<ffffffff8102c7d1>]  [<ffffffff8102c7d1>] flat_send_IPI_mask+0x6a/0xc0
[  262.112535] RSP: e02b:ffff88006778f968  EFLAGS: 00010086
[  262.112800] RAX: 00000000ff000000 RBX: ffffffff81790060 RCX: 00000000000160a0
[  262.113068] RDX: ffff88001898e000 RSI: 0000000000000002 RDI: ffffffff81816020
[  262.113337] RBP: ffff88006778f988 R08: 0000000000000000 R09: 0000000000000004
[  262.113605] R10: 0000000000000002 R11: 0000000000000004 R12: 0000000000000002
[  262.113877] R13: 0000000000000800 R14: 00000000000000ff R15: 0000000000000000
[  262.114149] FS:  00007fa78bcc5710(0063) GS:ffff88001898e000(0000) knlGS:0000000000000000
[  262.114636] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
[  262.114902] CR2: ffffffffff5fb310 CR3: 00000000641b4000 CR4: 0000000000002660
[  262.115171] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[  262.115438] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[  262.115707] Process vol_id (pid: 5181, threadinfo ffff88006778e000, task ffff88007db86250)
[  262.116194] Stack:
[  262.116451]  0000000000000000 0000000076e9ecd0 0000000000000000 0000000076e9ecd0
[  262.116825] <0> ffff88006778f998 ffffffff8102c841 ffff88006778f9b8 ffffffff81029f0d
[  262.117485] <0> ffff88007c75f520 ffff88007c75f520 ffff88006778f9f8 ffffffff81219eb3
[  262.118396] Call Trace:
[  262.118657]  [<ffffffff8102c841>] flat_send_IPI_all+0x1a/0x56
[  262.118925]  [<ffffffff81029f0d>] arch_trigger_all_cpu_backtrace+0x45/0x66
[  262.119195]  [<ffffffff81219eb3>] _raw_spin_lock+0x106/0x12f
[  262.119463]  [<ffffffff815293e6>] _spin_lock_irq+0x1e/0x20
[  262.119730]  [<ffffffff811fcc14>] __make_request+0x5e/0x402
[  262.119996]  [<ffffffff8101019f>] ? xen_restore_fl_direct_end+0x0/0x1
[  262.120264]  [<ffffffff811fa171>] generic_make_request+0x258/0x2f4
[  262.120532]  [<ffffffff811156d8>] ? bio_init+0x18/0x32
[  262.120799]  [<ffffffff811fbec8>] submit_bio+0xd0/0xd9
[  262.121066]  [<ffffffff81111574>] submit_bh+0xf7/0x11a
[  262.121333]  [<ffffffff8111448f>] block_read_full_page+0x246/0x264
[  262.121602]  [<ffffffff81117c13>] ? blkdev_get_block+0x0/0x4d
[  262.121870]  [<ffffffff815292c6>] ? _spin_unlock_irq+0x1e/0x20
[  262.122137]  [<ffffffff810ba73d>] ? add_to_page_cache_locked+0xa0/0xca
[  262.127766]  [<ffffffff81116ef9>] blkdev_readpage+0x13/0x15
[  262.128025]  [<ffffffff810c1d36>] __do_page_cache_readahead+0x144/0x177
[  262.128288]  [<ffffffff810c1f8f>] ondemand_readahead+0x126/0x18e
[  262.128548]  [<ffffffff810c20d7>] page_cache_sync_readahead+0x38/0x3a
[  262.128810]  [<ffffffff810bb833>] generic_file_aio_read+0x24c/0x5c1
[  262.129070]  [<ffffffff810f1808>] do_sync_read+0xe2/0x126
[  262.129329]  [<ffffffff81068a02>] ? autoremove_wake_function+0x0/0x38
[  262.129590]  [<ffffffff811d701c>] ? selinux_file_permission+0x5c/0x10e
[  262.129851]  [<ffffffff811ce9c4>] ? security_file_permission+0x11/0x13
[  262.130110]  [<ffffffff810f1f7b>] vfs_read+0xab/0x167
[  262.130368]  [<ffffffff810f2374>] sys_read+0x47/0x70
[  262.130624]  [<ffffffff81013d02>] system_call_fastpath+0x16/0x1b
[  262.130883] Code: 8b 05 b4 95 7e 00 83 fe 02 44 8b 68 34 75 0a ff 90 58 01 00 00 eb 0e f3 90 8b 04 25 00 b3 5f ff f6 c4 10 75 f2 44 89 f0 c1 e0 18 <89> 04 25 10 b3 5f ff 41 83 fc 02 74 08 44 89 e0 44 09 e8 eb 06
[  262.133866] RIP  [<ffffffff8102c7d1>] flat_send_IPI_mask+0x6a/0xc0
[  262.134164]  RSP <ffff88006778f968>
[  262.134419] CR2: ffffffffff5fb310
[  262.134673] ---[ end trace 2800cfa5aa85ca0b ]---

Can you try to recompile your PVOPS kernel with CONFIG_DEBUG_PAGEALLOC=y?

You can read more about this BUG at
http://lists.xensource.com/archives/html/xen-devel/2011-03/msg01756.html

I initially hit this BUG sometime Dec 2010... ...
http://lists.xensource.com/archives/html/xen-devel/2010-12/msg01501.html

Thanks.

Kindest regards,
Giam Teck Choon
Sorry for the delayed answer. The problem I'm facing now is not related to the mmu bug - that one I still am seeing at systems with xen 4.0.2 / kernel 2.6.32.24. Newer kernels have new bugs, much more troublesome, since I can not run a single VM instance.

Adding DEBUG_PAGEALLOC, the main difference is that the system is rebooting shortly after trying to start up a VM:

Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: I/O queue driver: lio
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: received 'attach' message (uuid = 0)
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: sending 'attach response' message (uuid = 0)
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: received 'open' message (uuid = 0)
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: block-aio open('/storage5/linux-centos-5-64b-base-7253/hda')
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: open(/storage5/linux-centos-5-64b-base-7253/hda) with O_DIRECT
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: Image size:      pre sector_shift  [134217728]     post sector_shift [262144]
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: opened image /storage5/linux-centos-5-64b-base-rip/hda (1 users, state: 0x00000001, type: 0)
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: VBD CHAIN:
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: /storage5/linux-centos-5-64b-base/hda: 0
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15563]: sending 'open response' message (uuid = 0)
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.644887] block tda: sector-size: 512 capacity: 262144
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657328] general protection fault: 0000 [#1] SMP DEBUG_PAGEALLOC
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657379] last sysfs file: /sys/block/tda/removable
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657400] CPU 0
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657421] Modules linked in: nfs fscache nfs_acl auth_rpcgss bridge stp ocfs2 ocfs2_dlmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager ocfs2_stackglue configfs arptable_filter arp_tables xt_esp ipt_ah xt_physdev xt_multiport dm_round_robin lockd sunrpc crc32c bonding iscsi_tcp libiscsi_tcp bnx2i libiscsi scsi_transport_iscsi cnic uio dm_multipath bnx2 megaraid_sas
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657736] Pid: 15566, comm: tapdisk2 Not tainted 2.6.32.36 #4 PowerEdge M610
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657763] RIP: e030:[<ffffffff8129fb89>]  [<ffffffff8129fb89>] blktap_device_end_request+0x4e/0x63
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657808] RSP: e02b:ffff88006da5dcd8  EFLAGS: 00010046
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657833] RAX: 6b6b6b6b6b6b6b6b RBX: ffff88006566c000 RCX: 0000000000000000
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657860] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff88006d8c7980
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.657887] RBP: ffff88006da5dcf8 R08: ffffffff817e66a0 R09: ffff88007d775790
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.658136] R10: ffffffff810ccfe4 R11: ffff8800280d8f60 R12: ffff88006720e7f8
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.658384] R13: ffff88006d8c77e0 R14: 0000000000000000 R15: ffff88006d8c77f0
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.658635] FS:  00007f86d75a3730(0000) GS:ffff8800280c7000(0000) knlGS:0000000000000000
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.659107] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.659351] CR2: 0000000045614ed8 CR3: 000000006d911000 CR4: 0000000000002660
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.659600] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.659847] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.660100] Process tapdisk2 (pid: 15566, threadinfo ffff88006da5c000, task ffff88007d7750f0)
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.660574] Stack:
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.660807]  ffff88007d7750f0 0000000000000000 0000000000000000 0000000000000000
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.661074] <0> ffff88006da5de88 ffffffff8129f2c0 ffffffff8100fedd 000000016da5ddc8
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.661574] <0> 00000000ffffffff ffff88006d8c77e0 00000001816fab27 0000000000000000
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.662287] Call Trace:
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.662522]  [<ffffffff8129f2c0>] blktap_ring_ioctl+0x183/0x2d8
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.662767]  [<ffffffff8100fedd>] ? xen_force_evtchn_callback+0xd/0xf
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.663012]  [<ffffffff811ebe9b>] ? inode_has_perm+0xa1/0xb3
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.663260]  [<ffffffff8101064f>] ? xen_restore_fl_direct_end+0x0/0x1
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.663509]  [<ffffffff81081d84>] ? lock_release+0x1b8/0x1c3
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.663756]  [<ffffffff81233d2c>] ? _raw_spin_unlock+0xab/0xb2
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.663999]  [<ffffffff8155618d>] ? _spin_unlock+0x26/0x2a
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.664248]  [<ffffffff81134aa1>] ? aio_read_evt+0x87/0x13a
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.664493]  [<ffffffff81134b36>] ? aio_read_evt+0x11c/0x13a
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.664740]  [<ffffffff81233ecc>] ? _raw_spin_lock+0x77/0x12f
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.664986]  [<ffffffff811ec054>] ? file_has_perm+0xb4/0xc6
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.665237]  [<ffffffff81112444>] vfs_ioctl+0x5e/0x77
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.665480]  [<ffffffff81112959>] do_vfs_ioctl+0x484/0x4d5
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.665726]  [<ffffffff81112a01>] sys_ioctl+0x57/0x7a
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.665970]  [<ffffffff81013d42>] system_call_fastpath+0x16/0x1b
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.666214] Code: e8 61 f4 ff ff 49 8b 44 24 40 48 8b b8 70 04 00 00 e8 79 67 2b 00 41 8b 54 24 60 44 89 f6 4c 89 e7 e8 76 3b f7 ff 49 8b 44 24 40 <48> 8b b8 70 04 00 00 e8 1d 65 2b 00 5b 41 5c 41 5d 41 5e c9 c3
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.667320] RIP  [<ffffffff8129fb89>] blktap_device_end_request+0x4e/0x63
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.667577]  RSP <ffff88006da5dcd8>
Apr  8 12:00:43 r2b16ch2x28p2 kernel: [ 8879.668069] ---[ end trace da218b929afc63f7 ]---
Apr  8 12:00:43 r2b16ch2x28p2 tapdisk2[15584]: I/O queue driver: lio
Apr  8 12:00:48 r2b16ch2x28p2 tap-ctl: tap-err:tap_ctl_read_message: failure reading message
Apr  8 12:00:48 r2b16ch2x28p2 tap-ctl: tap-err:tap_ctl_send_and_receive: failed to receive 'unknown' message

--

Gerd Jakobovitsch
Engenheiro de Produto

---------------------------------------------------------

ALOG Data Centers do Brasil
Excelência em Projetos de Hosting
Rua Dr. Miguel Couto, 58 – 01008-010 – São Paulo - SP
Telefone: (11) 3524-4970 / (11) 7152-0815

http://www.alog.com.br


"Como estão nossos serviços? Clique aqui e nos conte. Queremos escutar a sua opinião!"





_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel