WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-ppc-devel

[XenPPC] [xenppc-unstable] [IA64] VTI: simple format cleanup

To: xen-ppc-devel@xxxxxxxxxxxxxxxxxxx
Subject: [XenPPC] [xenppc-unstable] [IA64] VTI: simple format cleanup
From: Xen patchbot-xenppc-unstable <patchbot-xenppc-unstable@xxxxxxxxxxxxxxxxxxx>
Date: Fri, 02 Jun 2006 17:56:31 +0000
Delivery-date: Fri, 02 Jun 2006 10:58:18 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-ppc-devel-request@lists.xensource.com?subject=help>
List-id: Xen PPC development <xen-ppc-devel.lists.xensource.com>
List-post: <mailto:xen-ppc-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-ppc-devel>, <mailto:xen-ppc-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-ppc-devel>, <mailto:xen-ppc-devel-request@lists.xensource.com?subject=unsubscribe>
Reply-to: xen-ppc-devel@xxxxxxxxxxxxxxxxxxx
Sender: xen-ppc-devel-bounces@xxxxxxxxxxxxxxxxxxx
# HG changeset patch
# User awilliam@xxxxxxxxxxx
# Node ID 18b087bafac6197716b7b2290ddb1c7e656916fe
# Parent  c073ebdbde8c0f5c9437706b46c4a34f35033c0c
[IA64] VTI: simple format cleanup

Signed-off-by: Anthony Xu <anthony.xu@xxxxxxxxx>
---
 xen/arch/ia64/vmx/vmmu.c        |   28 -
 xen/arch/ia64/vmx/vmx_ivt.S     |  592 ++++++++++++++++++++--------------------
 xen/arch/ia64/vmx/vmx_process.c |   76 ++---
 xen/include/asm-ia64/vmmu.h     |   59 +--
 xen/include/asm-ia64/vmx_vcpu.h |    4 
 5 files changed, 376 insertions(+), 383 deletions(-)

diff -r c073ebdbde8c -r 18b087bafac6 xen/arch/ia64/vmx/vmmu.c
--- a/xen/arch/ia64/vmx/vmmu.c  Fri May 26 13:41:49 2006 -0600
+++ b/xen/arch/ia64/vmx/vmmu.c  Tue May 30 08:46:21 2006 -0600
@@ -319,17 +319,17 @@ fetch_code(VCPU *vcpu, u64 gip, u64 *cod
 //        if( tlb == NULL )
 //             tlb = vtlb_lookup(vcpu, gip, DSIDE_TLB );
         if (tlb)
-               gpip = (tlb->ppn >>(tlb->ps-12)<<tlb->ps) | ( gip & 
(PSIZE(tlb->ps)-1) );
+            gpip = (tlb->ppn >>(tlb->ps-12)<<tlb->ps) | ( gip & 
(PSIZE(tlb->ps)-1) );
     }
     if( gpip){
-        mfn = gmfn_to_mfn(vcpu->domain, gpip >>PAGE_SHIFT);
-       if( mfn == INVALID_MFN )  panic_domain(vcpu_regs(vcpu),"fetch_code: 
invalid memory\n");
-       vpa =(u64 *)__va( (gip & (PAGE_SIZE-1)) | (mfn<<PAGE_SHIFT));
+        mfn = gmfn_to_mfn(vcpu->domain, gpip >>PAGE_SHIFT);
+        if( mfn == INVALID_MFN )  panic_domain(vcpu_regs(vcpu),"fetch_code: 
invalid memory\n");
+        vpa =(u64 *)__va( (gip & (PAGE_SIZE-1)) | (mfn<<PAGE_SHIFT));
     }else{
-       tlb = vhpt_lookup(gip);
-       if( tlb == NULL)
-           panic_domain(vcpu_regs(vcpu),"No entry found in ITLB and DTLB\n");
-       vpa =(u64 
*)__va((tlb->ppn>>(PAGE_SHIFT-ARCH_PAGE_SHIFT)<<PAGE_SHIFT)|(gip&(PAGE_SIZE-1)));
+        tlb = vhpt_lookup(gip);
+        if( tlb == NULL)
+            panic_domain(vcpu_regs(vcpu),"No entry found in ITLB and DTLB\n");
+        vpa =(u64 
*)__va((tlb->ppn>>(PAGE_SHIFT-ARCH_PAGE_SHIFT)<<PAGE_SHIFT)|(gip&(PAGE_SIZE-1)));
     }
     *code1 = *vpa++;
     *code2 = *vpa;
@@ -530,7 +530,7 @@ IA64FAULT vmx_vcpu_tpa(VCPU *vcpu, UINT6
     visr.ir=pt_isr.ir;
     vpsr.val = vmx_vcpu_get_psr(vcpu);
     if(vpsr.ic==0){
-         visr.ni=1;
+        visr.ni=1;
     }
     visr.na=1;
     data = vtlb_lookup(vcpu, vadr, DSIDE_TLB);
@@ -648,14 +648,14 @@ long
 long
 __domain_va_to_ma(unsigned long va, unsigned long* ma, unsigned long *len)
 {
-    unsigned long      mpfn, gpfn, m, n = *len;
-    unsigned long      end;    /* end of the area mapped by current entry */
-    thash_data_t       *entry;
+    unsigned long  mpfn, gpfn, m, n = *len;
+    unsigned long  end;   /* end of the area mapped by current entry */
+    thash_data_t   *entry;
     struct vcpu *v = current;
 
     entry = vtlb_lookup(v, va, DSIDE_TLB);
     if (entry == NULL)
-       return -EFAULT;
+        return -EFAULT;
 
     gpfn =(entry->ppn>>(PAGE_SHIFT-12));
     gpfn =PAGEALIGN(gpfn,(entry->ps-PAGE_SHIFT));
@@ -668,7 +668,7 @@ __domain_va_to_ma(unsigned long va, unsi
     /*end = PAGEALIGN(m, entry->ps) + PSIZE(entry->ps);*/
     /* Current entry can't map all requested area */
     if ((m + n) > end)
-       n = end - m;
+        n = end - m;
 
     *ma = m;
     *len = n;
diff -r c073ebdbde8c -r 18b087bafac6 xen/arch/ia64/vmx/vmx_ivt.S
--- a/xen/arch/ia64/vmx/vmx_ivt.S       Fri May 26 13:41:49 2006 -0600
+++ b/xen/arch/ia64/vmx/vmx_ivt.S       Tue May 30 08:46:21 2006 -0600
@@ -2,10 +2,10 @@
  * arch/ia64/kernel/vmx_ivt.S
  *
  * Copyright (C) 1998-2001, 2003 Hewlett-Packard Co
- *     Stephane Eranian <eranian@xxxxxxxxxx>
- *     David Mosberger <davidm@xxxxxxxxxx>
+ *      Stephane Eranian <eranian@xxxxxxxxxx>
+ *      David Mosberger <davidm@xxxxxxxxxx>
  * Copyright (C) 2000, 2002-2003 Intel Co
- *     Asit Mallick <asit.k.mallick@xxxxxxxxx>
+ *      Asit Mallick <asit.k.mallick@xxxxxxxxx>
  *      Suresh Siddha <suresh.b.siddha@xxxxxxxxx>
  *      Kenneth Chen <kenneth.w.chen@xxxxxxxxx>
  *      Fenghua Yu <fenghua.yu@xxxxxxxxx>
@@ -31,7 +31,7 @@
  *
  *  For each entry, the comment is as follows:
  *
- *             // 0x1c00 Entry 7 (size 64 bundles) Data Key Miss (12,51)
+ *              // 0x1c00 Entry 7 (size 64 bundles) Data Key Miss (12,51)
  *  entry offset ----/     /         /                  /          /
  *  entry number ---------/         /                  /          /
  *  size of the entry -------------/                  /          /
@@ -96,13 +96,13 @@ vmx_fault_##n:;          \
     ;;                  \
 
 
-#define VMX_REFLECT(n)                         \
-       mov r31=pr;                                                             
        \
-       mov r19=n;                      /* prepare to save predicates */        
        \
-    mov r29=cr.ipsr;        \
+#define VMX_REFLECT(n)    \
+    mov r31=pr;           \
+    mov r19=n;       /* prepare to save predicates */ \
+    mov r29=cr.ipsr;      \
     ;;      \
     tbit.z p6,p7=r29,IA64_PSR_VM_BIT;       \
-(p7) br.sptk.many vmx_dispatch_reflection;        \
+(p7)br.sptk.many vmx_dispatch_reflection;        \
     VMX_FAULT(n);            \
 
 
@@ -115,10 +115,10 @@ END(vmx_panic)
 
 
 
-       .section .text.ivt,"ax"
-
-       .align 32768    // align on 32KB boundary
-       .global vmx_ia64_ivt
+    .section .text.ivt,"ax"
+
+    .align 32768    // align on 32KB boundary
+    .global vmx_ia64_ivt
 vmx_ia64_ivt:
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x0000 Entry 0 (size 64 bundles) VHPT Translation (8,20,47)
@@ -127,7 +127,7 @@ ENTRY(vmx_vhpt_miss)
     VMX_FAULT(0)
 END(vmx_vhpt_miss)
 
-       .org vmx_ia64_ivt+0x400
+    .org vmx_ia64_ivt+0x400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x0400 Entry 1 (size 64 bundles) ITLB (21)
 ENTRY(vmx_itlb_miss)
@@ -410,52 +410,52 @@ ENTRY(vmx_nested_dtlb_miss)
     VMX_FAULT(5)
 END(vmx_nested_dtlb_miss)
 
-       .org vmx_ia64_ivt+0x1800
+    .org vmx_ia64_ivt+0x1800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x1800 Entry 6 (size 64 bundles) Instruction Key Miss (24)
 ENTRY(vmx_ikey_miss)
     VMX_DBG_FAULT(6)
-       VMX_REFLECT(6)
+    VMX_REFLECT(6)
 END(vmx_ikey_miss)
 
-       .org vmx_ia64_ivt+0x1c00
+    .org vmx_ia64_ivt+0x1c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x1c00 Entry 7 (size 64 bundles) Data Key Miss (12,51)
 ENTRY(vmx_dkey_miss)
     VMX_DBG_FAULT(7)
-       VMX_REFLECT(7)
+    VMX_REFLECT(7)
 END(vmx_dkey_miss)
 
-       .org vmx_ia64_ivt+0x2000
+    .org vmx_ia64_ivt+0x2000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x2000 Entry 8 (size 64 bundles) Dirty-bit (54)
 ENTRY(vmx_dirty_bit)
     VMX_DBG_FAULT(8)
-       VMX_REFLECT(8)
+    VMX_REFLECT(8)
 END(vmx_dirty_bit)
 
-       .org vmx_ia64_ivt+0x2400
+    .org vmx_ia64_ivt+0x2400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x2400 Entry 9 (size 64 bundles) Instruction Access-bit (27)
 ENTRY(vmx_iaccess_bit)
     VMX_DBG_FAULT(9)
-       VMX_REFLECT(9)
+    VMX_REFLECT(9)
 END(vmx_iaccess_bit)
 
-       .org vmx_ia64_ivt+0x2800
+    .org vmx_ia64_ivt+0x2800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x2800 Entry 10 (size 64 bundles) Data Access-bit (15,55)
 ENTRY(vmx_daccess_bit)
     VMX_DBG_FAULT(10)
-       VMX_REFLECT(10)
+    VMX_REFLECT(10)
 END(vmx_daccess_bit)
 
-       .org vmx_ia64_ivt+0x2c00
+    .org vmx_ia64_ivt+0x2c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x2c00 Entry 11 (size 64 bundles) Break instruction (33)
 ENTRY(vmx_break_fault)
     VMX_DBG_FAULT(11)
-       mov r31=pr
+    mov r31=pr
     mov r19=11
     mov r30=cr.iim
     movl r29=0x1100
@@ -473,12 +473,12 @@ ENTRY(vmx_break_fault)
     VMX_FAULT(11);
 END(vmx_break_fault)
 
-       .org vmx_ia64_ivt+0x3000
+    .org vmx_ia64_ivt+0x3000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x3000 Entry 12 (size 64 bundles) External Interrupt (4)
 ENTRY(vmx_interrupt)
 //    VMX_DBG_FAULT(12)
-       mov r31=pr              // prepare to save predicates
+    mov r31=pr         // prepare to save predicates
     mov r19=12
     mov r29=cr.ipsr
     ;;
@@ -487,58 +487,58 @@ ENTRY(vmx_interrupt)
     ;;
 (p7) br.sptk vmx_dispatch_interrupt
     ;;
-       mov r27=ar.rsc                  /* M */
-       mov r20=r1                      /* A */
-       mov r25=ar.unat         /* M */
-       mov r26=ar.pfs                  /* I */
-       mov r28=cr.iip                  /* M */
-       cover               /* B (or nothing) */
-       ;;
-       mov r1=sp
-       ;;
-       invala                          /* M */
-       mov r30=cr.ifs
-       ;;
+    mov r27=ar.rsc             /* M */
+    mov r20=r1                 /* A */
+    mov r25=ar.unat            /* M */
+    mov r26=ar.pfs             /* I */
+    mov r28=cr.iip             /* M */
+    cover                      /* B (or nothing) */
+    ;;
+    mov r1=sp
+    ;;
+    invala                     /* M */
+    mov r30=cr.ifs
+    ;;
     addl r1=-IA64_PT_REGS_SIZE,r1
     ;;
-       adds r17=2*L1_CACHE_BYTES,r1            /* really: biggest cache-line 
size */
-       adds r16=PT(CR_IPSR),r1
-       ;;
-       lfetch.fault.excl.nt1 [r17],L1_CACHE_BYTES
-       st8 [r16]=r29           /* save cr.ipsr */
-       ;;
-       lfetch.fault.excl.nt1 [r17]
-       mov r29=b0
-       ;;
-       adds r16=PT(R8),r1      /* initialize first base pointer */
-       adds r17=PT(R9),r1      /* initialize second base pointer */
-       mov r18=r0                      /* make sure r18 isn't NaT */
-       ;;
+    adds r17=2*L1_CACHE_BYTES,r1       /* really: biggest cache-line size */
+    adds r16=PT(CR_IPSR),r1
+    ;;
+    lfetch.fault.excl.nt1 [r17],L1_CACHE_BYTES
+    st8 [r16]=r29                      /* save cr.ipsr */
+    ;;
+    lfetch.fault.excl.nt1 [r17]
+    mov r29=b0
+    ;;
+    adds r16=PT(R8),r1         /* initialize first base pointer */
+    adds r17=PT(R9),r1         /* initialize second base pointer */
+    mov r18=r0                 /* make sure r18 isn't NaT */
+    ;;
 .mem.offset 0,0; st8.spill [r16]=r8,16
 .mem.offset 8,0; st8.spill [r17]=r9,16
         ;;
 .mem.offset 0,0; st8.spill [r16]=r10,24
 .mem.offset 8,0; st8.spill [r17]=r11,24
         ;;
-       st8 [r16]=r28,16        /* save cr.iip */
-       st8 [r17]=r30,16        /* save cr.ifs */
-       mov r8=ar.fpsr          /* M */
-       mov r9=ar.csd
-       mov r10=ar.ssd
-       movl r11=FPSR_DEFAULT   /* L-unit */
-       ;;
-       st8 [r16]=r25,16        /* save ar.unat */
-       st8 [r17]=r26,16        /* save ar.pfs */
-       shl r18=r18,16          /* compute ar.rsc to be used for "loadrs" */
-       ;;
-    st8 [r16]=r27,16   /* save ar.rsc */
-    adds r17=16,r17    /* skip over ar_rnat field */
-    ;;          /* avoid RAW on r16 & r17 */
-    st8 [r17]=r31,16   /* save predicates */
-    adds r16=16,r16    /* skip over ar_bspstore field */
-    ;;
-    st8 [r16]=r29,16   /* save b0 */
-    st8 [r17]=r18,16   /* save ar.rsc value for "loadrs" */
+    st8 [r16]=r28,16           /* save cr.iip */
+    st8 [r17]=r30,16           /* save cr.ifs */
+    mov r8=ar.fpsr             /* M */
+    mov r9=ar.csd
+    mov r10=ar.ssd
+    movl r11=FPSR_DEFAULT      /* L-unit */
+    ;;
+    st8 [r16]=r25,16           /* save ar.unat */
+    st8 [r17]=r26,16           /* save ar.pfs */
+    shl r18=r18,16             /* compute ar.rsc to be used for "loadrs" */
+    ;;
+    st8 [r16]=r27,16           /* save ar.rsc */
+    adds r17=16,r17            /* skip over ar_rnat field */
+    ;;
+    st8 [r17]=r31,16           /* save predicates */
+    adds r16=16,r16            /* skip over ar_bspstore field */
+    ;;
+    st8 [r16]=r29,16           /* save b0 */
+    st8 [r17]=r18,16           /* save ar.rsc value for "loadrs" */
     ;;
 .mem.offset 0,0; st8.spill [r16]=r20,16    /* save original r1 */
 .mem.offset 8,0; st8.spill [r17]=r12,16
@@ -561,18 +561,18 @@ ENTRY(vmx_interrupt)
     ;;                                          \
     bsw.1
     ;;
-       alloc r14=ar.pfs,0,0,2,0 // must be first in an insn group
-       mov out0=cr.ivr         // pass cr.ivr as first arg
-       add out1=16,sp          // pass pointer to pt_regs as second arg
-
-       ssm psr.ic
+    alloc r14=ar.pfs,0,0,2,0   // must be first in an insn group
+    mov out0=cr.ivr            // pass cr.ivr as first arg
+    add out1=16,sp             // pass pointer to pt_regs as second arg
+
+    ssm psr.ic
     ;;
     srlz.i
-       ;;
+    ;;
     (p15) ssm psr.i
-       adds r3=8,r2            // set up second base pointer for SAVE_REST
-       srlz.i                  // ensure everybody knows psr.ic is back on
-       ;;
+    adds r3=8,r2               // set up second base pointer for SAVE_REST
+    srlz.i                     // ensure everybody knows psr.ic is back on
+    ;;
 .mem.offset 0,0; st8.spill [r2]=r16,16
 .mem.offset 8,0; st8.spill [r3]=r17,16
     ;;
@@ -599,8 +599,8 @@ ENTRY(vmx_interrupt)
 .mem.offset 0,0; st8.spill [r2]=r30,16
 .mem.offset 8,0; st8.spill [r3]=r31,32
     ;;
-    mov ar.fpsr=r11     /* M-unit */
-    st8 [r2]=r8,8      /* ar.ccv */
+    mov ar.fpsr=r11       /* M-unit */
+    st8 [r2]=r8,8         /* ar.ccv */
     adds r24=PT(B6)-PT(F7),r3
     ;;
     stf.spill [r2]=f6,32
@@ -619,95 +619,95 @@ ENTRY(vmx_interrupt)
     st8 [r24]=r9           /* ar.csd */
     st8 [r25]=r10          /* ar.ssd */
     ;;
-       srlz.d                  // make sure we see the effect of cr.ivr
-       movl r14=ia64_leave_nested
-       ;;
-       mov rp=r14
-       br.call.sptk.many b6=ia64_handle_irq
-       ;;
+    srlz.d             // make sure we see the effect of cr.ivr
+    movl r14=ia64_leave_nested
+    ;;
+    mov rp=r14
+    br.call.sptk.many b6=ia64_handle_irq
+    ;;
 END(vmx_interrupt)
 
-       .org vmx_ia64_ivt+0x3400
+    .org vmx_ia64_ivt+0x3400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x3400 Entry 13 (size 64 bundles) Reserved
 ENTRY(vmx_virtual_exirq)
-       VMX_DBG_FAULT(13)
-       mov r31=pr
-        mov r19=13
-        br.sptk vmx_dispatch_vexirq
+    VMX_DBG_FAULT(13)
+    mov r31=pr
+    mov r19=13
+    br.sptk vmx_dispatch_vexirq
 END(vmx_virtual_exirq)
 
-       .org vmx_ia64_ivt+0x3800
+    .org vmx_ia64_ivt+0x3800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x3800 Entry 14 (size 64 bundles) Reserved
     VMX_DBG_FAULT(14)
-       VMX_FAULT(14)
-
-
-       .org vmx_ia64_ivt+0x3c00
+    VMX_FAULT(14)
+
+
+    .org vmx_ia64_ivt+0x3c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x3c00 Entry 15 (size 64 bundles) Reserved
     VMX_DBG_FAULT(15)
-       VMX_FAULT(15)
-
-
-       .org vmx_ia64_ivt+0x4000
+    VMX_FAULT(15)
+
+
+    .org vmx_ia64_ivt+0x4000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x4000 Entry 16 (size 64 bundles) Reserved
     VMX_DBG_FAULT(16)
-       VMX_FAULT(16)
-
-       .org vmx_ia64_ivt+0x4400
+    VMX_FAULT(16)
+
+    .org vmx_ia64_ivt+0x4400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x4400 Entry 17 (size 64 bundles) Reserved
     VMX_DBG_FAULT(17)
-       VMX_FAULT(17)
-
-       .org vmx_ia64_ivt+0x4800
+    VMX_FAULT(17)
+
+    .org vmx_ia64_ivt+0x4800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x4800 Entry 18 (size 64 bundles) Reserved
     VMX_DBG_FAULT(18)
-       VMX_FAULT(18)
-
-       .org vmx_ia64_ivt+0x4c00
+    VMX_FAULT(18)
+
+    .org vmx_ia64_ivt+0x4c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x4c00 Entry 19 (size 64 bundles) Reserved
     VMX_DBG_FAULT(19)
-       VMX_FAULT(19)
+    VMX_FAULT(19)
 
     .org vmx_ia64_ivt+0x5000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5000 Entry 20 (size 16 bundles) Page Not Present
 ENTRY(vmx_page_not_present)
-       VMX_DBG_FAULT(20)
-       VMX_REFLECT(20)
+    VMX_DBG_FAULT(20)
+    VMX_REFLECT(20)
 END(vmx_page_not_present)
 
     .org vmx_ia64_ivt+0x5100
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5100 Entry 21 (size 16 bundles) Key Permission vector
 ENTRY(vmx_key_permission)
-       VMX_DBG_FAULT(21)
-       VMX_REFLECT(21)
+    VMX_DBG_FAULT(21)
+    VMX_REFLECT(21)
 END(vmx_key_permission)
 
     .org vmx_ia64_ivt+0x5200
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5200 Entry 22 (size 16 bundles) Instruction Access Rights (26)
 ENTRY(vmx_iaccess_rights)
-       VMX_DBG_FAULT(22)
-       VMX_REFLECT(22)
+    VMX_DBG_FAULT(22)
+    VMX_REFLECT(22)
 END(vmx_iaccess_rights)
 
-       .org vmx_ia64_ivt+0x5300
+    .org vmx_ia64_ivt+0x5300
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5300 Entry 23 (size 16 bundles) Data Access Rights (14,53)
 ENTRY(vmx_daccess_rights)
-       VMX_DBG_FAULT(23)
-       VMX_REFLECT(23)
+    VMX_DBG_FAULT(23)
+    VMX_REFLECT(23)
 END(vmx_daccess_rights)
 
-       .org vmx_ia64_ivt+0x5400
+    .org vmx_ia64_ivt+0x5400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5400 Entry 24 (size 16 bundles) General Exception (5,32,34,36,38,39)
 ENTRY(vmx_general_exception)
@@ -716,106 +716,106 @@ ENTRY(vmx_general_exception)
 //    VMX_FAULT(24)
 END(vmx_general_exception)
 
-       .org vmx_ia64_ivt+0x5500
+    .org vmx_ia64_ivt+0x5500
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5500 Entry 25 (size 16 bundles) Disabled FP-Register (35)
 ENTRY(vmx_disabled_fp_reg)
-       VMX_DBG_FAULT(25)
-       VMX_REFLECT(25)
+    VMX_DBG_FAULT(25)
+    VMX_REFLECT(25)
 END(vmx_disabled_fp_reg)
 
-       .org vmx_ia64_ivt+0x5600
+    .org vmx_ia64_ivt+0x5600
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5600 Entry 26 (size 16 bundles) Nat Consumption (11,23,37,50)
 ENTRY(vmx_nat_consumption)
-       VMX_DBG_FAULT(26)
-       VMX_REFLECT(26)
+    VMX_DBG_FAULT(26)
+    VMX_REFLECT(26)
 END(vmx_nat_consumption)
 
-       .org vmx_ia64_ivt+0x5700
+    .org vmx_ia64_ivt+0x5700
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5700 Entry 27 (size 16 bundles) Speculation (40)
 ENTRY(vmx_speculation_vector)
-       VMX_DBG_FAULT(27)
-       VMX_REFLECT(27)
+    VMX_DBG_FAULT(27)
+    VMX_REFLECT(27)
 END(vmx_speculation_vector)
 
-       .org vmx_ia64_ivt+0x5800
+    .org vmx_ia64_ivt+0x5800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5800 Entry 28 (size 16 bundles) Reserved
     VMX_DBG_FAULT(28)
-       VMX_FAULT(28)
-
-       .org vmx_ia64_ivt+0x5900
+    VMX_FAULT(28)
+
+    .org vmx_ia64_ivt+0x5900
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5900 Entry 29 (size 16 bundles) Debug (16,28,56)
 ENTRY(vmx_debug_vector)
     VMX_DBG_FAULT(29)
-       VMX_FAULT(29)
+    VMX_FAULT(29)
 END(vmx_debug_vector)
 
-       .org vmx_ia64_ivt+0x5a00
+    .org vmx_ia64_ivt+0x5a00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5a00 Entry 30 (size 16 bundles) Unaligned Reference (57)
 ENTRY(vmx_unaligned_access)
-       VMX_DBG_FAULT(30)
-       VMX_REFLECT(30)
+    VMX_DBG_FAULT(30)
+    VMX_REFLECT(30)
 END(vmx_unaligned_access)
 
-       .org vmx_ia64_ivt+0x5b00
+    .org vmx_ia64_ivt+0x5b00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5b00 Entry 31 (size 16 bundles) Unsupported Data Reference (57)
 ENTRY(vmx_unsupported_data_reference)
-       VMX_DBG_FAULT(31)
-       VMX_REFLECT(31)
+    VMX_DBG_FAULT(31)
+    VMX_REFLECT(31)
 END(vmx_unsupported_data_reference)
 
-       .org vmx_ia64_ivt+0x5c00
+    .org vmx_ia64_ivt+0x5c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5c00 Entry 32 (size 16 bundles) Floating-Point Fault (64)
 ENTRY(vmx_floating_point_fault)
-       VMX_DBG_FAULT(32)
-       VMX_REFLECT(32)
+    VMX_DBG_FAULT(32)
+    VMX_REFLECT(32)
 END(vmx_floating_point_fault)
 
-       .org vmx_ia64_ivt+0x5d00
+    .org vmx_ia64_ivt+0x5d00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5d00 Entry 33 (size 16 bundles) Floating Point Trap (66)
 ENTRY(vmx_floating_point_trap)
-       VMX_DBG_FAULT(33)
-       VMX_REFLECT(33)
+    VMX_DBG_FAULT(33)
+    VMX_REFLECT(33)
 END(vmx_floating_point_trap)
 
-       .org vmx_ia64_ivt+0x5e00
+    .org vmx_ia64_ivt+0x5e00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5e00 Entry 34 (size 16 bundles) Lower Privilege Transfer Trap (66)
 ENTRY(vmx_lower_privilege_trap)
-       VMX_DBG_FAULT(34)
-       VMX_REFLECT(34)
+    VMX_DBG_FAULT(34)
+    VMX_REFLECT(34)
 END(vmx_lower_privilege_trap)
 
-       .org vmx_ia64_ivt+0x5f00
+    .org vmx_ia64_ivt+0x5f00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x5f00 Entry 35 (size 16 bundles) Taken Branch Trap (68)
 ENTRY(vmx_taken_branch_trap)
-       VMX_DBG_FAULT(35)
-       VMX_REFLECT(35)
+    VMX_DBG_FAULT(35)
+    VMX_REFLECT(35)
 END(vmx_taken_branch_trap)
 
-       .org vmx_ia64_ivt+0x6000
+    .org vmx_ia64_ivt+0x6000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6000 Entry 36 (size 16 bundles) Single Step Trap (69)
 ENTRY(vmx_single_step_trap)
-       VMX_DBG_FAULT(36)
-       VMX_REFLECT(36)
+    VMX_DBG_FAULT(36)
+    VMX_REFLECT(36)
 END(vmx_single_step_trap)
 
-       .org vmx_ia64_ivt+0x6100
+    .org vmx_ia64_ivt+0x6100
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6100 Entry 37 (size 16 bundles) Virtualization Fault
 ENTRY(vmx_virtualization_fault)
 //    VMX_DBG_FAULT(37)
-       mov r31=pr
+    mov r31=pr
     mov r19=37
     adds r16 = IA64_VCPU_CAUSE_OFFSET,r21
     adds r17 = IA64_VCPU_OPCODE_OFFSET,r21
@@ -826,197 +826,197 @@ ENTRY(vmx_virtualization_fault)
     br.sptk vmx_dispatch_virtualization_fault
 END(vmx_virtualization_fault)
 
-       .org vmx_ia64_ivt+0x6200
+    .org vmx_ia64_ivt+0x6200
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6200 Entry 38 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(38)
-       VMX_FAULT(38)
-
-       .org vmx_ia64_ivt+0x6300
+    VMX_DBG_FAULT(38)
+    VMX_FAULT(38)
+
+    .org vmx_ia64_ivt+0x6300
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6300 Entry 39 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(39)
-       VMX_FAULT(39)
-
-       .org vmx_ia64_ivt+0x6400
+    VMX_DBG_FAULT(39)
+    VMX_FAULT(39)
+
+    .org vmx_ia64_ivt+0x6400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6400 Entry 40 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(40)
-       VMX_FAULT(40)
-
-       .org vmx_ia64_ivt+0x6500
+    VMX_DBG_FAULT(40)
+    VMX_FAULT(40)
+
+    .org vmx_ia64_ivt+0x6500
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6500 Entry 41 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(41)
-       VMX_FAULT(41)
-
-       .org vmx_ia64_ivt+0x6600
+    VMX_DBG_FAULT(41)
+    VMX_FAULT(41)
+
+    .org vmx_ia64_ivt+0x6600
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6600 Entry 42 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(42)
-       VMX_FAULT(42)
-
-       .org vmx_ia64_ivt+0x6700
+    VMX_DBG_FAULT(42)
+    VMX_FAULT(42)
+
+    .org vmx_ia64_ivt+0x6700
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6700 Entry 43 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(43)
-       VMX_FAULT(43)
-
-       .org vmx_ia64_ivt+0x6800
+    VMX_DBG_FAULT(43)
+    VMX_FAULT(43)
+
+    .org vmx_ia64_ivt+0x6800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6800 Entry 44 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(44)
-       VMX_FAULT(44)
-
-       .org vmx_ia64_ivt+0x6900
+    VMX_DBG_FAULT(44)
+    VMX_FAULT(44)
+
+    .org vmx_ia64_ivt+0x6900
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6900 Entry 45 (size 16 bundles) IA-32 Exeception 
(17,18,29,41,42,43,44,58,60,61,62,72,73,75,76,77)
 ENTRY(vmx_ia32_exception)
-       VMX_DBG_FAULT(45)
-       VMX_FAULT(45)
+    VMX_DBG_FAULT(45)
+    VMX_FAULT(45)
 END(vmx_ia32_exception)
 
-       .org vmx_ia64_ivt+0x6a00
+    .org vmx_ia64_ivt+0x6a00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6a00 Entry 46 (size 16 bundles) IA-32 Intercept  (30,31,59,70,71)
 ENTRY(vmx_ia32_intercept)
-       VMX_DBG_FAULT(46)
-       VMX_FAULT(46)
+    VMX_DBG_FAULT(46)
+    VMX_FAULT(46)
 END(vmx_ia32_intercept)
 
-       .org vmx_ia64_ivt+0x6b00
+    .org vmx_ia64_ivt+0x6b00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6b00 Entry 47 (size 16 bundles) IA-32 Interrupt  (74)
 ENTRY(vmx_ia32_interrupt)
-       VMX_DBG_FAULT(47)
-       VMX_FAULT(47)
+    VMX_DBG_FAULT(47)
+    VMX_FAULT(47)
 END(vmx_ia32_interrupt)
 
-       .org vmx_ia64_ivt+0x6c00
+    .org vmx_ia64_ivt+0x6c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6c00 Entry 48 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(48)
-       VMX_FAULT(48)
-
-       .org vmx_ia64_ivt+0x6d00
+    VMX_DBG_FAULT(48)
+    VMX_FAULT(48)
+
+    .org vmx_ia64_ivt+0x6d00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6d00 Entry 49 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(49)
-       VMX_FAULT(49)
-
-       .org vmx_ia64_ivt+0x6e00
+    VMX_DBG_FAULT(49)
+    VMX_FAULT(49)
+
+    .org vmx_ia64_ivt+0x6e00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6e00 Entry 50 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(50)
-       VMX_FAULT(50)
-
-       .org vmx_ia64_ivt+0x6f00
+    VMX_DBG_FAULT(50)
+    VMX_FAULT(50)
+
+    .org vmx_ia64_ivt+0x6f00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x6f00 Entry 51 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(51)
-       VMX_FAULT(51)
-
-       .org vmx_ia64_ivt+0x7000
+    VMX_DBG_FAULT(51)
+    VMX_FAULT(51)
+
+    .org vmx_ia64_ivt+0x7000
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7000 Entry 52 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(52)
-       VMX_FAULT(52)
-
-       .org vmx_ia64_ivt+0x7100
+    VMX_DBG_FAULT(52)
+    VMX_FAULT(52)
+
+    .org vmx_ia64_ivt+0x7100
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7100 Entry 53 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(53)
-       VMX_FAULT(53)
-
-       .org vmx_ia64_ivt+0x7200
+    VMX_DBG_FAULT(53)
+    VMX_FAULT(53)
+
+    .org vmx_ia64_ivt+0x7200
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7200 Entry 54 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(54)
-       VMX_FAULT(54)
-
-       .org vmx_ia64_ivt+0x7300
+    VMX_DBG_FAULT(54)
+    VMX_FAULT(54)
+
+    .org vmx_ia64_ivt+0x7300
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7300 Entry 55 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(55)
-       VMX_FAULT(55)
-
-       .org vmx_ia64_ivt+0x7400
+    VMX_DBG_FAULT(55)
+    VMX_FAULT(55)
+
+    .org vmx_ia64_ivt+0x7400
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7400 Entry 56 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(56)
-       VMX_FAULT(56)
-
-       .org vmx_ia64_ivt+0x7500
+    VMX_DBG_FAULT(56)
+    VMX_FAULT(56)
+
+    .org vmx_ia64_ivt+0x7500
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7500 Entry 57 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(57)
-       VMX_FAULT(57)
-
-       .org vmx_ia64_ivt+0x7600
+    VMX_DBG_FAULT(57)
+    VMX_FAULT(57)
+
+    .org vmx_ia64_ivt+0x7600
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7600 Entry 58 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(58)
-       VMX_FAULT(58)
-
-       .org vmx_ia64_ivt+0x7700
+    VMX_DBG_FAULT(58)
+    VMX_FAULT(58)
+
+    .org vmx_ia64_ivt+0x7700
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7700 Entry 59 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(59)
-       VMX_FAULT(59)
-
-       .org vmx_ia64_ivt+0x7800
+    VMX_DBG_FAULT(59)
+    VMX_FAULT(59)
+
+    .org vmx_ia64_ivt+0x7800
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7800 Entry 60 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(60)
-       VMX_FAULT(60)
-
-       .org vmx_ia64_ivt+0x7900
+    VMX_DBG_FAULT(60)
+    VMX_FAULT(60)
+
+    .org vmx_ia64_ivt+0x7900
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7900 Entry 61 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(61)
-       VMX_FAULT(61)
-
-       .org vmx_ia64_ivt+0x7a00
+    VMX_DBG_FAULT(61)
+    VMX_FAULT(61)
+
+    .org vmx_ia64_ivt+0x7a00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7a00 Entry 62 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(62)
-       VMX_FAULT(62)
-
-       .org vmx_ia64_ivt+0x7b00
+    VMX_DBG_FAULT(62)
+    VMX_FAULT(62)
+
+    .org vmx_ia64_ivt+0x7b00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7b00 Entry 63 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(63)
-       VMX_FAULT(63)
-
-       .org vmx_ia64_ivt+0x7c00
+    VMX_DBG_FAULT(63)
+    VMX_FAULT(63)
+
+    .org vmx_ia64_ivt+0x7c00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7c00 Entry 64 (size 16 bundles) Reserved
     VMX_DBG_FAULT(64)
-       VMX_FAULT(64)
-
-       .org vmx_ia64_ivt+0x7d00
+    VMX_FAULT(64)
+
+    .org vmx_ia64_ivt+0x7d00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7d00 Entry 65 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(65)
-       VMX_FAULT(65)
-
-       .org vmx_ia64_ivt+0x7e00
+    VMX_DBG_FAULT(65)
+    VMX_FAULT(65)
+
+    .org vmx_ia64_ivt+0x7e00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7e00 Entry 66 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(66)
-       VMX_FAULT(66)
-
-       .org vmx_ia64_ivt+0x7f00
+    VMX_DBG_FAULT(66)
+    VMX_FAULT(66)
+
+    .org vmx_ia64_ivt+0x7f00
 
/////////////////////////////////////////////////////////////////////////////////////////
 // 0x7f00 Entry 67 (size 16 bundles) Reserved
-       VMX_DBG_FAULT(67)
-       VMX_FAULT(67)
-
-       .org vmx_ia64_ivt+0x8000
-    // There is no particular reason for this code to be here, other than that
-    // there happens to be space here that would go unused otherwise.  If this
-    // fault ever gets "unreserved", simply moved the following code to a more
-    // suitable spot...
+    VMX_DBG_FAULT(67)
+    VMX_FAULT(67)
+
+    .org vmx_ia64_ivt+0x8000
+// There is no particular reason for this code to be here, other than that
+// there happens to be space here that would go unused otherwise.  If this
+// fault ever gets "unreserved", simply moved the following code to a more
+// suitable spot...
 
 
 ENTRY(vmx_dispatch_reflection)
@@ -1165,24 +1165,24 @@ END(vmx_hypercall_dispatch)
 
 
 ENTRY(vmx_dispatch_interrupt)
-       VMX_SAVE_MIN_WITH_COVER_R19     // uses r31; defines r2 and r3
-       ;;
-       alloc r14=ar.pfs,0,0,2,0 // must be first in an insn group
-       mov out0=cr.ivr         // pass cr.ivr as first arg
-       adds r3=8,r2            // set up second base pointer for SAVE_REST
-    ;;
-       ssm psr.ic
-       ;;
+    VMX_SAVE_MIN_WITH_COVER_R19        // uses r31; defines r2 and r3
+    ;;
+    alloc r14=ar.pfs,0,0,2,0 // must be first in an insn group
+    mov out0=cr.ivr            // pass cr.ivr as first arg
+    adds r3=8,r2               // set up second base pointer for SAVE_REST
+    ;;
+    ssm psr.ic
+    ;;
     srlz.i
     ;;
     (p15) ssm psr.i
-       movl r14=ia64_leave_hypervisor
-       ;;
-       VMX_SAVE_REST
-       mov rp=r14
-       ;;
-       add out1=16,sp          // pass pointer to pt_regs as second arg
-       br.call.sptk.many b6=ia64_handle_irq
+    movl r14=ia64_leave_hypervisor
+    ;;
+    VMX_SAVE_REST
+    mov rp=r14
+    ;;
+    add out1=16,sp             // pass pointer to pt_regs as second arg
+    br.call.sptk.many b6=ia64_handle_irq
 END(vmx_dispatch_interrupt)
 
 
diff -r c073ebdbde8c -r 18b087bafac6 xen/arch/ia64/vmx/vmx_process.c
--- a/xen/arch/ia64/vmx/vmx_process.c   Fri May 26 13:41:49 2006 -0600
+++ b/xen/arch/ia64/vmx/vmx_process.c   Tue May 30 08:46:21 2006 -0600
@@ -183,12 +183,12 @@ vmx_ia64_handle_break (unsigned long ifa
     struct vcpu *v = current;
 
 #ifdef CRASH_DEBUG
-       if ((iim == 0 || iim == CDB_BREAK_NUM) && !user_mode(regs) &&
+    if ((iim == 0 || iim == CDB_BREAK_NUM) && !user_mode(regs) &&
         IS_VMM_ADDRESS(regs->cr_iip)) {
-               if (iim == 0)
-                       show_registers(regs);
-               debugger_trap_fatal(0 /* don't care */, regs);
-       } else
+        if (iim == 0)
+            show_registers(regs);
+        debugger_trap_fatal(0 /* don't care */, regs);
+    } else
 #endif
     {
         if (iim == 0) 
@@ -247,45 +247,45 @@ void save_banked_regs_to_vpd(VCPU *v, RE
 // NEVER successful if already reflecting a trap/fault because psr.i==0
 void leave_hypervisor_tail(struct pt_regs *regs)
 {
-       struct domain *d = current->domain;
-       struct vcpu *v = current;
-       // FIXME: Will this work properly if doing an RFI???
-       if (!is_idle_domain(d) ) {      // always comes from guest
-               extern void vmx_dorfirfi(void);
-               struct pt_regs *user_regs = vcpu_regs(current);
-               if (local_softirq_pending())
-                       do_softirq();
-               local_irq_disable();
- 
-               if (user_regs != regs)
-                       printk("WARNING: checking pending interrupt in nested 
interrupt!!!\n");
-
-               /* VMX Domain N has other interrupt source, saying DM  */
-                if (test_bit(ARCH_VMX_INTR_ASSIST, &v->arch.arch_vmx.flags))
+    struct domain *d = current->domain;
+    struct vcpu *v = current;
+    // FIXME: Will this work properly if doing an RFI???
+    if (!is_idle_domain(d) ) { // always comes from guest
+        extern void vmx_dorfirfi(void);
+        struct pt_regs *user_regs = vcpu_regs(current);
+        if (local_softirq_pending())
+            do_softirq();
+        local_irq_disable();
+
+        if (user_regs != regs)
+            printk("WARNING: checking pending interrupt in nested 
interrupt!!!\n");
+
+        /* VMX Domain N has other interrupt source, saying DM  */
+        if (test_bit(ARCH_VMX_INTR_ASSIST, &v->arch.arch_vmx.flags))
                       vmx_intr_assist(v);
 
-               /* FIXME: Check event pending indicator, and set
-                * pending bit if necessary to inject back to guest.
-                * Should be careful about window between this check
-                * and above assist, since IOPACKET_PORT shouldn't be
-                * injected into vmx domain.
-                *
-                * Now hardcode the vector as 0x10 temporarily
-                */
-//             if (event_pending(v)&&(!(VLSAPIC_INSVC(v,0)&(1UL<<0x10)))) {
-//                     VCPU(v, irr[0]) |= 1UL << 0x10;
-//                     v->arch.irq_new_pending = 1;
-//             }
-
-               if ( v->arch.irq_new_pending ) {
-                       v->arch.irq_new_pending = 0;
-                       vmx_check_pending_irq(v);
-               }
+        /* FIXME: Check event pending indicator, and set
+         * pending bit if necessary to inject back to guest.
+         * Should be careful about window between this check
+         * and above assist, since IOPACKET_PORT shouldn't be
+         * injected into vmx domain.
+         *
+         * Now hardcode the vector as 0x10 temporarily
+         */
+//       if (event_pending(v)&&(!(VLSAPIC_INSVC(v,0)&(1UL<<0x10)))) {
+//           VCPU(v, irr[0]) |= 1UL << 0x10;
+//           v->arch.irq_new_pending = 1;
+//       }
+
+        if ( v->arch.irq_new_pending ) {
+            v->arch.irq_new_pending = 0;
+            vmx_check_pending_irq(v);
+        }
 //        if (VCPU(v,vac).a_bsw){
 //            save_banked_regs_to_vpd(v,regs);
 //        }
 
-       }
+    }
 }
 
 extern ia64_rr vmx_vcpu_rr(VCPU *vcpu,UINT64 vadr);
diff -r c073ebdbde8c -r 18b087bafac6 xen/include/asm-ia64/vmmu.h
--- a/xen/include/asm-ia64/vmmu.h       Fri May 26 13:41:49 2006 -0600
+++ b/xen/include/asm-ia64/vmmu.h       Tue May 30 08:46:21 2006 -0600
@@ -23,15 +23,15 @@
 #ifndef XEN_TLBthash_H
 #define XEN_TLBthash_H
 
-#define         MAX_CCN_DEPTH           15       // collision chain depth
-#define         VCPU_VTLB_SHIFT          (20)    // 1M for VTLB
-#define         VCPU_VTLB_SIZE           (1UL<<VCPU_VTLB_SHIFT)
-#define         VCPU_VTLB_ORDER          (VCPU_VTLB_SHIFT - PAGE_SHIFT)
-#define         VCPU_VHPT_SHIFT          (24)    // 16M for VTLB
-#define         VCPU_VHPT_SIZE           (1UL<<VCPU_VHPT_SHIFT)
-#define         VCPU_VHPT_ORDER          (VCPU_VHPT_SHIFT - PAGE_SHIFT)
-#define                VTLB(v,_x)              (v->arch.vtlb._x)
-#define                VHPT(v,_x)              (v->arch.vhpt._x)
+#define     MAX_CCN_DEPTH       (15)       // collision chain depth
+#define     VCPU_VTLB_SHIFT     (20)    // 1M for VTLB
+#define     VCPU_VTLB_SIZE      (1UL<<VCPU_VTLB_SHIFT)
+#define     VCPU_VTLB_ORDER     (VCPU_VTLB_SHIFT - PAGE_SHIFT)
+#define     VCPU_VHPT_SHIFT     (24)    // 16M for VTLB
+#define     VCPU_VHPT_SIZE      (1UL<<VCPU_VHPT_SHIFT)
+#define     VCPU_VHPT_ORDER     (VCPU_VHPT_SHIFT - PAGE_SHIFT)
+#define     VTLB(v,_x)          (v->arch.vtlb._x)
+#define     VHPT(v,_x)          (v->arch.vhpt._x)
 #ifndef __ASSEMBLY__
 
 #include <xen/config.h>
@@ -60,18 +60,18 @@
 #define LOW_32BITS(x)   bits(x,0,31)
 
 typedef union search_section {
-        struct {
-                u32 tr : 1;
-                u32 tc : 1;
-                u32 rsv: 30;
-        };
-        u32     v;
+    struct {
+        u32 tr : 1;
+        u32 tc : 1;
+        u32 rsv: 30;
+    };
+    u32     v;
 } search_section_t;
 
 
 enum {
-        ISIDE_TLB=0,
-        DSIDE_TLB=1
+    ISIDE_TLB=0,
+    DSIDE_TLB=1
 };
 #define VTLB_PTE_P_BIT      0
 #define VTLB_PTE_IO_BIT     60
@@ -93,16 +93,15 @@ typedef struct thash_data {
             u64 ig1  :  3; // 53-63
         };
         struct {
-            u64 __rv1 : 53;    // 0-52
+            u64 __rv1 : 53;     // 0-52
             u64 contiguous : 1; //53
-            u64 tc : 1;     // 54 TR or TC
-            u64 cl : 1; // 55 I side or D side cache line
-            // next extension to ig1, only for TLB instance
-            u64 len  :  4; // 56-59
+            u64 tc : 1;         // 54 TR or TC
+            u64 cl : 1;         // 55 I side or D side cache line
+            u64 len  :  4;      // 56-59
             u64 io  : 1;       // 60 entry is for io or not
-            u64 nomap : 1;   // 61 entry cann't be inserted into machine TLB.
-            u64 checked : 1; // 62 for VTLB/VHPT sanity check
-            u64 invalid : 1; // 63 invalid entry
+            u64 nomap : 1;      // 61 entry cann't be inserted into machine 
TLB.
+            u64 checked : 1;    // 62 for VTLB/VHPT sanity check
+            u64 invalid : 1;    // 63 invalid entry
         };
         u64 page_flags;
     };                  // same for VHPT and TLB
@@ -114,12 +113,6 @@ typedef struct thash_data {
             u64 key  : 24; // 8-31
             u64 rv4  : 32; // 32-63
         };
-//        struct {
-//            u64 __rv3  : 32; // 0-31
-            // next extension to rv4
-//            u64 rid  : 24;  // 32-55
-//            u64 __rv4  : 8; // 56-63
-//        };
         u64 itir;
     };
     union {
@@ -176,8 +169,8 @@ static inline u64 xen_to_arch_ppn(u64 xp
 }
 
 typedef enum {
-        THASH_TLB=0,
-        THASH_VHPT
+    THASH_TLB=0,
+    THASH_VHPT
 } THASH_TYPE;
 
 struct thash_cb;
diff -r c073ebdbde8c -r 18b087bafac6 xen/include/asm-ia64/vmx_vcpu.h
--- a/xen/include/asm-ia64/vmx_vcpu.h   Fri May 26 13:41:49 2006 -0600
+++ b/xen/include/asm-ia64/vmx_vcpu.h   Tue May 30 08:46:21 2006 -0600
@@ -44,7 +44,6 @@
 #define VRN7    0x7UL
 // for vlsapic
 #define  VLSAPIC_INSVC(vcpu, i) ((vcpu)->arch.insvc[i])
-//#define      VMX_VPD(x,y)    ((x)->arch.arch_vmx.vpd->y)
 
 #define VMX(x,y)  ((x)->arch.arch_vmx.y)
 
@@ -228,7 +227,8 @@ IA64FAULT vmx_vcpu_get_lrr0(VCPU *vcpu, 
 }
 static inline
 IA64FAULT vmx_vcpu_get_lrr1(VCPU *vcpu, UINT64 *pval)
-{    *pval = VCPU(vcpu,lrr1);
+{
+    *pval = VCPU(vcpu,lrr1);
     return (IA64_NO_FAULT);
 }
 static inline

_______________________________________________
Xen-ppc-devel mailing list
Xen-ppc-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-ppc-devel

<Prev in Thread] Current Thread [Next in Thread>
  • [XenPPC] [xenppc-unstable] [IA64] VTI: simple format cleanup, Xen patchbot-xenppc-unstable <=