WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-changelog

[Xen-changelog] [xen-unstable] domctl/sysctl: Clean up definitions

To: xen-changelog@xxxxxxxxxxxxxxxxxxx
Subject: [Xen-changelog] [xen-unstable] domctl/sysctl: Clean up definitions
From: Xen patchbot-unstable <patchbot-unstable@xxxxxxxxxxxxxxxxxxx>
Date: Tue, 22 Dec 2009 03:40:12 -0800
Delivery-date: Tue, 22 Dec 2009 03:41:10 -0800
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-changelog-request@lists.xensource.com?subject=help>
List-id: BK change log <xen-changelog.lists.xensource.com>
List-post: <mailto:xen-changelog@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-changelog>, <mailto:xen-changelog-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-changelog>, <mailto:xen-changelog-request@lists.xensource.com?subject=unsubscribe>
Reply-to: xen-devel@xxxxxxxxxxxxxxxxxxx
Sender: xen-changelog-bounces@xxxxxxxxxxxxxxxxxxx
# HG changeset patch
# User Keir Fraser <keir.fraser@xxxxxxxxxx>
# Date 1261481595 0
# Node ID bde24590c13a65a55b3bb240d47765cfd935fbc6
# Parent  3c93fffa650221e7f86d23337f5b5f2094e54e2f
domctl/sysctl: Clean up definitions
 - Use fixed-width types only
 - Use named unions only
 - Bump domctl version number

Signed-off-by: Keir Fraser <keir.fraser@xxxxxxxxxx>
---
 tools/libxc/xc_memshr.c       |   20 +++----
 tools/libxc/xc_pm.c           |   30 +++++------
 xen/arch/x86/mm/mem_sharing.c |   20 +++----
 xen/drivers/acpi/pmstat.c     |  108 +++++++++++++++++++++---------------------
 xen/include/public/domctl.h   |   64 ++++++++++++------------
 xen/include/public/sysctl.h   |    2 
 6 files changed, 123 insertions(+), 121 deletions(-)

diff -r 3c93fffa6502 -r bde24590c13a tools/libxc/xc_memshr.c
--- a/tools/libxc/xc_memshr.c   Mon Dec 21 16:51:40 2009 +0000
+++ b/tools/libxc/xc_memshr.c   Tue Dec 22 11:33:15 2009 +0000
@@ -37,7 +37,7 @@ int xc_memshr_control(int xc_handle,
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_CONTROL;
-    op->enable = enable;
+    op->u.enable = enable;
 
     return do_domctl(xc_handle, &domctl);
 }
@@ -56,10 +56,10 @@ int xc_memshr_nominate_gfn(int xc_handle
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_NOMINATE_GFN;
-    op->nominate.gfn = gfn;
+    op->u.nominate.u.gfn = gfn;
 
     ret = do_domctl(xc_handle, &domctl);
-    if(!ret) *handle = op->nominate.handle; 
+    if(!ret) *handle = op->u.nominate.handle; 
 
     return ret;
 }
@@ -78,10 +78,10 @@ int xc_memshr_nominate_gref(int xc_handl
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_NOMINATE_GREF;
-    op->nominate.grant_ref = gref;
+    op->u.nominate.u.grant_ref = gref;
 
     ret = do_domctl(xc_handle, &domctl);
-    if(!ret) *handle = op->nominate.handle; 
+    if(!ret) *handle = op->u.nominate.handle; 
 
     return ret;
 }
@@ -98,8 +98,8 @@ int xc_memshr_share(int xc_handle,
     domctl.domain = 0;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_SHARE;
-    op->share.source_handle = source_handle;
-    op->share.client_handle = client_handle;
+    op->u.share.source_handle = source_handle;
+    op->u.share.client_handle = client_handle;
 
     return do_domctl(xc_handle, &domctl);
 }
@@ -131,7 +131,7 @@ int xc_memshr_debug_gfn(int xc_handle,
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_DEBUG_GFN;
-    op->debug.gfn = gfn;
+    op->u.debug.u.gfn = gfn;
 
     return do_domctl(xc_handle, &domctl);
 }
@@ -148,7 +148,7 @@ int xc_memshr_debug_mfn(int xc_handle,
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_DEBUG_MFN;
-    op->debug.mfn = mfn;
+    op->u.debug.u.mfn = mfn;
 
     return do_domctl(xc_handle, &domctl);
 }
@@ -165,7 +165,7 @@ int xc_memshr_debug_gref(int xc_handle,
     domctl.domain = (domid_t)domid;
     op = &(domctl.u.mem_sharing_op);
     op->op = XEN_DOMCTL_MEM_SHARING_OP_DEBUG_GREF;
-    op->debug.gref = gref;
+    op->u.debug.u.gref = gref;
 
     return do_domctl(xc_handle, &domctl);
 }
diff -r 3c93fffa6502 -r bde24590c13a tools/libxc/xc_pm.c
--- a/tools/libxc/xc_pm.c       Mon Dec 21 16:51:40 2009 +0000
+++ b/tools/libxc/xc_pm.c       Tue Dec 22 11:33:15 2009 +0000
@@ -184,7 +184,7 @@ int xc_get_cpufreq_para(int xc_handle, i
 {
     DECLARE_SYSCTL;
     int ret = 0;
-    struct xen_get_cpufreq_para *sys_para = &sysctl.u.pm_op.get_para;
+    struct xen_get_cpufreq_para *sys_para = &sysctl.u.pm_op.u.get_para;
     bool has_num = user_para->cpu_num &&
                      user_para->freq_num &&
                      user_para->gov_num;
@@ -276,7 +276,7 @@ int xc_set_cpufreq_gov(int xc_handle, in
 int xc_set_cpufreq_gov(int xc_handle, int cpuid, char *govname)
 {
     DECLARE_SYSCTL;
-    char *scaling_governor = sysctl.u.pm_op.set_gov.scaling_governor;
+    char *scaling_governor = sysctl.u.pm_op.u.set_gov.scaling_governor;
 
     if ( (xc_handle < 0) || (!govname) )
         return -EINVAL;
@@ -301,8 +301,8 @@ int xc_set_cpufreq_para(int xc_handle, i
     sysctl.cmd = XEN_SYSCTL_pm_op;
     sysctl.u.pm_op.cmd = SET_CPUFREQ_PARA;
     sysctl.u.pm_op.cpuid = cpuid;
-    sysctl.u.pm_op.set_para.ctrl_type = ctrl_type;
-    sysctl.u.pm_op.set_para.ctrl_value = ctrl_value;
+    sysctl.u.pm_op.u.set_para.ctrl_type = ctrl_type;
+    sysctl.u.pm_op.u.set_para.ctrl_value = ctrl_value;
 
     return xc_sysctl(xc_handle, &sysctl);
 }
@@ -320,7 +320,7 @@ int xc_get_cpufreq_avgfreq(int xc_handle
     sysctl.u.pm_op.cpuid = cpuid;
     ret = xc_sysctl(xc_handle, &sysctl);
 
-    *avg_freq = sysctl.u.pm_op.get_avgfreq;
+    *avg_freq = sysctl.u.pm_op.u.get_avgfreq;
 
     return ret;
 }
@@ -333,14 +333,14 @@ int xc_get_cputopo(int xc_handle, struct
     sysctl.cmd = XEN_SYSCTL_pm_op;
     sysctl.u.pm_op.cmd = XEN_SYSCTL_pm_op_get_cputopo;
     sysctl.u.pm_op.cpuid = 0;
-    set_xen_guest_handle( sysctl.u.pm_op.get_topo.cpu_to_core,
+    set_xen_guest_handle( sysctl.u.pm_op.u.get_topo.cpu_to_core,
                          info->cpu_to_core );
-    set_xen_guest_handle( sysctl.u.pm_op.get_topo.cpu_to_socket,
+    set_xen_guest_handle( sysctl.u.pm_op.u.get_topo.cpu_to_socket,
                          info->cpu_to_socket );
-    sysctl.u.pm_op.get_topo.max_cpus = info->max_cpus;
+    sysctl.u.pm_op.u.get_topo.max_cpus = info->max_cpus;
 
     rc = do_sysctl(xc_handle, &sysctl);
-    info->nr_cpus = sysctl.u.pm_op.get_topo.nr_cpus;
+    info->nr_cpus = sysctl.u.pm_op.u.get_topo.nr_cpus;
 
     return rc;
 }
@@ -356,7 +356,7 @@ int xc_set_sched_opt_smt(int xc_handle, 
    sysctl.cmd = XEN_SYSCTL_pm_op;
    sysctl.u.pm_op.cmd = XEN_SYSCTL_pm_op_set_sched_opt_smt;
    sysctl.u.pm_op.cpuid = 0;
-   sysctl.u.pm_op.set_sched_opt_smt = value;
+   sysctl.u.pm_op.u.set_sched_opt_smt = value;
    rc = do_sysctl(xc_handle, &sysctl);
 
    return rc;
@@ -370,7 +370,7 @@ int xc_set_vcpu_migration_delay(int xc_h
    sysctl.cmd = XEN_SYSCTL_pm_op;
    sysctl.u.pm_op.cmd = XEN_SYSCTL_pm_op_set_vcpu_migration_delay;
    sysctl.u.pm_op.cpuid = 0;
-   sysctl.u.pm_op.set_vcpu_migration_delay = value;
+   sysctl.u.pm_op.u.set_vcpu_migration_delay = value;
    rc = do_sysctl(xc_handle, &sysctl);
 
    return rc;
@@ -387,7 +387,7 @@ int xc_get_vcpu_migration_delay(int xc_h
    rc = do_sysctl(xc_handle, &sysctl);
 
    if (!rc && value)
-       *value = sysctl.u.pm_op.get_vcpu_migration_delay;
+       *value = sysctl.u.pm_op.u.get_vcpu_migration_delay;
 
    return rc;
 }
@@ -403,9 +403,9 @@ int xc_get_cpuidle_max_cstate(int xc_han
     sysctl.cmd = XEN_SYSCTL_pm_op;
     sysctl.u.pm_op.cmd = XEN_SYSCTL_pm_op_get_max_cstate;
     sysctl.u.pm_op.cpuid = 0;
-    sysctl.u.pm_op.get_max_cstate = 0;
+    sysctl.u.pm_op.u.get_max_cstate = 0;
     rc = do_sysctl(xc_handle, &sysctl);
-    *value = sysctl.u.pm_op.get_max_cstate;
+    *value = sysctl.u.pm_op.u.get_max_cstate;
 
     return rc;
 }
@@ -420,7 +420,7 @@ int xc_set_cpuidle_max_cstate(int xc_han
     sysctl.cmd = XEN_SYSCTL_pm_op;
     sysctl.u.pm_op.cmd = XEN_SYSCTL_pm_op_set_max_cstate;
     sysctl.u.pm_op.cpuid = 0;
-    sysctl.u.pm_op.set_max_cstate = value;
+    sysctl.u.pm_op.u.set_max_cstate = value;
 
     return do_sysctl(xc_handle, &sysctl);
 }
diff -r 3c93fffa6502 -r bde24590c13a xen/arch/x86/mm/mem_sharing.c
--- a/xen/arch/x86/mm/mem_sharing.c     Mon Dec 21 16:51:40 2009 +0000
+++ b/xen/arch/x86/mm/mem_sharing.c     Tue Dec 22 11:33:15 2009 +0000
@@ -734,7 +734,7 @@ int mem_sharing_domctl(struct domain *d,
             rc = 0;
             if(!hap_enabled(d))
                 return -EINVAL;
-            d->arch.hvm_domain.mem_sharing_enabled = mec->enable;
+            d->arch.hvm_domain.mem_sharing_enabled = mec->u.enable;
             mem_sharing_audit();
             return 0; 
         }
@@ -742,19 +742,19 @@ int mem_sharing_domctl(struct domain *d,
 
         case XEN_DOMCTL_MEM_SHARING_OP_NOMINATE_GFN:
         {
-            unsigned long gfn = mec->nominate.gfn;
+            unsigned long gfn = mec->u.nominate.u.gfn;
             shr_handle_t handle;
             if(!mem_sharing_enabled(d))
                 return -EINVAL;
             rc = mem_sharing_nominate_page(d, gfn, 0, &handle);
-            mec->nominate.handle = handle;
+            mec->u.nominate.handle = handle;
             mem_sharing_audit();
         }
         break;
 
         case XEN_DOMCTL_MEM_SHARING_OP_NOMINATE_GREF:
         {
-            grant_ref_t gref = mec->nominate.grant_ref;
+            grant_ref_t gref = mec->u.nominate.u.grant_ref;
             unsigned long gfn;
             shr_handle_t handle;
 
@@ -763,15 +763,15 @@ int mem_sharing_domctl(struct domain *d,
             if(mem_sharing_gref_to_gfn(d, gref, &gfn) < 0)
                 return -EINVAL;
             rc = mem_sharing_nominate_page(d, gfn, 3, &handle);
-            mec->nominate.handle = handle;
+            mec->u.nominate.handle = handle;
             mem_sharing_audit();
         }
         break;
 
         case XEN_DOMCTL_MEM_SHARING_OP_SHARE:
         {
-            shr_handle_t sh = mec->share.source_handle;
-            shr_handle_t ch = mec->share.client_handle;
+            shr_handle_t sh = mec->u.share.source_handle;
+            shr_handle_t ch = mec->u.share.client_handle;
             rc = mem_sharing_share_pages(sh, ch); 
             mem_sharing_audit();
         }
@@ -788,7 +788,7 @@ int mem_sharing_domctl(struct domain *d,
 
         case XEN_DOMCTL_MEM_SHARING_OP_DEBUG_GFN:
         {
-            unsigned long gfn = mec->debug.gfn;
+            unsigned long gfn = mec->u.debug.u.gfn;
             rc = mem_sharing_debug_gfn(d, gfn);
             mem_sharing_audit();
         }
@@ -796,7 +796,7 @@ int mem_sharing_domctl(struct domain *d,
 
         case XEN_DOMCTL_MEM_SHARING_OP_DEBUG_MFN:
         {
-            unsigned long mfn = mec->debug.mfn;
+            unsigned long mfn = mec->u.debug.u.mfn;
             rc = mem_sharing_debug_mfn(mfn);
             mem_sharing_audit();
         }
@@ -804,7 +804,7 @@ int mem_sharing_domctl(struct domain *d,
 
         case XEN_DOMCTL_MEM_SHARING_OP_DEBUG_GREF:
         {
-            grant_ref_t gref = mec->debug.gref;
+            grant_ref_t gref = mec->u.debug.u.gref;
             rc = mem_sharing_debug_gref(d, gref);
             mem_sharing_audit();
         }
diff -r 3c93fffa6502 -r bde24590c13a xen/drivers/acpi/pmstat.c
--- a/xen/drivers/acpi/pmstat.c Mon Dec 21 16:51:40 2009 +0000
+++ b/xen/drivers/acpi/pmstat.c Tue Dec 22 11:33:15 2009 +0000
@@ -212,37 +212,37 @@ static int get_cpufreq_para(struct xen_s
     list_for_each(pos, &cpufreq_governor_list)
         gov_num++;
 
-    if ( (op->get_para.cpu_num  != cpus_weight(policy->cpus)) ||
-         (op->get_para.freq_num != pmpt->perf.state_count)    ||
-         (op->get_para.gov_num  != gov_num) )
-    {
-        op->get_para.cpu_num =  cpus_weight(policy->cpus);
-        op->get_para.freq_num = pmpt->perf.state_count;
-        op->get_para.gov_num  = gov_num;
+    if ( (op->u.get_para.cpu_num  != cpus_weight(policy->cpus)) ||
+         (op->u.get_para.freq_num != pmpt->perf.state_count)    ||
+         (op->u.get_para.gov_num  != gov_num) )
+    {
+        op->u.get_para.cpu_num =  cpus_weight(policy->cpus);
+        op->u.get_para.freq_num = pmpt->perf.state_count;
+        op->u.get_para.gov_num  = gov_num;
         return -EAGAIN;
     }
 
-    if ( !(affected_cpus = xmalloc_array(uint32_t, op->get_para.cpu_num)) )
+    if ( !(affected_cpus = xmalloc_array(uint32_t, op->u.get_para.cpu_num)) )
         return -ENOMEM;
-    memset(affected_cpus, 0, op->get_para.cpu_num * sizeof(uint32_t));
+    memset(affected_cpus, 0, op->u.get_para.cpu_num * sizeof(uint32_t));
     for_each_cpu_mask(cpu, policy->cpus)
         affected_cpus[j++] = cpu;
-    ret = copy_to_guest(op->get_para.affected_cpus,
-                       affected_cpus, op->get_para.cpu_num);
+    ret = copy_to_guest(op->u.get_para.affected_cpus,
+                       affected_cpus, op->u.get_para.cpu_num);
     xfree(affected_cpus);
     if ( ret )
         return ret;
 
     if ( !(scaling_available_frequencies =
-        xmalloc_array(uint32_t, op->get_para.freq_num)) )
+        xmalloc_array(uint32_t, op->u.get_para.freq_num)) )
         return -ENOMEM;
     memset(scaling_available_frequencies, 0,
-           op->get_para.freq_num * sizeof(uint32_t));
-    for ( i = 0; i < op->get_para.freq_num; i++ )
+           op->u.get_para.freq_num * sizeof(uint32_t));
+    for ( i = 0; i < op->u.get_para.freq_num; i++ )
         scaling_available_frequencies[i] =
                         pmpt->perf.states[i].core_frequency * 1000;
-    ret = copy_to_guest(op->get_para.scaling_available_frequencies,
-                   scaling_available_frequencies, op->get_para.freq_num);
+    ret = copy_to_guest(op->u.get_para.scaling_available_frequencies,
+                   scaling_available_frequencies, op->u.get_para.freq_num);
     xfree(scaling_available_frequencies);
     if ( ret )
         return ret;
@@ -258,47 +258,47 @@ static int get_cpufreq_para(struct xen_s
         xfree(scaling_available_governors);
         return ret;
     }
-    ret = copy_to_guest(op->get_para.scaling_available_governors,
+    ret = copy_to_guest(op->u.get_para.scaling_available_governors,
                 scaling_available_governors, gov_num * CPUFREQ_NAME_LEN);
     xfree(scaling_available_governors);
     if ( ret )
         return ret;
 
-    op->get_para.cpuinfo_cur_freq =
+    op->u.get_para.cpuinfo_cur_freq =
         cpufreq_driver->get ? cpufreq_driver->get(op->cpuid) : policy->cur;
-    op->get_para.cpuinfo_max_freq = policy->cpuinfo.max_freq;
-    op->get_para.cpuinfo_min_freq = policy->cpuinfo.min_freq;
-    op->get_para.scaling_cur_freq = policy->cur;
-    op->get_para.scaling_max_freq = policy->max;
-    op->get_para.scaling_min_freq = policy->min;
+    op->u.get_para.cpuinfo_max_freq = policy->cpuinfo.max_freq;
+    op->u.get_para.cpuinfo_min_freq = policy->cpuinfo.min_freq;
+    op->u.get_para.scaling_cur_freq = policy->cur;
+    op->u.get_para.scaling_max_freq = policy->max;
+    op->u.get_para.scaling_min_freq = policy->min;
 
     if ( cpufreq_driver->name )
-        strlcpy(op->get_para.scaling_driver, 
+        strlcpy(op->u.get_para.scaling_driver, 
             cpufreq_driver->name, CPUFREQ_NAME_LEN);
     else
-        strlcpy(op->get_para.scaling_driver, "Unknown", CPUFREQ_NAME_LEN);
+        strlcpy(op->u.get_para.scaling_driver, "Unknown", CPUFREQ_NAME_LEN);
 
     if ( policy->governor->name )
-        strlcpy(op->get_para.scaling_governor, 
+        strlcpy(op->u.get_para.scaling_governor, 
             policy->governor->name, CPUFREQ_NAME_LEN);
     else
-        strlcpy(op->get_para.scaling_governor, "Unknown", CPUFREQ_NAME_LEN);
+        strlcpy(op->u.get_para.scaling_governor, "Unknown", CPUFREQ_NAME_LEN);
 
     /* governor specific para */
-    if ( !strnicmp(op->get_para.scaling_governor, 
+    if ( !strnicmp(op->u.get_para.scaling_governor, 
                    "userspace", CPUFREQ_NAME_LEN) )
     {
-        op->get_para.u.userspace.scaling_setspeed = policy->cur;
-    }
-
-    if ( !strnicmp(op->get_para.scaling_governor, 
+        op->u.get_para.u.userspace.scaling_setspeed = policy->cur;
+    }
+
+    if ( !strnicmp(op->u.get_para.scaling_governor, 
                    "ondemand", CPUFREQ_NAME_LEN) )
     {
         ret = get_cpufreq_ondemand_para(
-            &op->get_para.u.ondemand.sampling_rate_max,
-            &op->get_para.u.ondemand.sampling_rate_min,
-            &op->get_para.u.ondemand.sampling_rate,
-            &op->get_para.u.ondemand.up_threshold); 
+            &op->u.get_para.u.ondemand.sampling_rate_max,
+            &op->u.get_para.u.ondemand.sampling_rate_min,
+            &op->u.get_para.u.ondemand.sampling_rate,
+            &op->u.get_para.u.ondemand.up_threshold); 
     }
 
     return ret;
@@ -317,7 +317,7 @@ static int set_cpufreq_gov(struct xen_sy
 
     memcpy(&new_policy, old_policy, sizeof(struct cpufreq_policy));
 
-    new_policy.governor = __find_governor(op->set_gov.scaling_governor);
+    new_policy.governor = __find_governor(op->u.set_gov.scaling_governor);
     if (new_policy.governor == NULL)
         return -EINVAL;
 
@@ -336,14 +336,14 @@ static int set_cpufreq_para(struct xen_s
     if ( !policy || !policy->governor )
         return -EINVAL;
 
-    switch(op->set_para.ctrl_type)
+    switch(op->u.set_para.ctrl_type)
     {
     case SCALING_MAX_FREQ:
     {
         struct cpufreq_policy new_policy;
 
         memcpy(&new_policy, policy, sizeof(struct cpufreq_policy));
-        new_policy.max = op->set_para.ctrl_value;
+        new_policy.max = op->u.set_para.ctrl_value;
         ret = __cpufreq_set_policy(policy, &new_policy);
 
         break;
@@ -354,7 +354,7 @@ static int set_cpufreq_para(struct xen_s
         struct cpufreq_policy new_policy;
 
         memcpy(&new_policy, policy, sizeof(struct cpufreq_policy));
-        new_policy.min = op->set_para.ctrl_value;
+        new_policy.min = op->u.set_para.ctrl_value;
         ret = __cpufreq_set_policy(policy, &new_policy);
 
         break;
@@ -362,7 +362,7 @@ static int set_cpufreq_para(struct xen_s
 
     case SCALING_SETSPEED:
     {
-        unsigned int freq =op->set_para.ctrl_value;
+        unsigned int freq =op->u.set_para.ctrl_value;
 
         if ( !strnicmp(policy->governor->name,
                        "userspace", CPUFREQ_NAME_LEN) )
@@ -375,7 +375,7 @@ static int set_cpufreq_para(struct xen_s
 
     case SAMPLING_RATE:
     {
-        unsigned int sampling_rate = op->set_para.ctrl_value;
+        unsigned int sampling_rate = op->u.set_para.ctrl_value;
 
         if ( !strnicmp(policy->governor->name,
                        "ondemand", CPUFREQ_NAME_LEN) )
@@ -388,7 +388,7 @@ static int set_cpufreq_para(struct xen_s
 
     case UP_THRESHOLD:
     {
-        unsigned int up_threshold = op->set_para.ctrl_value;
+        unsigned int up_threshold = op->u.set_para.ctrl_value;
 
         if ( !strnicmp(policy->governor->name,
                        "ondemand", CPUFREQ_NAME_LEN) )
@@ -412,7 +412,7 @@ static int get_cpufreq_avgfreq(struct xe
     if ( !op || !cpu_online(op->cpuid) )
         return -EINVAL;
 
-    op->get_avgfreq = cpufreq_driver_getavg(op->cpuid, USR_GETAVG);
+    op->u.get_avgfreq = cpufreq_driver_getavg(op->cpuid, USR_GETAVG);
 
     return 0;
 }
@@ -424,9 +424,9 @@ static int get_cputopo (struct xen_sysct
     XEN_GUEST_HANDLE_64(uint32) cpu_to_socket_arr;
     int arr_size, ret=0;
 
-    cpu_to_core_arr = op->get_topo.cpu_to_core;
-    cpu_to_socket_arr = op->get_topo.cpu_to_socket;
-    arr_size= min_t(uint32_t, op->get_topo.max_cpus, NR_CPUS);
+    cpu_to_core_arr = op->u.get_topo.cpu_to_core;
+    cpu_to_socket_arr = op->u.get_topo.cpu_to_socket;
+    arr_size= min_t(uint32_t, op->u.get_topo.max_cpus, NR_CPUS);
 
     if ( guest_handle_is_null( cpu_to_core_arr ) ||
             guest_handle_is_null(  cpu_to_socket_arr) )
@@ -458,7 +458,7 @@ static int get_cputopo (struct xen_sysct
         }
     }
 
-    op->get_topo.nr_cpus = nr_cpus + 1;
+    op->u.get_topo.nr_cpus = nr_cpus + 1;
 out:
     return ret;
 }
@@ -519,33 +519,33 @@ int do_pm_op(struct xen_sysctl_pm_op *op
         uint32_t saved_value;
 
         saved_value = sched_smt_power_savings;
-        sched_smt_power_savings = !!op->set_sched_opt_smt;
-        op->set_sched_opt_smt = saved_value;
+        sched_smt_power_savings = !!op->u.set_sched_opt_smt;
+        op->u.set_sched_opt_smt = saved_value;
 
         break;
     }
 
     case XEN_SYSCTL_pm_op_set_vcpu_migration_delay:
     {
-        set_vcpu_migration_delay(op->set_vcpu_migration_delay);
+        set_vcpu_migration_delay(op->u.set_vcpu_migration_delay);
         break;
     }
 
     case XEN_SYSCTL_pm_op_get_vcpu_migration_delay:
     {
-        op->get_vcpu_migration_delay = get_vcpu_migration_delay();
+        op->u.get_vcpu_migration_delay = get_vcpu_migration_delay();
         break;
     }
 
     case XEN_SYSCTL_pm_op_get_max_cstate:
     {
-        op->get_max_cstate = acpi_get_cstate_limit();
+        op->u.get_max_cstate = acpi_get_cstate_limit();
         break;
     }
 
     case XEN_SYSCTL_pm_op_set_max_cstate:
     {
-        acpi_set_cstate_limit(op->set_max_cstate);
+        acpi_set_cstate_limit(op->u.set_max_cstate);
         break;
     }
 
diff -r 3c93fffa6502 -r bde24590c13a xen/include/public/domctl.h
--- a/xen/include/public/domctl.h       Mon Dec 21 16:51:40 2009 +0000
+++ b/xen/include/public/domctl.h       Tue Dec 22 11:33:15 2009 +0000
@@ -35,7 +35,7 @@
 #include "xen.h"
 #include "grant_table.h"
 
-#define XEN_DOMCTL_INTERFACE_VERSION 0x00000005
+#define XEN_DOMCTL_INTERFACE_VERSION 0x00000006
 
 struct xenctl_cpumap {
     XEN_GUEST_HANDLE_64(uint8) bitmap;
@@ -539,7 +539,7 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_iopor
 #define XEN_DOMCTL_MEM_CACHEATTR_UCM 7
 struct xen_domctl_pin_mem_cacheattr {
     uint64_aligned_t start, end;
-    unsigned int type; /* XEN_DOMCTL_MEM_CACHEATTR_* */
+    uint32_t type; /* XEN_DOMCTL_MEM_CACHEATTR_* */
 };
 typedef struct xen_domctl_pin_mem_cacheattr xen_domctl_pin_mem_cacheattr_t;
 DEFINE_XEN_GUEST_HANDLE(xen_domctl_pin_mem_cacheattr_t);
@@ -598,11 +598,11 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_set_t
 # define XEN_CPUID_INPUT_UNUSED  0xFFFFFFFF
 # define XEN_DOMCTL_set_cpuid 49
 struct xen_domctl_cpuid {
-  unsigned int  input[2];
-  unsigned int  eax;
-  unsigned int  ebx;
-  unsigned int  ecx;
-  unsigned int  edx;
+  uint32_t input[2];
+  uint32_t eax;
+  uint32_t ebx;
+  uint32_t ecx;
+  uint32_t edx;
 };
 typedef struct xen_domctl_cpuid xen_domctl_cpuid_t;
 DEFINE_XEN_GUEST_HANDLE(xen_domctl_cpuid_t);
@@ -661,7 +661,7 @@ struct xen_guest_tsc_info {
     uint32_t gtsc_khz;
     uint32_t incarnation;
     uint32_t pad;
-    uint64_t elapsed_nsec;
+    uint64_aligned_t elapsed_nsec;
 };
 typedef struct xen_guest_tsc_info xen_guest_tsc_info_t;
 DEFINE_XEN_GUEST_HANDLE(xen_guest_tsc_info_t);
@@ -672,12 +672,14 @@ typedef struct xen_domctl_tsc_info {
 
 #define XEN_DOMCTL_gdbsx_guestmemio     1000 /* guest mem io */
 struct xen_domctl_gdbsx_memio {
+    /* IN */
     uint64_aligned_t pgd3val;/* optional: init_mm.pgd[3] value */
     uint64_aligned_t gva;    /* guest virtual address */
     uint64_aligned_t uva;    /* user buffer virtual address */
-    int              len;    /* number of bytes to read/write */
-    int              gwr;    /* 0 = read from guest. 1 = write to guest */
-    int              remain; /* bytes remaining to be copied */
+    uint32_t         len;    /* number of bytes to read/write */
+    uint8_t          gwr;    /* 0 = read from guest. 1 = write to guest */
+    /* OUT */
+    uint32_t         remain; /* bytes remaining to be copied */
 };
 
 #define XEN_DOMCTL_gdbsx_pausevcpu   1001  
@@ -688,10 +690,10 @@ struct xen_domctl_gdbsx_pauseunp_vcpu { 
 
 #define XEN_DOMCTL_gdbsx_domstatus   1003  
 struct xen_domctl_gdbsx_domstatus {
-    int              paused;     /* is the domain paused */
+    /* OUT */
+    uint8_t          paused;     /* is the domain paused */
     uint32_t         vcpu_id;    /* any vcpu in an event? */
     uint32_t         vcpu_ev;    /* if yes, what event? */
-
 };
 
 /*
@@ -720,11 +722,11 @@ struct xen_domctl_mem_event_op {
     uint32_t       mode;         /* XEN_DOMCTL_MEM_EVENT_ENABLE_* */
 
     /* OP_ENABLE */
-    unsigned long  shared_addr;  /* IN:  Virtual address of shared page */
-    unsigned long  ring_addr;    /* IN:  Virtual address of ring page */
+    uint64_aligned_t shared_addr;  /* IN:  Virtual address of shared page */
+    uint64_aligned_t ring_addr;    /* IN:  Virtual address of ring page */
 
     /* Other OPs */
-    unsigned long  gfn;          /* IN:  gfn of page being operated on */
+    uint64_aligned_t gfn;          /* IN:  gfn of page being operated on */
 };
 typedef struct xen_domctl_mem_event_op xen_domctl_mem_event_op_t;
 DEFINE_XEN_GUEST_HANDLE(xen_domctl_mem_event_op_t);
@@ -747,30 +749,30 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_mem_e
 #define XEN_DOMCTL_MEM_SHARING_C_HANDLE_INVALID  (-9)
 
 struct xen_domctl_mem_sharing_op {
-    uint8_t       op;            /* XEN_DOMCTL_MEM_EVENT_OP_* */
+    uint8_t op; /* XEN_DOMCTL_MEM_EVENT_OP_* */
 
     union {
-        int enable;                       /* for OP_CONTROL            */
-
-        struct mem_sharing_op_nominate {  /* for OP_NOMINATE           */
+        uint8_t enable;                   /* OP_CONTROL                */
+
+        struct mem_sharing_op_nominate {  /* OP_NOMINATE_xxx           */
             union {
-                unsigned long gfn;        /* IN: gfn to nominate       */
+                uint64_aligned_t gfn;     /* IN: gfn to nominate       */
                 uint32_t      grant_ref;  /* IN: grant ref to nominate */
-            };
-            uint64_t       handle;        /* OUT: the handle           */
+            } u;
+            uint64_aligned_t  handle;     /* OUT: the handle           */
         } nominate;
-        struct mem_sharing_op_share {
-            uint64_t source_handle;       /* IN: handle to the source page */
-            uint64_t client_handle;       /* IN: handle to the client page */
+        struct mem_sharing_op_share {     /* OP_SHARE */
+            uint64_aligned_t source_handle; /* IN: handle to the source page */
+            uint64_aligned_t client_handle; /* IN: handle to the client page */
         } share; 
-        struct mem_sharing_op_debug {
+        struct mem_sharing_op_debug {     /* OP_DEBUG_xxx */
             union {
-                unsigned long  gfn;        /* IN: gfn to debug          */
-                unsigned long  mfn;        /* IN: mfn to debug          */
+                uint64_aligned_t gfn;      /* IN: gfn to debug          */
+                uint64_aligned_t mfn;      /* IN: mfn to debug          */
                 grant_ref_t    gref;       /* IN: gref to debug         */
-            };
+            } u;
         } debug;
-    };
+    } u;
 };
 typedef struct xen_domctl_mem_sharing_op xen_domctl_mem_sharing_op_t;
 DEFINE_XEN_GUEST_HANDLE(xen_domctl_mem_sharing_op_t);
diff -r 3c93fffa6502 -r bde24590c13a xen/include/public/sysctl.h
--- a/xen/include/public/sysctl.h       Mon Dec 21 16:51:40 2009 +0000
+++ b/xen/include/public/sysctl.h       Tue Dec 22 11:33:15 2009 +0000
@@ -403,7 +403,7 @@ struct xen_sysctl_pm_op {
         uint32_t                    set_max_cstate;
         uint32_t                    get_vcpu_migration_delay;
         uint32_t                    set_vcpu_migration_delay;
-    };
+    } u;
 };
 
 #define XEN_SYSCTL_page_offline_op        14

_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog

<Prev in Thread] Current Thread [Next in Thread>
  • [Xen-changelog] [xen-unstable] domctl/sysctl: Clean up definitions, Xen patchbot-unstable <=