[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH 2/6] xen/sched: Use %*pb[l] instead of cpumask_scn{, list}printf()



>>> On 06.09.18 at 14:08, <andrew.cooper3@xxxxxxxxxx> wrote:
> @@ -2059,11 +2058,10 @@ csched_dump_pcpu(const struct scheduler *ops, int cpu)
>      spc = CSCHED_PCPU(cpu);
>      runq = &spc->runq;
>  
> -    cpumask_scnprintf(cpustr, sizeof(cpustr), per_cpu(cpu_sibling_mask, 
> cpu));
> -    printk("CPU[%02d] nr_run=%d, sort=%d, sibling=%s, ",
> -           cpu, spc->nr_runnable, spc->runq_sort_last, cpustr);
> -    cpumask_scnprintf(cpustr, sizeof(cpustr), per_cpu(cpu_core_mask, cpu));
> -    printk("core=%s\n", cpustr);
> +    printk("CPU[%02d] nr_run=%d, sort=%d, sibling=%*pb, core=%*pb\n",
> +           cpu, spc->nr_runnable, spc->runq_sort_last,
> +           nr_cpu_ids, per_cpu(cpu_sibling_mask, cpu),
> +           nr_cpu_ids, per_cpu(cpu_core_mask, cpu));

Strictly speaking here and elsewhere you should wrap the CPU mask
accesses in cpumask_bits(). Then again I wonder whether a special
case for CPU masks wouldn't be warranted, making it unnecessary for
callers to pass in nr_cpu_ids explicitly.

Jan



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.