[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH RFC 01/49] xen/sched: call cpu_disable_scheduler() via cpu notifier



Hi,

On 4/1/19 3:23 PM, Juergen Gross wrote:
On 01/04/2019 16:01, Julien Grall wrote:
Hi,

On 4/1/19 2:33 PM, Juergen Gross wrote:
On 01/04/2019 15:21, Julien Grall wrote:
Hi Juergen,

On 4/1/19 11:37 AM, Juergen Gross wrote:
On 01/04/2019 12:29, Julien Grall wrote:
Hi,

On 4/1/19 10:40 AM, Juergen Gross wrote:
On 01/04/2019 11:21, Julien Grall wrote:
Hi,

On 3/29/19 3:08 PM, Juergen Gross wrote:
cpu_disable_scheduler() is being called from __cpu_disable() today.
There is no need to execute it on the cpu just being disabled,
so use
the CPU_DEAD case of the cpu notifier chain. Moving the call out of
stop_machine() context is fine, as we just need to hold the domain
RCU
lock and need the scheduler percpu data to be still allocated.

Add another hook for CPU_DOWN_PREPARE to bail out early in case
cpu_disable_scheduler() would fail. This will avoid crashes in rare
cases for cpu hotplug or suspend.

While at it remove a superfluous smp_mb() in the ARM
__cpu_disable()
incarnation.

This is not obvious why the smp_mb() is superfluous. Can you please
provide more details on why this is not necessary?

cpumask_clear_cpu() should already have the needed semantics, no?
It is based on clear_bit() which is defined to be atomic.

atomicity does not mean the store/load cannot be re-ordered by the
CPU.
You would need a barrier to prevent re-ordering.

cpumask_clear_cpu() and clear_bit() does not contain any barrier, so
store/load can be re-ordered.

Uh, couldn't this lead to problems, e.g. in vcpu_block()? The comment
there suggests the sequence of setting the blocked bit and doing the
test is important for avoiding a race...

Hmmm... looking at the other usage (such as in do_poll), on non-x86
platform, there is a smp_mb() between set_bit(...) and checking the
event with a similar comment above.

I don't know enough the scheduler code to know why the barrier is
needed. But for consistency, it seems to me the smp_mb() would be
required in vcpu_block() as well.

Also, it is quite interesting that the barrier is not presence for x86.
If I understand correctly the comment on top of set_bit/clear_bit, it
could as well be re-ordered. So we seem to relying on the underlying
implementation of set_bit/clear_bit.

On x86 reads and writes can't be reordered with locked operations (SDM
Vol 3 8.2.2). So the barrier is really not needed AFAIU.

include/asm-x86/bitops.h:

   * clear_bit() is atomic and may not be reordered.

I interpreted the "may not" as you should not rely on the re-ordering to
not happen.

In place were re-ordering should not happen (e.g test_and_set_bit) we
use the wording "cannot".

The SDM is very clear here:

"Reads or writes cannot be reordered with I/O instructions, locked
  instructions, or serializing instructions."

This is what the specification says not the intended semantic. Helper may have a more relaxed semantics to accommodate other architecture.

I believe, this is the case here. The semantic is more relaxed than the implementation. So you don't have to impose a barrier in architecture with a more relaxed memory ordering.


Wouldn't it make sense to try to uniformize the semantics? Maybe by
introducing a new helper?

Or adding the barrier on ARM for the atomic operations?

On which basis?  Why should we impact every users for fixing a bug in
the scheduler?

I'm assuming there are more places like this either in common code or
code copied verbatim from arch/x86 to arch/arm with that problem.

Adding it in the *_set helpers is just the poor's man fix. If we do that this is going to stick for a long time and impact performance.

Instead we should fix the scheduler code (and hopefully only that) where the ordering is necessary.


So I take it you'd rather let me add that smp_mb() in __cpu_disable()
again.

Removing/Adding barriers should be accompanied with a proper justifications in the commit message. Additionally, new barrier should have a comment explaining what they are for.

In this case, I don't know what is the correct answer. It feels to me we should keep it until we have a better understanding of this code. But then it raises the question whether a barrier would also be necessary after calling cpu_disable_scheduler().

Cheers,

--
Julien Grall

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.