[Intel-gfx] [patch 08/13] sched: Clenaup PREEMPT_COUNT leftovers
Valentin Schneider
valentin.schneider at arm.com
Wed Sep 16 10:56:23 UTC 2020
On 14/09/20 21:42, Thomas Gleixner wrote:
> CONFIG_PREEMPT_COUNT is now unconditionally enabled and will be
> removed. Cleanup the leftovers before doing so.
>
> Signed-off-by: Thomas Gleixner <tglx at linutronix.de>
> Cc: Ingo Molnar <mingo at redhat.com>
> Cc: Peter Zijlstra <peterz at infradead.org>
> Cc: Juri Lelli <juri.lelli at redhat.com>
> Cc: Vincent Guittot <vincent.guittot at linaro.org>
> Cc: Dietmar Eggemann <dietmar.eggemann at arm.com>
> Cc: Steven Rostedt <rostedt at goodmis.org>
> Cc: Ben Segall <bsegall at google.com>
> Cc: Mel Gorman <mgorman at suse.de>
> Cc: Daniel Bristot de Oliveira <bristot at redhat.com>
Small nit below;
Reviewed-by: Valentin Schneider <valentin.schneider at arm.com>
> ---
> kernel/sched/core.c | 6 +-----
> lib/Kconfig.debug | 1 -
> 2 files changed, 1 insertion(+), 6 deletions(-)
>
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -3706,8 +3706,7 @@ asmlinkage __visible void schedule_tail(
> * finish_task_switch() for details.
> *
> * finish_task_switch() will drop rq->lock() and lower preempt_count
> - * and the preempt_enable() will end up enabling preemption (on
> - * PREEMPT_COUNT kernels).
I suppose this wanted to be s/PREEMPT_COUNT/PREEMPT/ in the first place,
which ought to be still relevant.
> + * and the preempt_enable() will end up enabling preemption.
> */
>
> rq = finish_task_switch(prev);
More information about the Intel-gfx
mailing list