From 5c6b0c3cb093782507f05fbbe7e070641a69ab45 Mon Sep 17 00:00:00 2001 Message-Id: <5c6b0c3cb093782507f05fbbe7e070641a69ab45.1601675152.git.zanussi@kernel.org> In-Reply-To: <5b5a156f9808b1acf1205606e03da117214549ea.1601675151.git.zanussi@kernel.org> References: <5b5a156f9808b1acf1205606e03da117214549ea.1601675151.git.zanussi@kernel.org> From: Thomas Gleixner Date: Sun, 17 Jul 2011 22:51:33 +0200 Subject: [PATCH 111/333] sched: Use the proper LOCK_OFFSET for cond_resched() Origin: https://www.kernel.org/pub/linux/kernel/projects/rt/4.19/older/patches-4.19.148-rt64.tar.xz RT does not increment preempt count when a 'sleeping' spinlock is locked. Update PREEMPT_LOCK_OFFSET for that case. Signed-off-by: Thomas Gleixner --- include/linux/preempt.h | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/include/linux/preempt.h b/include/linux/preempt.h index f7a17fcc3fec..b7fe717eb1f4 100644 --- a/include/linux/preempt.h +++ b/include/linux/preempt.h @@ -118,7 +118,11 @@ /* * The preempt_count offset after spin_lock() */ +#if !defined(CONFIG_PREEMPT_RT_FULL) #define PREEMPT_LOCK_OFFSET PREEMPT_DISABLE_OFFSET +#else +#define PREEMPT_LOCK_OFFSET 0 +#endif /* * The preempt_count offset needed for things like: -- 2.17.1