@@ -28,6 +28,8 @@ typedef void (*postgp_func_t)(struct rcu_tasks *rtp);
28
28
* @kthread_ptr: This flavor's grace-period/callback-invocation kthread.
29
29
* @gp_func: This flavor's grace-period-wait function.
30
30
* @gp_state: Grace period's most recent state transition (debugging).
31
+ * @gp_sleep: Per-grace-period sleep to prevent CPU-bound looping.
32
+ * @init_fract: Initial backoff sleep interval.
31
33
* @gp_jiffies: Time of last @gp_state transition.
32
34
* @gp_start: Most recent grace-period start in jiffies.
33
35
* @n_gps: Number of grace periods completed since boot.
@@ -48,6 +50,8 @@ struct rcu_tasks {
48
50
struct wait_queue_head cbs_wq ;
49
51
raw_spinlock_t cbs_lock ;
50
52
int gp_state ;
53
+ int gp_sleep ;
54
+ int init_fract ;
51
55
unsigned long gp_jiffies ;
52
56
unsigned long gp_start ;
53
57
unsigned long n_gps ;
@@ -81,7 +85,7 @@ static struct rcu_tasks rt_name = \
81
85
DEFINE_STATIC_SRCU (tasks_rcu_exit_srcu );
82
86
83
87
/* Avoid IPIing CPUs early in the grace period. */
84
- #define RCU_TASK_IPI_DELAY (HZ / 2)
88
+ #define RCU_TASK_IPI_DELAY (IS_ENABLED(CONFIG_TASKS_TRACE_RCU_READ_MB) ? HZ / 2 : 0 )
85
89
static int rcu_task_ipi_delay __read_mostly = RCU_TASK_IPI_DELAY ;
86
90
module_param (rcu_task_ipi_delay , int , 0644 );
87
91
@@ -231,7 +235,7 @@ static int __noreturn rcu_tasks_kthread(void *arg)
231
235
cond_resched ();
232
236
}
233
237
/* Paranoid sleep to keep this from entering a tight loop */
234
- schedule_timeout_idle (HZ / 10 );
238
+ schedule_timeout_idle (rtp -> gp_sleep );
235
239
236
240
set_tasks_gp_state (rtp , RTGS_WAIT_CBS );
237
241
}
@@ -329,8 +333,10 @@ static void rcu_tasks_wait_gp(struct rcu_tasks *rtp)
329
333
*/
330
334
lastreport = jiffies ;
331
335
332
- /* Start off with HZ/10 wait and slowly back off to 1 HZ wait. */
333
- fract = 10 ;
336
+ // Start off with initial wait and slowly back off to 1 HZ wait.
337
+ fract = rtp -> init_fract ;
338
+ if (fract > HZ )
339
+ fract = HZ ;
334
340
335
341
for (;;) {
336
342
bool firstreport ;
@@ -553,6 +559,8 @@ EXPORT_SYMBOL_GPL(rcu_barrier_tasks);
553
559
554
560
static int __init rcu_spawn_tasks_kthread (void )
555
561
{
562
+ rcu_tasks .gp_sleep = HZ / 10 ;
563
+ rcu_tasks .init_fract = 10 ;
556
564
rcu_tasks .pregp_func = rcu_tasks_pregp_step ;
557
565
rcu_tasks .pertask_func = rcu_tasks_pertask ;
558
566
rcu_tasks .postscan_func = rcu_tasks_postscan ;
@@ -685,6 +693,7 @@ EXPORT_SYMBOL_GPL(rcu_barrier_tasks_rude);
685
693
686
694
static int __init rcu_spawn_tasks_rude_kthread (void )
687
695
{
696
+ rcu_tasks_rude .gp_sleep = HZ / 10 ;
688
697
rcu_spawn_tasks_kthread_generic (& rcu_tasks_rude );
689
698
return 0 ;
690
699
}
@@ -745,9 +754,9 @@ static DEFINE_PER_CPU(bool, trc_ipi_to_cpu);
745
754
746
755
// The number of detections of task quiescent state relying on
747
756
// heavyweight readers executing explicit memory barriers.
748
- unsigned long n_heavy_reader_attempts ;
749
- unsigned long n_heavy_reader_updates ;
750
- unsigned long n_heavy_reader_ofl_updates ;
757
+ static unsigned long n_heavy_reader_attempts ;
758
+ static unsigned long n_heavy_reader_updates ;
759
+ static unsigned long n_heavy_reader_ofl_updates ;
751
760
752
761
void call_rcu_tasks_trace (struct rcu_head * rhp , rcu_callback_t func );
753
762
DEFINE_RCU_TASKS (rcu_tasks_trace , rcu_tasks_wait_gp , call_rcu_tasks_trace ,
@@ -821,6 +830,12 @@ static void trc_read_check_handler(void *t_in)
821
830
WRITE_ONCE (t -> trc_reader_checked , true);
822
831
goto reset_ipi ;
823
832
}
833
+ // If we are racing with an rcu_read_unlock_trace(), try again later.
834
+ if (unlikely (t -> trc_reader_nesting < 0 )) {
835
+ if (WARN_ON_ONCE (atomic_dec_and_test (& trc_n_readers_need_end )))
836
+ wake_up (& trc_wait );
837
+ goto reset_ipi ;
838
+ }
824
839
WRITE_ONCE (t -> trc_reader_checked , true);
825
840
826
841
// Get here if the task is in a read-side critical section. Set
@@ -911,7 +926,8 @@ static void trc_wait_for_one_reader(struct task_struct *t,
911
926
912
927
// If currently running, send an IPI, either way, add to list.
913
928
trc_add_holdout (t , bhp );
914
- if (task_curr (t ) && time_after (jiffies , rcu_tasks_trace .gp_start + rcu_task_ipi_delay )) {
929
+ if (task_curr (t ) &&
930
+ time_after (jiffies + 1 , rcu_tasks_trace .gp_start + rcu_task_ipi_delay )) {
915
931
// The task is currently running, so try IPIing it.
916
932
cpu = task_cpu (t );
917
933
@@ -1072,15 +1088,17 @@ static void rcu_tasks_trace_postgp(struct rcu_tasks *rtp)
1072
1088
if (ret )
1073
1089
break ; // Count reached zero.
1074
1090
// Stall warning time, so make a list of the offenders.
1091
+ rcu_read_lock ();
1075
1092
for_each_process_thread (g , t )
1076
1093
if (READ_ONCE (t -> trc_reader_special .b .need_qs ))
1077
1094
trc_add_holdout (t , & holdouts );
1095
+ rcu_read_unlock ();
1078
1096
firstreport = true;
1079
- list_for_each_entry_safe (t , g , & holdouts , trc_holdout_list )
1080
- if (READ_ONCE (t -> trc_reader_special .b .need_qs )) {
1097
+ list_for_each_entry_safe (t , g , & holdouts , trc_holdout_list ) {
1098
+ if (READ_ONCE (t -> trc_reader_special .b .need_qs ))
1081
1099
show_stalled_task_trace (t , & firstreport );
1082
- trc_del_holdout (t );
1083
- }
1100
+ trc_del_holdout (t ); // Release task_struct reference.
1101
+ }
1084
1102
if (firstreport )
1085
1103
pr_err ("INFO: rcu_tasks_trace detected stalls? (Counter/taskslist mismatch?)\n" );
1086
1104
show_stalled_ipi_trace ();
@@ -1163,6 +1181,17 @@ EXPORT_SYMBOL_GPL(rcu_barrier_tasks_trace);
1163
1181
1164
1182
static int __init rcu_spawn_tasks_trace_kthread (void )
1165
1183
{
1184
+ if (IS_ENABLED (CONFIG_TASKS_TRACE_RCU_READ_MB )) {
1185
+ rcu_tasks_trace .gp_sleep = HZ / 10 ;
1186
+ rcu_tasks_trace .init_fract = 10 ;
1187
+ } else {
1188
+ rcu_tasks_trace .gp_sleep = HZ / 200 ;
1189
+ if (rcu_tasks_trace .gp_sleep <= 0 )
1190
+ rcu_tasks_trace .gp_sleep = 1 ;
1191
+ rcu_tasks_trace .init_fract = HZ / 5 ;
1192
+ if (rcu_tasks_trace .init_fract <= 0 )
1193
+ rcu_tasks_trace .init_fract = 1 ;
1194
+ }
1166
1195
rcu_tasks_trace .pregp_func = rcu_tasks_trace_pregp_step ;
1167
1196
rcu_tasks_trace .pertask_func = rcu_tasks_trace_pertask ;
1168
1197
rcu_tasks_trace .postscan_func = rcu_tasks_trace_postscan ;
0 commit comments