]> bbs.cooldavid.org Git - net-next-2.6.git/commitdiff
perf: Avoid RCU vs preemption assumptions
authorPeter Zijlstra <a.p.zijlstra@chello.nl>
Thu, 16 Sep 2010 17:17:24 +0000 (19:17 +0200)
committerIngo Molnar <mingo@elte.hu>
Tue, 21 Sep 2010 11:55:44 +0000 (13:55 +0200)
The per-pmu per-cpu context patch converted things from
get_cpu_var() to this_cpu_ptr(), but that only works if
rcu_read_lock() actually disables preemption, and since
there is no such guarantee, we need to fix that.

Use the newly introduced {get,put}_cpu_ptr().

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Tejun Heo <tj@kernel.org>
LKML-Reference: <20100917093009.308453028@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
kernel/perf_event.c

index baae1367e9450a229f0c3654680302fd2a98a873..c16158c77dfd015a64ee0d0efac9b7617224dd39 100644 (file)
@@ -3836,18 +3836,20 @@ static void perf_event_task_event(struct perf_task_event *task_event)
 
        rcu_read_lock();
        list_for_each_entry_rcu(pmu, &pmus, entry) {
-               cpuctx = this_cpu_ptr(pmu->pmu_cpu_context);
+               cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
                perf_event_task_ctx(&cpuctx->ctx, task_event);
 
                ctx = task_event->task_ctx;
                if (!ctx) {
                        ctxn = pmu->task_ctx_nr;
                        if (ctxn < 0)
-                               continue;
+                               goto next;
                        ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
                }
                if (ctx)
                        perf_event_task_ctx(ctx, task_event);
+next:
+               put_cpu_ptr(pmu->pmu_cpu_context);
        }
        rcu_read_unlock();
 }
@@ -3969,16 +3971,18 @@ static void perf_event_comm_event(struct perf_comm_event *comm_event)
 
        rcu_read_lock();
        list_for_each_entry_rcu(pmu, &pmus, entry) {
-               cpuctx = this_cpu_ptr(pmu->pmu_cpu_context);
+               cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
                perf_event_comm_ctx(&cpuctx->ctx, comm_event);
 
                ctxn = pmu->task_ctx_nr;
                if (ctxn < 0)
-                       continue;
+                       goto next;
 
                ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
                if (ctx)
                        perf_event_comm_ctx(ctx, comm_event);
+next:
+               put_cpu_ptr(pmu->pmu_cpu_context);
        }
        rcu_read_unlock();
 }
@@ -4152,19 +4156,21 @@ got_name:
 
        rcu_read_lock();
        list_for_each_entry_rcu(pmu, &pmus, entry) {
-               cpuctx = this_cpu_ptr(pmu->pmu_cpu_context);
+               cpuctx = get_cpu_ptr(pmu->pmu_cpu_context);
                perf_event_mmap_ctx(&cpuctx->ctx, mmap_event,
                                        vma->vm_flags & VM_EXEC);
 
                ctxn = pmu->task_ctx_nr;
                if (ctxn < 0)
-                       continue;
+                       goto next;
 
                ctx = rcu_dereference(current->perf_event_ctxp[ctxn]);
                if (ctx) {
                        perf_event_mmap_ctx(ctx, mmap_event,
                                        vma->vm_flags & VM_EXEC);
                }
+next:
+               put_cpu_ptr(pmu->pmu_cpu_context);
        }
        rcu_read_unlock();