Skip to content

Commit

Permalink
perf_counter: Don't swap contexts containing locked mutex
Browse files Browse the repository at this point in the history
Peter Zijlstra pointed out that under some circumstances, we can take
the mutex in a context or a counter and then swap that context or
counter to another task, potentially leading to lock order inversions
or the mutexes not protecting what they are supposed to protect.

This fixes the problem by making sure that we never take a mutex in a
context or counter which could get swapped to another task.  Most of
the cases where we take a mutex is on a top-level counter or context,
i.e. a counter which has an fd associated with it or a context that
contains such a counter.  This adds WARN_ON_ONCE statements to verify
that.

The two cases where we need to take the mutex on a context that is a
clone of another are in perf_counter_exit_task and
perf_counter_init_task.  The perf_counter_exit_task case is solved by
uncloning the context before starting to remove the counters from it.
The perf_counter_init_task is a little trickier; we temporarily
disable context swapping for the parent (forking) task by setting its
ctx->parent_gen to the all-1s value after locking the context, if it
is a cloned context, and restore the ctx->parent_gen value at the end
if the context didn't get uncloned in the meantime.

This also moves the increment of the context generation count to be
within the same critical section, protected by the context mutex, that
adds the new counter to the context.  That way, taking the mutex is
sufficient to ensure that both the counter list and the generation
count are stable.

[ Impact: fix hangs, races with inherited and PID counters ]

Signed-off-by: Paul Mackerras <paulus@samba.org>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <18975.31580.520676.619896@drongo.ozlabs.ibm.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
  • Loading branch information
Paul Mackerras authored and Ingo Molnar committed May 29, 2009
1 parent be1ac0d commit ad3a37d
Showing 1 changed file with 79 additions and 10 deletions.
89 changes: 79 additions & 10 deletions kernel/perf_counter.c
Original file line number Diff line number Diff line change
Expand Up @@ -919,7 +919,8 @@ static int context_equiv(struct perf_counter_context *ctx1,
struct perf_counter_context *ctx2)
{
return ctx1->parent_ctx && ctx1->parent_ctx == ctx2->parent_ctx
&& ctx1->parent_gen == ctx2->parent_gen;
&& ctx1->parent_gen == ctx2->parent_gen
&& ctx1->parent_gen != ~0ull;
}

/*
Expand Down Expand Up @@ -1339,7 +1340,6 @@ static struct perf_counter_context *find_get_context(pid_t pid, int cpu)
put_ctx(parent_ctx);
ctx->parent_ctx = NULL; /* no longer a clone */
}
++ctx->generation;
/*
* Get an extra reference before dropping the lock so that
* this context won't get freed if the task exits.
Expand Down Expand Up @@ -1414,6 +1414,7 @@ static int perf_release(struct inode *inode, struct file *file)

file->private_data = NULL;

WARN_ON_ONCE(ctx->parent_ctx);
mutex_lock(&ctx->mutex);
perf_counter_remove_from_context(counter);
mutex_unlock(&ctx->mutex);
Expand Down Expand Up @@ -1445,6 +1446,7 @@ perf_read_hw(struct perf_counter *counter, char __user *buf, size_t count)
if (counter->state == PERF_COUNTER_STATE_ERROR)
return 0;

WARN_ON_ONCE(counter->ctx->parent_ctx);
mutex_lock(&counter->child_mutex);
values[0] = perf_counter_read(counter);
n = 1;
Expand Down Expand Up @@ -1504,6 +1506,7 @@ static void perf_counter_for_each_sibling(struct perf_counter *counter,
struct perf_counter_context *ctx = counter->ctx;
struct perf_counter *sibling;

WARN_ON_ONCE(ctx->parent_ctx);
mutex_lock(&ctx->mutex);
counter = counter->group_leader;

Expand All @@ -1524,6 +1527,7 @@ static void perf_counter_for_each_child(struct perf_counter *counter,
{
struct perf_counter *child;

WARN_ON_ONCE(counter->ctx->parent_ctx);
mutex_lock(&counter->child_mutex);
func(counter);
list_for_each_entry(child, &counter->child_list, child_list)
Expand All @@ -1536,6 +1540,7 @@ static void perf_counter_for_each(struct perf_counter *counter,
{
struct perf_counter *child;

WARN_ON_ONCE(counter->ctx->parent_ctx);
mutex_lock(&counter->child_mutex);
perf_counter_for_each_sibling(counter, func);
list_for_each_entry(child, &counter->child_list, child_list)
Expand Down Expand Up @@ -1741,6 +1746,7 @@ static void perf_mmap_close(struct vm_area_struct *vma)
{
struct perf_counter *counter = vma->vm_file->private_data;

WARN_ON_ONCE(counter->ctx->parent_ctx);
if (atomic_dec_and_mutex_lock(&counter->mmap_count,
&counter->mmap_mutex)) {
struct user_struct *user = current_user();
Expand Down Expand Up @@ -1788,6 +1794,7 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma)
if (vma->vm_pgoff != 0)
return -EINVAL;

WARN_ON_ONCE(counter->ctx->parent_ctx);
mutex_lock(&counter->mmap_mutex);
if (atomic_inc_not_zero(&counter->mmap_count)) {
if (nr_pages != counter->data->nr_pages)
Expand Down Expand Up @@ -3349,8 +3356,10 @@ SYSCALL_DEFINE5(perf_counter_open,
goto err_free_put_context;

counter->filp = counter_file;
WARN_ON_ONCE(ctx->parent_ctx);
mutex_lock(&ctx->mutex);
perf_install_in_context(ctx, counter, cpu);
++ctx->generation;
mutex_unlock(&ctx->mutex);

counter->owner = current;
Expand Down Expand Up @@ -3436,6 +3445,7 @@ inherit_counter(struct perf_counter *parent_counter,
/*
* Link this into the parent counter's child list
*/
WARN_ON_ONCE(parent_counter->ctx->parent_ctx);
mutex_lock(&parent_counter->child_mutex);
list_add_tail(&child_counter->child_list, &parent_counter->child_list);
mutex_unlock(&parent_counter->child_mutex);
Expand Down Expand Up @@ -3485,6 +3495,7 @@ static void sync_child_counter(struct perf_counter *child_counter,
/*
* Remove this counter from the parent's list
*/
WARN_ON_ONCE(parent_counter->ctx->parent_ctx);
mutex_lock(&parent_counter->child_mutex);
list_del_init(&child_counter->child_list);
mutex_unlock(&parent_counter->child_mutex);
Expand Down Expand Up @@ -3527,12 +3538,17 @@ void perf_counter_exit_task(struct task_struct *child)
struct perf_counter_context *child_ctx;
unsigned long flags;

child_ctx = child->perf_counter_ctxp;

if (likely(!child_ctx))
if (likely(!child->perf_counter_ctxp))
return;

local_irq_save(flags);
/*
* We can't reschedule here because interrupts are disabled,
* and either child is current or it is a task that can't be
* scheduled, so we are now safe from rescheduling changing
* our context.
*/
child_ctx = child->perf_counter_ctxp;
__perf_counter_task_sched_out(child_ctx);

/*
Expand All @@ -3542,6 +3558,15 @@ void perf_counter_exit_task(struct task_struct *child)
*/
spin_lock(&child_ctx->lock);
child->perf_counter_ctxp = NULL;
if (child_ctx->parent_ctx) {
/*
* This context is a clone; unclone it so it can't get
* swapped to another process while we're removing all
* the counters from it.
*/
put_ctx(child_ctx->parent_ctx);
child_ctx->parent_ctx = NULL;
}
spin_unlock(&child_ctx->lock);
local_irq_restore(flags);

Expand Down Expand Up @@ -3571,18 +3596,19 @@ void perf_counter_exit_task(struct task_struct *child)
int perf_counter_init_task(struct task_struct *child)
{
struct perf_counter_context *child_ctx, *parent_ctx;
struct perf_counter_context *cloned_ctx;
struct perf_counter *counter;
struct task_struct *parent = current;
int inherited_all = 1;
u64 cloned_gen;
int ret = 0;

child->perf_counter_ctxp = NULL;

mutex_init(&child->perf_counter_mutex);
INIT_LIST_HEAD(&child->perf_counter_list);

parent_ctx = parent->perf_counter_ctxp;
if (likely(!parent_ctx || !parent_ctx->nr_counters))
if (likely(!parent->perf_counter_ctxp))
return 0;

/*
Expand All @@ -3599,6 +3625,34 @@ int perf_counter_init_task(struct task_struct *child)
child->perf_counter_ctxp = child_ctx;
get_task_struct(child);

/*
* If the parent's context is a clone, temporarily set its
* parent_gen to an impossible value (all 1s) so it won't get
* swapped under us. The rcu_read_lock makes sure that
* parent_ctx continues to exist even if it gets swapped to
* another process and then freed while we are trying to get
* its lock.
*/
rcu_read_lock();
retry:
parent_ctx = rcu_dereference(parent->perf_counter_ctxp);
/*
* No need to check if parent_ctx != NULL here; since we saw
* it non-NULL earlier, the only reason for it to become NULL
* is if we exit, and since we're currently in the middle of
* a fork we can't be exiting at the same time.
*/
spin_lock_irq(&parent_ctx->lock);
if (parent_ctx != rcu_dereference(parent->perf_counter_ctxp)) {
spin_unlock_irq(&parent_ctx->lock);
goto retry;
}
cloned_gen = parent_ctx->parent_gen;
if (parent_ctx->parent_ctx)
parent_ctx->parent_gen = ~0ull;
spin_unlock_irq(&parent_ctx->lock);
rcu_read_unlock();

/*
* Lock the parent list. No need to lock the child - not PID
* hashed yet and not running, so nobody can access it.
Expand Down Expand Up @@ -3630,10 +3684,15 @@ int perf_counter_init_task(struct task_struct *child)
/*
* Mark the child context as a clone of the parent
* context, or of whatever the parent is a clone of.
* Note that if the parent is a clone, it could get
* uncloned at any point, but that doesn't matter
* because the list of counters and the generation
* count can't have changed since we took the mutex.
*/
if (parent_ctx->parent_ctx) {
child_ctx->parent_ctx = parent_ctx->parent_ctx;
child_ctx->parent_gen = parent_ctx->parent_gen;
cloned_ctx = rcu_dereference(parent_ctx->parent_ctx);
if (cloned_ctx) {
child_ctx->parent_ctx = cloned_ctx;
child_ctx->parent_gen = cloned_gen;
} else {
child_ctx->parent_ctx = parent_ctx;
child_ctx->parent_gen = parent_ctx->generation;
Expand All @@ -3643,6 +3702,16 @@ int perf_counter_init_task(struct task_struct *child)

mutex_unlock(&parent_ctx->mutex);

/*
* Restore the clone status of the parent.
*/
if (parent_ctx->parent_ctx) {
spin_lock_irq(&parent_ctx->lock);
if (parent_ctx->parent_ctx)
parent_ctx->parent_gen = cloned_gen;
spin_unlock_irq(&parent_ctx->lock);
}

return ret;
}

Expand Down

0 comments on commit ad3a37d

Please sign in to comment.