diff options
author | Ian Rogers <irogers@google.com> | 2020-02-13 23:51:32 -0800 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2020-03-06 11:57:00 +0100 |
commit | c2283c9368d41063f2077cb58def02217360526d (patch) | |
tree | d6aa52773c38c4076a0b8e7eb444db8a6f8546b0 /kernel/events | |
parent | 836196beb377e59e54ec9e04f7402076ef7a8bd8 (diff) |
perf/cgroup: Grow per perf_cpu_context heap storage
Allow the per-CPU min heap storage to have sufficient space for per-cgroup
iterators.
Based-on-work-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Ian Rogers <irogers@google.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Link: https://lkml.kernel.org/r/20200214075133.181299-6-irogers@google.com
Diffstat (limited to 'kernel/events')
-rw-r--r-- | kernel/events/core.c | 47 |
1 files changed, 47 insertions, 0 deletions
diff --git a/kernel/events/core.c b/kernel/events/core.c index 7529e76a2e36..8065949a865e 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c @@ -892,6 +892,47 @@ static inline void perf_cgroup_sched_in(struct task_struct *prev, rcu_read_unlock(); } +static int perf_cgroup_ensure_storage(struct perf_event *event, + struct cgroup_subsys_state *css) +{ + struct perf_cpu_context *cpuctx; + struct perf_event **storage; + int cpu, heap_size, ret = 0; + + /* + * Allow storage to have sufficent space for an iterator for each + * possibly nested cgroup plus an iterator for events with no cgroup. + */ + for (heap_size = 1; css; css = css->parent) + heap_size++; + + for_each_possible_cpu(cpu) { + cpuctx = per_cpu_ptr(event->pmu->pmu_cpu_context, cpu); + if (heap_size <= cpuctx->heap_size) + continue; + + storage = kmalloc_node(heap_size * sizeof(struct perf_event *), + GFP_KERNEL, cpu_to_node(cpu)); + if (!storage) { + ret = -ENOMEM; + break; + } + + raw_spin_lock_irq(&cpuctx->ctx.lock); + if (cpuctx->heap_size < heap_size) { + swap(cpuctx->heap, storage); + if (storage == cpuctx->heap_default) + storage = NULL; + cpuctx->heap_size = heap_size; + } + raw_spin_unlock_irq(&cpuctx->ctx.lock); + + kfree(storage); + } + + return ret; +} + static inline int perf_cgroup_connect(int fd, struct perf_event *event, struct perf_event_attr *attr, struct perf_event *group_leader) @@ -911,6 +952,10 @@ static inline int perf_cgroup_connect(int fd, struct perf_event *event, goto out; } + ret = perf_cgroup_ensure_storage(event, css); + if (ret) + goto out; + cgrp = container_of(css, struct perf_cgroup, css); event->cgrp = cgrp; @@ -3440,6 +3485,8 @@ static noinline int visit_groups_merge(struct perf_cpu_context *cpuctx, .nr = 0, .size = cpuctx->heap_size, }; + + lockdep_assert_held(&cpuctx->ctx.lock); } else { event_heap = (struct min_heap){ .data = itrs, |