perf/core: Remove perf_event::group_entry
Now that all the grouping is done with RB trees, we no longer need group_entry and can replace the whole thing with sibling_list. Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Acked-by: Mark Rutland <mark.rutland@arm.com> Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com> Cc: Alexey Budankov <alexey.budankov@linux.intel.com> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: David Carrillo-Cisneros <davidcc@google.com> Cc: Dmitri Prokhorov <Dmitry.Prohorov@intel.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: Kan Liang <kan.liang@intel.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Stephane Eranian <eranian@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Valery Cherepennikov <valery.cherepennikov@intel.com> Cc: Vince Weaver <vincent.weaver@maine.edu> Cc: linux-kernel@vger.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
parent
1cac7b1ae3
commit
8343aae661
19 changed files with 36 additions and 43 deletions
|
@ -351,7 +351,7 @@ static int collect_events(struct perf_event *group, int max_count,
|
|||
evtype[n] = group->hw.event_base;
|
||||
current_idx[n++] = PMC_NO_INDEX;
|
||||
}
|
||||
list_for_each_entry(pe, &group->sibling_list, group_entry) {
|
||||
list_for_each_entry(pe, &group->sibling_list, sibling_list) {
|
||||
if (!is_software_event(pe) && pe->state != PERF_EVENT_STATE_OFF) {
|
||||
if (n >= max_count)
|
||||
return -1;
|
||||
|
|
|
@ -269,7 +269,7 @@ static bool mmdc_pmu_group_is_valid(struct perf_event *event)
|
|||
return false;
|
||||
}
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (!mmdc_pmu_group_event_is_valid(sibling, pmu, &counter_mask))
|
||||
return false;
|
||||
}
|
||||
|
|
|
@ -293,7 +293,7 @@ static bool l2x0_pmu_group_is_valid(struct perf_event *event)
|
|||
else if (!is_software_event(leader))
|
||||
return false;
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (sibling->pmu == pmu)
|
||||
num_hw++;
|
||||
else if (!is_software_event(sibling))
|
||||
|
|
|
@ -711,7 +711,7 @@ static int validate_group(struct perf_event *event)
|
|||
if (mipsxx_pmu_alloc_counter(&fake_cpuc, &leader->hw) < 0)
|
||||
return -EINVAL;
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (mipsxx_pmu_alloc_counter(&fake_cpuc, &sibling->hw) < 0)
|
||||
return -EINVAL;
|
||||
}
|
||||
|
|
|
@ -1426,7 +1426,7 @@ static int collect_events(struct perf_event *group, int max_count,
|
|||
flags[n] = group->hw.event_base;
|
||||
events[n++] = group->hw.config;
|
||||
}
|
||||
list_for_each_entry(event, &group->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &group->sibling_list, sibling_list) {
|
||||
if (event->pmu->task_ctx_nr == perf_hw_context &&
|
||||
event->state != PERF_EVENT_STATE_OFF) {
|
||||
if (n >= max_count)
|
||||
|
|
|
@ -277,7 +277,7 @@ static int collect_events(struct perf_event *group, int max_count,
|
|||
ctrs[n] = group;
|
||||
n++;
|
||||
}
|
||||
list_for_each_entry(event, &group->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &group->sibling_list, sibling_list) {
|
||||
if (!is_software_event(event) &&
|
||||
event->state != PERF_EVENT_STATE_OFF) {
|
||||
if (n >= max_count)
|
||||
|
|
|
@ -1342,7 +1342,7 @@ static int collect_events(struct perf_event *group, int max_count,
|
|||
events[n] = group->hw.event_base;
|
||||
current_idx[n++] = PIC_NO_INDEX;
|
||||
}
|
||||
list_for_each_entry(event, &group->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &group->sibling_list, sibling_list) {
|
||||
if (!is_software_event(event) &&
|
||||
event->state != PERF_EVENT_STATE_OFF) {
|
||||
if (n >= max_count)
|
||||
|
|
|
@ -990,7 +990,7 @@ static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader,
|
|||
if (!dogrp)
|
||||
return n;
|
||||
|
||||
list_for_each_entry(event, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &leader->sibling_list, sibling_list) {
|
||||
if (!is_x86_event(event) ||
|
||||
event->state <= PERF_EVENT_STATE_OFF)
|
||||
continue;
|
||||
|
|
|
@ -354,7 +354,7 @@ uncore_collect_events(struct intel_uncore_box *box, struct perf_event *leader,
|
|||
if (!dogrp)
|
||||
return n;
|
||||
|
||||
list_for_each_entry(event, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &leader->sibling_list, sibling_list) {
|
||||
if (!is_box_event(box, event) ||
|
||||
event->state <= PERF_EVENT_STATE_OFF)
|
||||
continue;
|
||||
|
|
|
@ -1311,7 +1311,7 @@ validate_group(struct perf_event *event)
|
|||
if (!validate_event(event->pmu, &fake_pmu, leader))
|
||||
return -EINVAL;
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (!validate_event(event->pmu, &fake_pmu, sibling))
|
||||
return -EINVAL;
|
||||
}
|
||||
|
|
|
@ -847,7 +847,7 @@ static int arm_ccn_pmu_event_init(struct perf_event *event)
|
|||
return -EINVAL;
|
||||
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list,
|
||||
group_entry)
|
||||
sibling_list)
|
||||
if (sibling->pmu != event->pmu &&
|
||||
!is_software_event(sibling))
|
||||
return -EINVAL;
|
||||
|
|
|
@ -536,7 +536,7 @@ static bool dsu_pmu_validate_group(struct perf_event *event)
|
|||
memset(fake_hw.used_mask, 0, sizeof(fake_hw.used_mask));
|
||||
if (!dsu_pmu_validate_event(event->pmu, &fake_hw, leader))
|
||||
return false;
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (!dsu_pmu_validate_event(event->pmu, &fake_hw, sibling))
|
||||
return false;
|
||||
}
|
||||
|
|
|
@ -311,7 +311,7 @@ validate_group(struct perf_event *event)
|
|||
if (!validate_event(event->pmu, &fake_pmu, leader))
|
||||
return -EINVAL;
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (!validate_event(event->pmu, &fake_pmu, sibling))
|
||||
return -EINVAL;
|
||||
}
|
||||
|
|
|
@ -82,8 +82,7 @@ static bool hisi_validate_event_group(struct perf_event *event)
|
|||
counters++;
|
||||
}
|
||||
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list,
|
||||
group_entry) {
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list, sibling_list) {
|
||||
if (is_software_event(sibling))
|
||||
continue;
|
||||
if (sibling->pmu != event->pmu)
|
||||
|
|
|
@ -535,7 +535,7 @@ static int l2_cache_event_init(struct perf_event *event)
|
|||
}
|
||||
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list,
|
||||
group_entry)
|
||||
sibling_list)
|
||||
if (sibling->pmu != event->pmu &&
|
||||
!is_software_event(sibling)) {
|
||||
dev_dbg_ratelimited(&l2cache_pmu->pdev->dev,
|
||||
|
@ -572,7 +572,7 @@ static int l2_cache_event_init(struct perf_event *event)
|
|||
}
|
||||
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list,
|
||||
group_entry) {
|
||||
sibling_list) {
|
||||
if ((sibling != event) &&
|
||||
!is_software_event(sibling) &&
|
||||
(L2_EVT_GROUP(sibling->attr.config) ==
|
||||
|
|
|
@ -468,7 +468,7 @@ static bool qcom_l3_cache__validate_event_group(struct perf_event *event)
|
|||
counters = event_num_counters(event);
|
||||
counters += event_num_counters(leader);
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list) {
|
||||
if (is_software_event(sibling))
|
||||
continue;
|
||||
if (sibling->pmu != event->pmu)
|
||||
|
|
|
@ -950,7 +950,7 @@ static int xgene_perf_event_init(struct perf_event *event)
|
|||
return -EINVAL;
|
||||
|
||||
list_for_each_entry(sibling, &event->group_leader->sibling_list,
|
||||
group_entry)
|
||||
sibling_list)
|
||||
if (sibling->pmu != event->pmu &&
|
||||
!is_software_event(sibling))
|
||||
return -EINVAL;
|
||||
|
|
|
@ -549,14 +549,9 @@ struct perf_event {
|
|||
struct list_head event_entry;
|
||||
|
||||
/*
|
||||
* XXX: group_entry and sibling_list should be mutually exclusive;
|
||||
* either you're a sibling on a group, or you're the group leader.
|
||||
* Rework the code to always use the same list element.
|
||||
*
|
||||
* Locked for modification by both ctx->mutex and ctx->lock; holding
|
||||
* either sufficies for read.
|
||||
*/
|
||||
struct list_head group_entry;
|
||||
struct list_head sibling_list;
|
||||
/*
|
||||
* Node on the pinned or flexible tree located at the event context;
|
||||
|
|
|
@ -643,7 +643,7 @@ static void perf_event_update_sibling_time(struct perf_event *leader)
|
|||
{
|
||||
struct perf_event *sibling;
|
||||
|
||||
list_for_each_entry(sibling, &leader->sibling_list, group_entry)
|
||||
list_for_each_entry(sibling, &leader->sibling_list, sibling_list)
|
||||
perf_event_update_time(sibling);
|
||||
}
|
||||
|
||||
|
@ -1835,12 +1835,12 @@ static void perf_group_attach(struct perf_event *event)
|
|||
|
||||
group_leader->group_caps &= event->event_caps;
|
||||
|
||||
list_add_tail(&event->group_entry, &group_leader->sibling_list);
|
||||
list_add_tail(&event->sibling_list, &group_leader->sibling_list);
|
||||
group_leader->nr_siblings++;
|
||||
|
||||
perf_event__header_size(group_leader);
|
||||
|
||||
list_for_each_entry(pos, &group_leader->sibling_list, group_entry)
|
||||
list_for_each_entry(pos, &group_leader->sibling_list, sibling_list)
|
||||
perf_event__header_size(pos);
|
||||
}
|
||||
|
||||
|
@ -1904,7 +1904,7 @@ static void perf_group_detach(struct perf_event *event)
|
|||
* If this is a sibling, remove it from its group.
|
||||
*/
|
||||
if (event->group_leader != event) {
|
||||
list_del_init(&event->group_entry);
|
||||
list_del_init(&event->sibling_list);
|
||||
event->group_leader->nr_siblings--;
|
||||
goto out;
|
||||
}
|
||||
|
@ -1914,7 +1914,7 @@ static void perf_group_detach(struct perf_event *event)
|
|||
* upgrade the siblings to singleton events by adding them
|
||||
* to whatever list we are on.
|
||||
*/
|
||||
list_for_each_entry_safe(sibling, tmp, &event->sibling_list, group_entry) {
|
||||
list_for_each_entry_safe(sibling, tmp, &event->sibling_list, sibling_list) {
|
||||
|
||||
sibling->group_leader = sibling;
|
||||
|
||||
|
@ -1922,7 +1922,7 @@ static void perf_group_detach(struct perf_event *event)
|
|||
sibling->group_caps = event->group_caps;
|
||||
|
||||
if (!RB_EMPTY_NODE(&event->group_node)) {
|
||||
list_del_init(&sibling->group_entry);
|
||||
list_del_init(&sibling->sibling_list);
|
||||
add_event_to_groups(sibling, event->ctx);
|
||||
}
|
||||
|
||||
|
@ -1932,7 +1932,7 @@ static void perf_group_detach(struct perf_event *event)
|
|||
out:
|
||||
perf_event__header_size(event->group_leader);
|
||||
|
||||
list_for_each_entry(tmp, &event->group_leader->sibling_list, group_entry)
|
||||
list_for_each_entry(tmp, &event->group_leader->sibling_list, sibling_list)
|
||||
perf_event__header_size(tmp);
|
||||
}
|
||||
|
||||
|
@ -1960,7 +1960,7 @@ static inline int pmu_filter_match(struct perf_event *event)
|
|||
if (!__pmu_filter_match(event))
|
||||
return 0;
|
||||
|
||||
list_for_each_entry(child, &event->sibling_list, group_entry) {
|
||||
list_for_each_entry(child, &event->sibling_list, sibling_list) {
|
||||
if (!__pmu_filter_match(child))
|
||||
return 0;
|
||||
}
|
||||
|
@ -2028,7 +2028,7 @@ group_sched_out(struct perf_event *group_event,
|
|||
/*
|
||||
* Schedule out siblings (if any):
|
||||
*/
|
||||
list_for_each_entry(event, &group_event->sibling_list, group_entry)
|
||||
list_for_each_entry(event, &group_event->sibling_list, sibling_list)
|
||||
event_sched_out(event, cpuctx, ctx);
|
||||
|
||||
perf_pmu_enable(ctx->pmu);
|
||||
|
@ -2307,7 +2307,7 @@ group_sched_in(struct perf_event *group_event,
|
|||
/*
|
||||
* Schedule in siblings as one group (if any):
|
||||
*/
|
||||
list_for_each_entry(event, &group_event->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &group_event->sibling_list, sibling_list) {
|
||||
if (event_sched_in(event, cpuctx, ctx)) {
|
||||
partial_group = event;
|
||||
goto group_error;
|
||||
|
@ -2323,7 +2323,7 @@ group_sched_in(struct perf_event *group_event,
|
|||
* partial group before returning:
|
||||
* The events up to the failed event are scheduled out normally.
|
||||
*/
|
||||
list_for_each_entry(event, &group_event->sibling_list, group_entry) {
|
||||
list_for_each_entry(event, &group_event->sibling_list, sibling_list) {
|
||||
if (event == partial_group)
|
||||
break;
|
||||
|
||||
|
@ -3796,7 +3796,7 @@ static void __perf_event_read(void *info)
|
|||
|
||||
pmu->read(event);
|
||||
|
||||
list_for_each_entry(sub, &event->sibling_list, group_entry) {
|
||||
list_for_each_entry(sub, &event->sibling_list, sibling_list) {
|
||||
if (sub->state == PERF_EVENT_STATE_ACTIVE) {
|
||||
/*
|
||||
* Use sibling's PMU rather than @event's since
|
||||
|
@ -4642,7 +4642,7 @@ static int __perf_read_group_add(struct perf_event *leader,
|
|||
if (read_format & PERF_FORMAT_ID)
|
||||
values[n++] = primary_event_id(leader);
|
||||
|
||||
list_for_each_entry(sub, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sub, &leader->sibling_list, sibling_list) {
|
||||
values[n++] += perf_event_count(sub);
|
||||
if (read_format & PERF_FORMAT_ID)
|
||||
values[n++] = primary_event_id(sub);
|
||||
|
@ -4836,7 +4836,7 @@ static void perf_event_for_each(struct perf_event *event,
|
|||
event = event->group_leader;
|
||||
|
||||
perf_event_for_each_child(event, func);
|
||||
list_for_each_entry(sibling, &event->sibling_list, group_entry)
|
||||
list_for_each_entry(sibling, &event->sibling_list, sibling_list)
|
||||
perf_event_for_each_child(sibling, func);
|
||||
}
|
||||
|
||||
|
@ -5995,7 +5995,7 @@ static void perf_output_read_group(struct perf_output_handle *handle,
|
|||
|
||||
__output_copy(handle, values, n * sizeof(u64));
|
||||
|
||||
list_for_each_entry(sub, &leader->sibling_list, group_entry) {
|
||||
list_for_each_entry(sub, &leader->sibling_list, sibling_list) {
|
||||
n = 0;
|
||||
|
||||
if ((sub != event) &&
|
||||
|
@ -9813,7 +9813,6 @@ perf_event_alloc(struct perf_event_attr *attr, int cpu,
|
|||
mutex_init(&event->child_mutex);
|
||||
INIT_LIST_HEAD(&event->child_list);
|
||||
|
||||
INIT_LIST_HEAD(&event->group_entry);
|
||||
INIT_LIST_HEAD(&event->event_entry);
|
||||
INIT_LIST_HEAD(&event->sibling_list);
|
||||
init_event_group(event);
|
||||
|
@ -10581,7 +10580,7 @@ SYSCALL_DEFINE5(perf_event_open,
|
|||
put_ctx(gctx);
|
||||
|
||||
list_for_each_entry(sibling, &group_leader->sibling_list,
|
||||
group_entry) {
|
||||
sibling_list) {
|
||||
perf_remove_from_context(sibling, 0);
|
||||
put_ctx(gctx);
|
||||
}
|
||||
|
@ -10603,7 +10602,7 @@ SYSCALL_DEFINE5(perf_event_open,
|
|||
* reachable through the group lists.
|
||||
*/
|
||||
list_for_each_entry(sibling, &group_leader->sibling_list,
|
||||
group_entry) {
|
||||
sibling_list) {
|
||||
perf_event__state_init(sibling);
|
||||
perf_install_in_context(ctx, sibling, sibling->cpu);
|
||||
get_ctx(ctx);
|
||||
|
@ -11242,7 +11241,7 @@ static int inherit_group(struct perf_event *parent_event,
|
|||
* case inherit_event() will create individual events, similar to what
|
||||
* perf_group_detach() would do anyway.
|
||||
*/
|
||||
list_for_each_entry(sub, &parent_event->sibling_list, group_entry) {
|
||||
list_for_each_entry(sub, &parent_event->sibling_list, sibling_list) {
|
||||
child_ctr = inherit_event(sub, parent, parent_ctx,
|
||||
child, leader, child_ctx);
|
||||
if (IS_ERR(child_ctr))
|
||||
|
|
Loading…
Reference in a new issue