{
struct perf_event *sibling;
- list_for_each_entry(sibling, &leader->sibling_list, group_entry)
+ list_for_each_entry(sibling, &leader->sibling_list, sibling_list)
perf_event_update_time(sibling);
}
group_leader->group_caps &= event->event_caps;
- list_add_tail(&event->group_entry, &group_leader->sibling_list);
+ list_add_tail(&event->sibling_list, &group_leader->sibling_list);
group_leader->nr_siblings++;
perf_event__header_size(group_leader);
- list_for_each_entry(pos, &group_leader->sibling_list, group_entry)
+ list_for_each_entry(pos, &group_leader->sibling_list, sibling_list)
perf_event__header_size(pos);
}
* If this is a sibling, remove it from its group.
*/
if (event->group_leader != event) {
- list_del_init(&event->group_entry);
+ list_del_init(&event->sibling_list);
event->group_leader->nr_siblings--;
goto out;
}
* upgrade the siblings to singleton events by adding them
* to whatever list we are on.
*/
- list_for_each_entry_safe(sibling, tmp, &event->sibling_list, group_entry) {
+ list_for_each_entry_safe(sibling, tmp, &event->sibling_list, sibling_list) {
sibling->group_leader = sibling;
sibling->group_caps = event->group_caps;
if (!RB_EMPTY_NODE(&event->group_node)) {
- list_del_init(&sibling->group_entry);
+ list_del_init(&sibling->sibling_list);
add_event_to_groups(sibling, event->ctx);
}
out:
perf_event__header_size(event->group_leader);
- list_for_each_entry(tmp, &event->group_leader->sibling_list, group_entry)
+ list_for_each_entry(tmp, &event->group_leader->sibling_list, sibling_list)
perf_event__header_size(tmp);
}
if (!__pmu_filter_match(event))
return 0;
- list_for_each_entry(child, &event->sibling_list, group_entry) {
+ list_for_each_entry(child, &event->sibling_list, sibling_list) {
if (!__pmu_filter_match(child))
return 0;
}
/*
* Schedule out siblings (if any):
*/
- list_for_each_entry(event, &group_event->sibling_list, group_entry)
+ list_for_each_entry(event, &group_event->sibling_list, sibling_list)
event_sched_out(event, cpuctx, ctx);
perf_pmu_enable(ctx->pmu);
/*
* Schedule in siblings as one group (if any):
*/
- list_for_each_entry(event, &group_event->sibling_list, group_entry) {
+ list_for_each_entry(event, &group_event->sibling_list, sibling_list) {
if (event_sched_in(event, cpuctx, ctx)) {
partial_group = event;
goto group_error;
* partial group before returning:
* The events up to the failed event are scheduled out normally.
*/
- list_for_each_entry(event, &group_event->sibling_list, group_entry) {
+ list_for_each_entry(event, &group_event->sibling_list, sibling_list) {
if (event == partial_group)
break;
pmu->read(event);
- list_for_each_entry(sub, &event->sibling_list, group_entry) {
+ list_for_each_entry(sub, &event->sibling_list, sibling_list) {
if (sub->state == PERF_EVENT_STATE_ACTIVE) {
/*
* Use sibling's PMU rather than @event's since
if (read_format & PERF_FORMAT_ID)
values[n++] = primary_event_id(leader);
- list_for_each_entry(sub, &leader->sibling_list, group_entry) {
+ list_for_each_entry(sub, &leader->sibling_list, sibling_list) {
values[n++] += perf_event_count(sub);
if (read_format & PERF_FORMAT_ID)
values[n++] = primary_event_id(sub);
event = event->group_leader;
perf_event_for_each_child(event, func);
- list_for_each_entry(sibling, &event->sibling_list, group_entry)
+ list_for_each_entry(sibling, &event->sibling_list, sibling_list)
perf_event_for_each_child(sibling, func);
}
__output_copy(handle, values, n * sizeof(u64));
- list_for_each_entry(sub, &leader->sibling_list, group_entry) {
+ list_for_each_entry(sub, &leader->sibling_list, sibling_list) {
n = 0;
if ((sub != event) &&
mutex_init(&event->child_mutex);
INIT_LIST_HEAD(&event->child_list);
- INIT_LIST_HEAD(&event->group_entry);
INIT_LIST_HEAD(&event->event_entry);
INIT_LIST_HEAD(&event->sibling_list);
init_event_group(event);
put_ctx(gctx);
list_for_each_entry(sibling, &group_leader->sibling_list,
- group_entry) {
+ sibling_list) {
perf_remove_from_context(sibling, 0);
put_ctx(gctx);
}
* reachable through the group lists.
*/
list_for_each_entry(sibling, &group_leader->sibling_list,
- group_entry) {
+ sibling_list) {
perf_event__state_init(sibling);
perf_install_in_context(ctx, sibling, sibling->cpu);
get_ctx(ctx);
* case inherit_event() will create individual events, similar to what
* perf_group_detach() would do anyway.
*/
- list_for_each_entry(sub, &parent_event->sibling_list, group_entry) {
+ list_for_each_entry(sub, &parent_event->sibling_list, sibling_list) {
child_ctr = inherit_event(sub, parent, parent_ctx,
child, leader, child_ctx);
if (IS_ERR(child_ctr))