struct perf_counter *counter, *group_leader = sibling->group_leader;
list_for_each_entry(counter, &group_leader->sibling_list, list_entry) {
- perf_swcounter_update(counter);
+ counter->hw_ops->read(counter);
perf_swcounter_store_irq(sibling, counter->hw_event.type);
perf_swcounter_store_irq(sibling, atomic64_read(&counter->count));
}
static void perf_swcounter_interrupt(struct perf_counter *counter,
int nmi, struct pt_regs *regs)
{
- perf_swcounter_save_and_restart(counter);
-
switch (counter->hw_event.record_type) {
case PERF_RECORD_SIMPLE:
break;
wake_up(&counter->waitq);
}
+static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
+{
+ struct perf_counter *counter;
+ struct pt_regs *regs;
+
+ counter = container_of(hrtimer, struct perf_counter, hw.hrtimer);
+ counter->hw_ops->read(counter);
+
+ regs = get_irq_regs();
+ /*
+ * In case we exclude kernel IPs or are somehow not in interrupt
+ * context, provide the next best thing, the user IP.
+ */
+ if ((counter->hw_event.exclude_kernel || !regs) &&
+ !counter->hw_event.exclude_user)
+ regs = task_pt_regs(current);
+
+ if (regs)
+ perf_swcounter_interrupt(counter, 0, regs);
+
+ hrtimer_forward_now(hrtimer, ns_to_ktime(counter->hw.irq_period));
+
+ return HRTIMER_RESTART;
+}
+
+static void perf_swcounter_overflow(struct perf_counter *counter,
+ int nmi, struct pt_regs *regs)
+{
+ perf_swcounter_save_and_restart(counter);
+ perf_swcounter_interrupt(counter, nmi, regs);
+}
+
static int perf_swcounter_match(struct perf_counter *counter,
enum hw_event_types event,
struct pt_regs *regs)
return 1;
}
+static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
+ int nmi, struct pt_regs *regs)
+{
+ int neg = atomic64_add_negative(nr, &counter->hw.count);
+ if (counter->hw.irq_period && !neg)
+ perf_swcounter_overflow(counter, nmi, regs);
+}
+
static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
enum hw_event_types event, u64 nr,
int nmi, struct pt_regs *regs)
{
struct perf_counter *counter;
unsigned long flags;
- int neg;
if (list_empty(&ctx->counter_list))
return;
* XXX: make counter_list RCU safe
*/
list_for_each_entry(counter, &ctx->counter_list, list_entry) {
- if (perf_swcounter_match(counter, event, regs)) {
- neg = atomic64_add_negative(nr, &counter->hw.count);
- if (counter->hw.irq_period && !neg)
- perf_swcounter_interrupt(counter, nmi, regs);
- }
+ if (perf_swcounter_match(counter, event, regs))
+ perf_swcounter_add(counter, nr, nmi, regs);
}
spin_unlock_irqrestore(&ctx->lock, flags);
* Software counter: cpu wall time clock
*/
-static int cpu_clock_perf_counter_enable(struct perf_counter *counter)
-{
- int cpu = raw_smp_processor_id();
-
- atomic64_set(&counter->hw.prev_count, cpu_clock(cpu));
- return 0;
-}
-
static void cpu_clock_perf_counter_update(struct perf_counter *counter)
{
int cpu = raw_smp_processor_id();
atomic64_add(now - prev, &counter->count);
}
+static int cpu_clock_perf_counter_enable(struct perf_counter *counter)
+{
+ struct hw_perf_counter *hwc = &counter->hw;
+ int cpu = raw_smp_processor_id();
+
+ atomic64_set(&hwc->prev_count, cpu_clock(cpu));
+ if (hwc->irq_period) {
+ hrtimer_init(&hwc->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
+ hwc->hrtimer.function = perf_swcounter_hrtimer;
+ __hrtimer_start_range_ns(&hwc->hrtimer,
+ ns_to_ktime(hwc->irq_period), 0,
+ HRTIMER_MODE_REL, 0);
+ }
+
+ return 0;
+}
+
static void cpu_clock_perf_counter_disable(struct perf_counter *counter)
{
+ hrtimer_cancel(&counter->hw.hrtimer);
cpu_clock_perf_counter_update(counter);
}
atomic64_add(delta, &counter->count);
}
-static void task_clock_perf_counter_read(struct perf_counter *counter)
-{
- u64 now = task_clock_perf_counter_val(counter, 1);
-
- task_clock_perf_counter_update(counter, now);
-}
-
static int task_clock_perf_counter_enable(struct perf_counter *counter)
{
- if (counter->prev_state <= PERF_COUNTER_STATE_OFF)
- atomic64_set(&counter->hw.prev_count,
- task_clock_perf_counter_val(counter, 0));
+ struct hw_perf_counter *hwc = &counter->hw;
+
+ atomic64_set(&hwc->prev_count, task_clock_perf_counter_val(counter, 0));
+ if (hwc->irq_period) {
+ hrtimer_init(&hwc->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
+ hwc->hrtimer.function = perf_swcounter_hrtimer;
+ __hrtimer_start_range_ns(&hwc->hrtimer,
+ ns_to_ktime(hwc->irq_period), 0,
+ HRTIMER_MODE_REL, 0);
+ }
return 0;
}
static void task_clock_perf_counter_disable(struct perf_counter *counter)
{
- u64 now = task_clock_perf_counter_val(counter, 0);
+ hrtimer_cancel(&counter->hw.hrtimer);
+ task_clock_perf_counter_update(counter,
+ task_clock_perf_counter_val(counter, 0));
+}
- task_clock_perf_counter_update(counter, now);
+static void task_clock_perf_counter_read(struct perf_counter *counter)
+{
+ task_clock_perf_counter_update(counter,
+ task_clock_perf_counter_val(counter, 1));
}
static const struct hw_perf_counter_ops perf_ops_task_clock = {
*/
switch (counter->hw_event.type) {
case PERF_COUNT_CPU_CLOCK:
- if (!(counter->hw_event.exclude_user ||
- counter->hw_event.exclude_kernel ||
- counter->hw_event.exclude_hv))
- hw_ops = &perf_ops_cpu_clock;
+ hw_ops = &perf_ops_cpu_clock;
+
+ if (hw_event->irq_period && hw_event->irq_period < 10000)
+ hw_event->irq_period = 10000;
break;
case PERF_COUNT_TASK_CLOCK:
- if (counter->hw_event.exclude_user ||
- counter->hw_event.exclude_kernel ||
- counter->hw_event.exclude_hv)
- break;
/*
* If the user instantiates this as a per-cpu counter,
* use the cpu_clock counter instead.
hw_ops = &perf_ops_task_clock;
else
hw_ops = &perf_ops_cpu_clock;
+
+ if (hw_event->irq_period && hw_event->irq_period < 10000)
+ hw_event->irq_period = 10000;
break;
case PERF_COUNT_PAGE_FAULTS:
case PERF_COUNT_PAGE_FAULTS_MIN: